diff --git "a/adapter/checkpoint-1220/trainer_state.json" "b/adapter/checkpoint-1220/trainer_state.json" new file mode 100644--- /dev/null +++ "b/adapter/checkpoint-1220/trainer_state.json" @@ -0,0 +1,8561 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.9975440032746623, + "eval_steps": 500, + "global_step": 1220, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0, + "grad_norm": 0.3300960063934326, + "learning_rate": 2.9999999999999997e-05, + "loss": 0.9966, + "step": 1 + }, + { + "epoch": 0.0, + "grad_norm": 0.4113194942474365, + "learning_rate": 5.9999999999999995e-05, + "loss": 1.1253, + "step": 2 + }, + { + "epoch": 0.0, + "grad_norm": 0.2486647665500641, + "learning_rate": 8.999999999999999e-05, + "loss": 1.0721, + "step": 3 + }, + { + "epoch": 0.01, + "grad_norm": 0.2249160259962082, + "learning_rate": 0.00011999999999999999, + "loss": 0.9033, + "step": 4 + }, + { + "epoch": 0.01, + "grad_norm": 0.3706735074520111, + "learning_rate": 0.00015, + "loss": 1.0498, + "step": 5 + }, + { + "epoch": 0.01, + "grad_norm": 0.28104931116104126, + "learning_rate": 0.00017999999999999998, + "loss": 0.9108, + "step": 6 + }, + { + "epoch": 0.01, + "grad_norm": 0.27497801184654236, + "learning_rate": 0.00020999999999999998, + "loss": 0.9038, + "step": 7 + }, + { + "epoch": 0.01, + "grad_norm": 0.30283215641975403, + "learning_rate": 0.00023999999999999998, + "loss": 0.8605, + "step": 8 + }, + { + "epoch": 0.01, + "grad_norm": 0.33457252383232117, + "learning_rate": 0.00027, + "loss": 0.9049, + "step": 9 + }, + { + "epoch": 0.02, + "grad_norm": 0.37725692987442017, + "learning_rate": 0.0003, + "loss": 0.772, + "step": 10 + }, + { + "epoch": 0.02, + "grad_norm": 0.27986466884613037, + "learning_rate": 0.00029975206611570246, + "loss": 0.7666, + "step": 11 + }, + { + "epoch": 0.02, + "grad_norm": 0.30687034130096436, + "learning_rate": 0.00029950413223140494, + "loss": 0.8312, + "step": 12 + }, + { + "epoch": 0.02, + "grad_norm": 0.3321741819381714, + "learning_rate": 0.0002992561983471074, + "loss": 0.8308, + "step": 13 + }, + { + "epoch": 0.02, + "grad_norm": 0.29080134630203247, + "learning_rate": 0.0002990082644628099, + "loss": 0.7597, + "step": 14 + }, + { + "epoch": 0.02, + "grad_norm": 0.33823856711387634, + "learning_rate": 0.0002987603305785124, + "loss": 0.8693, + "step": 15 + }, + { + "epoch": 0.03, + "grad_norm": 0.3461182117462158, + "learning_rate": 0.0002985123966942149, + "loss": 1.0571, + "step": 16 + }, + { + "epoch": 0.03, + "grad_norm": 0.22306275367736816, + "learning_rate": 0.0002982644628099173, + "loss": 0.7706, + "step": 17 + }, + { + "epoch": 0.03, + "grad_norm": 154.4940643310547, + "learning_rate": 0.0002980165289256198, + "loss": 2.6519, + "step": 18 + }, + { + "epoch": 0.03, + "grad_norm": 0.22956405580043793, + "learning_rate": 0.00029776859504132227, + "loss": 0.6897, + "step": 19 + }, + { + "epoch": 0.03, + "grad_norm": 0.25711989402770996, + "learning_rate": 0.00029752066115702476, + "loss": 0.7338, + "step": 20 + }, + { + "epoch": 0.03, + "grad_norm": 0.2565441131591797, + "learning_rate": 0.00029727272727272724, + "loss": 0.8211, + "step": 21 + }, + { + "epoch": 0.04, + "grad_norm": 0.2437434047460556, + "learning_rate": 0.0002970247933884297, + "loss": 0.8027, + "step": 22 + }, + { + "epoch": 0.04, + "grad_norm": 0.21284469962120056, + "learning_rate": 0.0002967768595041322, + "loss": 0.7944, + "step": 23 + }, + { + "epoch": 0.04, + "grad_norm": 0.23338356614112854, + "learning_rate": 0.0002965289256198347, + "loss": 0.7696, + "step": 24 + }, + { + "epoch": 0.04, + "grad_norm": 0.25512659549713135, + "learning_rate": 0.0002962809917355372, + "loss": 0.7693, + "step": 25 + }, + { + "epoch": 0.04, + "grad_norm": 0.19500921666622162, + "learning_rate": 0.0002960330578512396, + "loss": 0.7599, + "step": 26 + }, + { + "epoch": 0.04, + "grad_norm": 0.2554054260253906, + "learning_rate": 0.00029578512396694214, + "loss": 0.966, + "step": 27 + }, + { + "epoch": 0.05, + "grad_norm": 0.17682747542858124, + "learning_rate": 0.0002955371900826446, + "loss": 0.676, + "step": 28 + }, + { + "epoch": 0.05, + "grad_norm": 0.20516635477542877, + "learning_rate": 0.0002952892561983471, + "loss": 0.8144, + "step": 29 + }, + { + "epoch": 0.05, + "grad_norm": 0.3275119662284851, + "learning_rate": 0.0002950413223140496, + "loss": 0.7704, + "step": 30 + }, + { + "epoch": 0.05, + "grad_norm": 0.22231778502464294, + "learning_rate": 0.000294793388429752, + "loss": 0.7614, + "step": 31 + }, + { + "epoch": 0.05, + "grad_norm": 0.17065812647342682, + "learning_rate": 0.0002945454545454545, + "loss": 0.5634, + "step": 32 + }, + { + "epoch": 0.05, + "grad_norm": 0.1771956831216812, + "learning_rate": 0.000294297520661157, + "loss": 0.7607, + "step": 33 + }, + { + "epoch": 0.06, + "grad_norm": 0.26693442463874817, + "learning_rate": 0.00029404958677685947, + "loss": 0.8171, + "step": 34 + }, + { + "epoch": 0.06, + "grad_norm": 1.409070611000061, + "learning_rate": 0.00029380165289256196, + "loss": 0.7791, + "step": 35 + }, + { + "epoch": 0.06, + "grad_norm": 0.20727217197418213, + "learning_rate": 0.00029355371900826444, + "loss": 0.7357, + "step": 36 + }, + { + "epoch": 0.06, + "grad_norm": 0.2145707905292511, + "learning_rate": 0.0002933057851239669, + "loss": 0.8458, + "step": 37 + }, + { + "epoch": 0.06, + "grad_norm": 0.2068527340888977, + "learning_rate": 0.0002930578512396694, + "loss": 0.78, + "step": 38 + }, + { + "epoch": 0.06, + "grad_norm": 0.22432388365268707, + "learning_rate": 0.00029280991735537184, + "loss": 0.8523, + "step": 39 + }, + { + "epoch": 0.07, + "grad_norm": 0.19982610642910004, + "learning_rate": 0.0002925619834710743, + "loss": 0.7372, + "step": 40 + }, + { + "epoch": 0.07, + "grad_norm": 6.248472213745117, + "learning_rate": 0.00029231404958677686, + "loss": 0.7399, + "step": 41 + }, + { + "epoch": 0.07, + "grad_norm": 0.2269737422466278, + "learning_rate": 0.00029206611570247934, + "loss": 0.7842, + "step": 42 + }, + { + "epoch": 0.07, + "grad_norm": 0.23117898404598236, + "learning_rate": 0.0002918181818181818, + "loss": 0.7111, + "step": 43 + }, + { + "epoch": 0.07, + "grad_norm": 0.22466522455215454, + "learning_rate": 0.00029157024793388425, + "loss": 0.8979, + "step": 44 + }, + { + "epoch": 0.07, + "grad_norm": 0.20770332217216492, + "learning_rate": 0.00029132231404958674, + "loss": 0.774, + "step": 45 + }, + { + "epoch": 0.08, + "grad_norm": 0.2376495748758316, + "learning_rate": 0.0002910743801652892, + "loss": 0.7216, + "step": 46 + }, + { + "epoch": 0.08, + "grad_norm": 0.2470778226852417, + "learning_rate": 0.0002908264462809917, + "loss": 0.7369, + "step": 47 + }, + { + "epoch": 0.08, + "grad_norm": 0.23465900123119354, + "learning_rate": 0.0002905785123966942, + "loss": 0.7528, + "step": 48 + }, + { + "epoch": 0.08, + "grad_norm": 0.5718627572059631, + "learning_rate": 0.00029033057851239667, + "loss": 0.7535, + "step": 49 + }, + { + "epoch": 0.08, + "grad_norm": 0.21493370831012726, + "learning_rate": 0.00029008264462809916, + "loss": 0.8593, + "step": 50 + }, + { + "epoch": 0.08, + "grad_norm": 0.21197210252285004, + "learning_rate": 0.00028983471074380164, + "loss": 0.8013, + "step": 51 + }, + { + "epoch": 0.09, + "grad_norm": 0.20836398005485535, + "learning_rate": 0.0002895867768595041, + "loss": 0.7905, + "step": 52 + }, + { + "epoch": 0.09, + "grad_norm": 0.2096678912639618, + "learning_rate": 0.00028933884297520655, + "loss": 0.6754, + "step": 53 + }, + { + "epoch": 0.09, + "grad_norm": 0.25898435711860657, + "learning_rate": 0.00028909090909090904, + "loss": 0.7725, + "step": 54 + }, + { + "epoch": 0.09, + "grad_norm": 0.23370735347270966, + "learning_rate": 0.0002888429752066116, + "loss": 0.7007, + "step": 55 + }, + { + "epoch": 0.09, + "grad_norm": 0.23006942868232727, + "learning_rate": 0.00028859504132231406, + "loss": 0.7534, + "step": 56 + }, + { + "epoch": 0.09, + "grad_norm": 0.20855402946472168, + "learning_rate": 0.0002883471074380165, + "loss": 0.9491, + "step": 57 + }, + { + "epoch": 0.09, + "grad_norm": 0.24340493977069855, + "learning_rate": 0.00028809917355371897, + "loss": 0.8089, + "step": 58 + }, + { + "epoch": 0.1, + "grad_norm": 0.20169466733932495, + "learning_rate": 0.00028785123966942145, + "loss": 0.64, + "step": 59 + }, + { + "epoch": 0.1, + "grad_norm": 0.23272906243801117, + "learning_rate": 0.00028760330578512394, + "loss": 0.8456, + "step": 60 + }, + { + "epoch": 0.1, + "grad_norm": 0.1767100691795349, + "learning_rate": 0.0002873553719008264, + "loss": 0.6686, + "step": 61 + }, + { + "epoch": 0.1, + "grad_norm": 0.24511106312274933, + "learning_rate": 0.0002871074380165289, + "loss": 0.6998, + "step": 62 + }, + { + "epoch": 0.1, + "grad_norm": 0.22284479439258575, + "learning_rate": 0.0002868595041322314, + "loss": 0.6699, + "step": 63 + }, + { + "epoch": 0.1, + "grad_norm": 0.21842750906944275, + "learning_rate": 0.00028661157024793387, + "loss": 0.7413, + "step": 64 + }, + { + "epoch": 0.11, + "grad_norm": 0.2669163644313812, + "learning_rate": 0.00028636363636363636, + "loss": 0.931, + "step": 65 + }, + { + "epoch": 0.11, + "grad_norm": 0.1864808052778244, + "learning_rate": 0.0002861157024793388, + "loss": 0.5652, + "step": 66 + }, + { + "epoch": 0.11, + "grad_norm": 0.18369853496551514, + "learning_rate": 0.00028586776859504127, + "loss": 0.6847, + "step": 67 + }, + { + "epoch": 0.11, + "grad_norm": 0.22353056073188782, + "learning_rate": 0.00028561983471074375, + "loss": 0.598, + "step": 68 + }, + { + "epoch": 0.11, + "grad_norm": 0.20269523561000824, + "learning_rate": 0.0002853719008264463, + "loss": 0.8688, + "step": 69 + }, + { + "epoch": 0.11, + "grad_norm": 0.2291198968887329, + "learning_rate": 0.0002851239669421488, + "loss": 0.7535, + "step": 70 + }, + { + "epoch": 0.12, + "grad_norm": 0.22033120691776276, + "learning_rate": 0.0002848760330578512, + "loss": 0.8377, + "step": 71 + }, + { + "epoch": 0.12, + "grad_norm": 0.2687983214855194, + "learning_rate": 0.0002846280991735537, + "loss": 0.6926, + "step": 72 + }, + { + "epoch": 0.12, + "grad_norm": 0.1933681070804596, + "learning_rate": 0.00028438016528925617, + "loss": 0.6276, + "step": 73 + }, + { + "epoch": 0.12, + "grad_norm": 0.2820705473423004, + "learning_rate": 0.00028413223140495865, + "loss": 0.848, + "step": 74 + }, + { + "epoch": 0.12, + "grad_norm": 0.19532324373722076, + "learning_rate": 0.00028388429752066114, + "loss": 0.6198, + "step": 75 + }, + { + "epoch": 0.12, + "grad_norm": 0.25057846307754517, + "learning_rate": 0.0002836363636363636, + "loss": 0.6838, + "step": 76 + }, + { + "epoch": 0.13, + "grad_norm": 0.2168462574481964, + "learning_rate": 0.0002833884297520661, + "loss": 0.7885, + "step": 77 + }, + { + "epoch": 0.13, + "grad_norm": 0.2106674313545227, + "learning_rate": 0.0002831404958677686, + "loss": 0.6757, + "step": 78 + }, + { + "epoch": 0.13, + "grad_norm": 0.24460363388061523, + "learning_rate": 0.000282892561983471, + "loss": 0.7414, + "step": 79 + }, + { + "epoch": 0.13, + "grad_norm": 0.3706071078777313, + "learning_rate": 0.0002826446280991735, + "loss": 0.621, + "step": 80 + }, + { + "epoch": 0.13, + "grad_norm": 0.2251998782157898, + "learning_rate": 0.000282396694214876, + "loss": 0.7453, + "step": 81 + }, + { + "epoch": 0.13, + "grad_norm": 0.24521738290786743, + "learning_rate": 0.00028214876033057847, + "loss": 0.6985, + "step": 82 + }, + { + "epoch": 0.14, + "grad_norm": 0.2262742966413498, + "learning_rate": 0.000281900826446281, + "loss": 0.6316, + "step": 83 + }, + { + "epoch": 0.14, + "grad_norm": 0.19723354279994965, + "learning_rate": 0.00028165289256198344, + "loss": 0.4798, + "step": 84 + }, + { + "epoch": 0.14, + "grad_norm": 0.20684833824634552, + "learning_rate": 0.0002814049586776859, + "loss": 0.7993, + "step": 85 + }, + { + "epoch": 0.14, + "grad_norm": 0.19534814357757568, + "learning_rate": 0.0002811570247933884, + "loss": 0.7735, + "step": 86 + }, + { + "epoch": 0.14, + "grad_norm": 0.2585545480251312, + "learning_rate": 0.0002809090909090909, + "loss": 0.8126, + "step": 87 + }, + { + "epoch": 0.14, + "grad_norm": 0.2510583996772766, + "learning_rate": 0.00028066115702479337, + "loss": 0.6973, + "step": 88 + }, + { + "epoch": 0.15, + "grad_norm": 0.1884051263332367, + "learning_rate": 0.00028041322314049585, + "loss": 0.701, + "step": 89 + }, + { + "epoch": 0.15, + "grad_norm": 0.2526257038116455, + "learning_rate": 0.00028016528925619834, + "loss": 0.7132, + "step": 90 + }, + { + "epoch": 0.15, + "grad_norm": 0.200734481215477, + "learning_rate": 0.0002799173553719008, + "loss": 0.7024, + "step": 91 + }, + { + "epoch": 0.15, + "grad_norm": 0.2404022514820099, + "learning_rate": 0.0002796694214876033, + "loss": 0.704, + "step": 92 + }, + { + "epoch": 0.15, + "grad_norm": 0.23063871264457703, + "learning_rate": 0.00027942148760330573, + "loss": 0.6312, + "step": 93 + }, + { + "epoch": 0.15, + "grad_norm": 0.1759747564792633, + "learning_rate": 0.0002791735537190082, + "loss": 0.6577, + "step": 94 + }, + { + "epoch": 0.16, + "grad_norm": 0.2009582370519638, + "learning_rate": 0.0002789256198347107, + "loss": 0.8036, + "step": 95 + }, + { + "epoch": 0.16, + "grad_norm": 0.2200164943933487, + "learning_rate": 0.0002786776859504132, + "loss": 0.7101, + "step": 96 + }, + { + "epoch": 0.16, + "grad_norm": 0.19693537056446075, + "learning_rate": 0.00027842975206611567, + "loss": 0.6221, + "step": 97 + }, + { + "epoch": 0.16, + "grad_norm": 0.23269779980182648, + "learning_rate": 0.00027818181818181815, + "loss": 0.8264, + "step": 98 + }, + { + "epoch": 0.16, + "grad_norm": 0.2440226823091507, + "learning_rate": 0.00027793388429752064, + "loss": 0.8051, + "step": 99 + }, + { + "epoch": 0.16, + "grad_norm": 0.2307034134864807, + "learning_rate": 0.0002776859504132231, + "loss": 0.631, + "step": 100 + }, + { + "epoch": 0.17, + "grad_norm": 0.2530567944049835, + "learning_rate": 0.0002774380165289256, + "loss": 0.8616, + "step": 101 + }, + { + "epoch": 0.17, + "grad_norm": 0.2808806300163269, + "learning_rate": 0.0002771900826446281, + "loss": 0.8333, + "step": 102 + }, + { + "epoch": 0.17, + "grad_norm": 0.20667941868305206, + "learning_rate": 0.00027694214876033057, + "loss": 0.7212, + "step": 103 + }, + { + "epoch": 0.17, + "grad_norm": 0.17540781199932098, + "learning_rate": 0.00027669421487603305, + "loss": 0.5964, + "step": 104 + }, + { + "epoch": 0.17, + "grad_norm": 0.2526637613773346, + "learning_rate": 0.00027644628099173554, + "loss": 0.6868, + "step": 105 + }, + { + "epoch": 0.17, + "grad_norm": 0.2137339860200882, + "learning_rate": 0.00027619834710743797, + "loss": 0.6155, + "step": 106 + }, + { + "epoch": 0.18, + "grad_norm": 0.21061092615127563, + "learning_rate": 0.00027595041322314045, + "loss": 0.813, + "step": 107 + }, + { + "epoch": 0.18, + "grad_norm": 0.21619191765785217, + "learning_rate": 0.00027570247933884293, + "loss": 0.8046, + "step": 108 + }, + { + "epoch": 0.18, + "grad_norm": 0.2212170660495758, + "learning_rate": 0.0002754545454545454, + "loss": 0.6706, + "step": 109 + }, + { + "epoch": 0.18, + "grad_norm": 0.23427413403987885, + "learning_rate": 0.0002752066115702479, + "loss": 0.7152, + "step": 110 + }, + { + "epoch": 0.18, + "grad_norm": 0.20566123723983765, + "learning_rate": 0.0002749586776859504, + "loss": 0.6568, + "step": 111 + }, + { + "epoch": 0.18, + "grad_norm": 0.22977930307388306, + "learning_rate": 0.00027471074380165287, + "loss": 0.7832, + "step": 112 + }, + { + "epoch": 0.19, + "grad_norm": 0.28307485580444336, + "learning_rate": 0.00027446280991735535, + "loss": 0.7446, + "step": 113 + }, + { + "epoch": 0.19, + "grad_norm": 0.19567596912384033, + "learning_rate": 0.00027421487603305784, + "loss": 0.6394, + "step": 114 + }, + { + "epoch": 0.19, + "grad_norm": 0.24577689170837402, + "learning_rate": 0.0002739669421487603, + "loss": 0.6389, + "step": 115 + }, + { + "epoch": 0.19, + "grad_norm": 0.2180463820695877, + "learning_rate": 0.0002737190082644628, + "loss": 0.7814, + "step": 116 + }, + { + "epoch": 0.19, + "grad_norm": 0.19546380639076233, + "learning_rate": 0.0002734710743801653, + "loss": 0.8312, + "step": 117 + }, + { + "epoch": 0.19, + "grad_norm": 0.22698360681533813, + "learning_rate": 0.00027322314049586777, + "loss": 0.7443, + "step": 118 + }, + { + "epoch": 0.19, + "grad_norm": 0.22987066209316254, + "learning_rate": 0.0002729752066115702, + "loss": 0.7839, + "step": 119 + }, + { + "epoch": 0.2, + "grad_norm": 0.20548178255558014, + "learning_rate": 0.0002727272727272727, + "loss": 0.7805, + "step": 120 + }, + { + "epoch": 0.2, + "grad_norm": 0.2477702796459198, + "learning_rate": 0.00027247933884297517, + "loss": 0.5694, + "step": 121 + }, + { + "epoch": 0.2, + "grad_norm": 0.20593340694904327, + "learning_rate": 0.00027223140495867765, + "loss": 0.6479, + "step": 122 + }, + { + "epoch": 0.2, + "grad_norm": 0.23635917901992798, + "learning_rate": 0.00027198347107438013, + "loss": 0.8107, + "step": 123 + }, + { + "epoch": 0.2, + "grad_norm": 0.25808119773864746, + "learning_rate": 0.0002717355371900826, + "loss": 0.7876, + "step": 124 + }, + { + "epoch": 0.2, + "grad_norm": 0.22156469523906708, + "learning_rate": 0.0002714876033057851, + "loss": 0.7261, + "step": 125 + }, + { + "epoch": 0.21, + "grad_norm": 0.19892215728759766, + "learning_rate": 0.0002712396694214876, + "loss": 0.6874, + "step": 126 + }, + { + "epoch": 0.21, + "grad_norm": 0.24936752021312714, + "learning_rate": 0.00027099173553719007, + "loss": 0.6155, + "step": 127 + }, + { + "epoch": 0.21, + "grad_norm": 0.23287539184093475, + "learning_rate": 0.0002707438016528925, + "loss": 0.602, + "step": 128 + }, + { + "epoch": 0.21, + "grad_norm": 0.2086639404296875, + "learning_rate": 0.00027049586776859504, + "loss": 0.7198, + "step": 129 + }, + { + "epoch": 0.21, + "grad_norm": 0.24974922835826874, + "learning_rate": 0.0002702479338842975, + "loss": 0.6873, + "step": 130 + }, + { + "epoch": 0.21, + "grad_norm": 0.2066827118396759, + "learning_rate": 0.00027, + "loss": 0.5821, + "step": 131 + }, + { + "epoch": 0.22, + "grad_norm": 0.28004395961761475, + "learning_rate": 0.0002697520661157025, + "loss": 0.7864, + "step": 132 + }, + { + "epoch": 0.22, + "grad_norm": 0.22391608357429504, + "learning_rate": 0.0002695041322314049, + "loss": 0.6773, + "step": 133 + }, + { + "epoch": 0.22, + "grad_norm": 0.2821199297904968, + "learning_rate": 0.0002692561983471074, + "loss": 0.6806, + "step": 134 + }, + { + "epoch": 0.22, + "grad_norm": 0.21736428141593933, + "learning_rate": 0.0002690082644628099, + "loss": 0.6662, + "step": 135 + }, + { + "epoch": 0.22, + "grad_norm": 0.23889939486980438, + "learning_rate": 0.00026876033057851237, + "loss": 0.6356, + "step": 136 + }, + { + "epoch": 0.22, + "grad_norm": 0.21096719801425934, + "learning_rate": 0.00026851239669421485, + "loss": 0.6762, + "step": 137 + }, + { + "epoch": 0.23, + "grad_norm": 0.22622421383857727, + "learning_rate": 0.00026826446280991733, + "loss": 0.8085, + "step": 138 + }, + { + "epoch": 0.23, + "grad_norm": 0.19824957847595215, + "learning_rate": 0.0002680165289256198, + "loss": 0.6031, + "step": 139 + }, + { + "epoch": 0.23, + "grad_norm": 0.24482691287994385, + "learning_rate": 0.0002677685950413223, + "loss": 0.6649, + "step": 140 + }, + { + "epoch": 0.23, + "grad_norm": 0.21291929483413696, + "learning_rate": 0.0002675206611570248, + "loss": 0.6671, + "step": 141 + }, + { + "epoch": 0.23, + "grad_norm": 0.2202674299478531, + "learning_rate": 0.0002672727272727272, + "loss": 0.6469, + "step": 142 + }, + { + "epoch": 0.23, + "grad_norm": 0.23572632670402527, + "learning_rate": 0.0002670247933884297, + "loss": 0.7377, + "step": 143 + }, + { + "epoch": 0.24, + "grad_norm": 0.2051907777786255, + "learning_rate": 0.00026677685950413224, + "loss": 0.6217, + "step": 144 + }, + { + "epoch": 0.24, + "grad_norm": 0.23270072042942047, + "learning_rate": 0.0002665289256198347, + "loss": 0.7933, + "step": 145 + }, + { + "epoch": 0.24, + "grad_norm": 0.20652809739112854, + "learning_rate": 0.00026628099173553715, + "loss": 0.6007, + "step": 146 + }, + { + "epoch": 0.24, + "grad_norm": 0.23084674775600433, + "learning_rate": 0.00026603305785123963, + "loss": 0.701, + "step": 147 + }, + { + "epoch": 0.24, + "grad_norm": 0.25663891434669495, + "learning_rate": 0.0002657851239669421, + "loss": 0.7271, + "step": 148 + }, + { + "epoch": 0.24, + "grad_norm": 0.25880497694015503, + "learning_rate": 0.0002655371900826446, + "loss": 0.6562, + "step": 149 + }, + { + "epoch": 0.25, + "grad_norm": 0.19349205493927002, + "learning_rate": 0.0002652892561983471, + "loss": 0.5016, + "step": 150 + }, + { + "epoch": 0.25, + "grad_norm": 0.2401740401983261, + "learning_rate": 0.00026504132231404957, + "loss": 0.6978, + "step": 151 + }, + { + "epoch": 0.25, + "grad_norm": 0.19495394825935364, + "learning_rate": 0.00026479338842975205, + "loss": 0.5562, + "step": 152 + }, + { + "epoch": 0.25, + "grad_norm": 0.21485286951065063, + "learning_rate": 0.00026454545454545453, + "loss": 0.7847, + "step": 153 + }, + { + "epoch": 0.25, + "grad_norm": 0.241348534822464, + "learning_rate": 0.000264297520661157, + "loss": 0.7513, + "step": 154 + }, + { + "epoch": 0.25, + "grad_norm": 0.3316986858844757, + "learning_rate": 0.00026404958677685945, + "loss": 0.664, + "step": 155 + }, + { + "epoch": 0.26, + "grad_norm": 0.2419958859682083, + "learning_rate": 0.00026380165289256193, + "loss": 0.7322, + "step": 156 + }, + { + "epoch": 0.26, + "grad_norm": 0.2868640124797821, + "learning_rate": 0.0002635537190082644, + "loss": 0.7004, + "step": 157 + }, + { + "epoch": 0.26, + "grad_norm": 0.24806949496269226, + "learning_rate": 0.00026330578512396695, + "loss": 0.6497, + "step": 158 + }, + { + "epoch": 0.26, + "grad_norm": 0.23873400688171387, + "learning_rate": 0.00026305785123966944, + "loss": 0.7543, + "step": 159 + }, + { + "epoch": 0.26, + "grad_norm": 0.2480355203151703, + "learning_rate": 0.00026280991735537187, + "loss": 0.6048, + "step": 160 + }, + { + "epoch": 0.26, + "grad_norm": 0.2619112730026245, + "learning_rate": 0.00026256198347107435, + "loss": 0.762, + "step": 161 + }, + { + "epoch": 0.27, + "grad_norm": 0.22763262689113617, + "learning_rate": 0.00026231404958677683, + "loss": 0.6557, + "step": 162 + }, + { + "epoch": 0.27, + "grad_norm": 0.3291528522968292, + "learning_rate": 0.0002620661157024793, + "loss": 0.7059, + "step": 163 + }, + { + "epoch": 0.27, + "grad_norm": 0.2959338426589966, + "learning_rate": 0.0002618181818181818, + "loss": 0.6622, + "step": 164 + }, + { + "epoch": 0.27, + "grad_norm": 0.23001112043857574, + "learning_rate": 0.0002615702479338843, + "loss": 0.6465, + "step": 165 + }, + { + "epoch": 0.27, + "grad_norm": 0.1998877376317978, + "learning_rate": 0.00026132231404958677, + "loss": 0.666, + "step": 166 + }, + { + "epoch": 0.27, + "grad_norm": 0.23009613156318665, + "learning_rate": 0.00026107438016528925, + "loss": 0.8793, + "step": 167 + }, + { + "epoch": 0.28, + "grad_norm": 0.24525685608386993, + "learning_rate": 0.0002608264462809917, + "loss": 0.8009, + "step": 168 + }, + { + "epoch": 0.28, + "grad_norm": 0.21605077385902405, + "learning_rate": 0.00026057851239669416, + "loss": 0.5459, + "step": 169 + }, + { + "epoch": 0.28, + "grad_norm": 0.2576725482940674, + "learning_rate": 0.00026033057851239665, + "loss": 0.6818, + "step": 170 + }, + { + "epoch": 0.28, + "grad_norm": 0.23385170102119446, + "learning_rate": 0.00026008264462809913, + "loss": 0.7559, + "step": 171 + }, + { + "epoch": 0.28, + "grad_norm": 0.1973017454147339, + "learning_rate": 0.00025983471074380167, + "loss": 0.6798, + "step": 172 + }, + { + "epoch": 0.28, + "grad_norm": 0.22262559831142426, + "learning_rate": 0.0002595867768595041, + "loss": 0.5566, + "step": 173 + }, + { + "epoch": 0.28, + "grad_norm": 0.23010462522506714, + "learning_rate": 0.0002593388429752066, + "loss": 0.7101, + "step": 174 + }, + { + "epoch": 0.29, + "grad_norm": 0.21676452457904816, + "learning_rate": 0.00025909090909090907, + "loss": 0.7038, + "step": 175 + }, + { + "epoch": 0.29, + "grad_norm": 0.22475261986255646, + "learning_rate": 0.00025884297520661155, + "loss": 0.7812, + "step": 176 + }, + { + "epoch": 0.29, + "grad_norm": 0.28893202543258667, + "learning_rate": 0.00025859504132231403, + "loss": 0.5925, + "step": 177 + }, + { + "epoch": 0.29, + "grad_norm": 0.22777552902698517, + "learning_rate": 0.0002583471074380165, + "loss": 0.7319, + "step": 178 + }, + { + "epoch": 0.29, + "grad_norm": 0.2287953644990921, + "learning_rate": 0.000258099173553719, + "loss": 0.7775, + "step": 179 + }, + { + "epoch": 0.29, + "grad_norm": 0.2049843668937683, + "learning_rate": 0.0002578512396694215, + "loss": 0.7448, + "step": 180 + }, + { + "epoch": 0.3, + "grad_norm": 0.22585280239582062, + "learning_rate": 0.00025760330578512397, + "loss": 0.59, + "step": 181 + }, + { + "epoch": 0.3, + "grad_norm": 0.23159150779247284, + "learning_rate": 0.0002573553719008264, + "loss": 0.737, + "step": 182 + }, + { + "epoch": 0.3, + "grad_norm": 0.3393082320690155, + "learning_rate": 0.0002571074380165289, + "loss": 0.6948, + "step": 183 + }, + { + "epoch": 0.3, + "grad_norm": 0.2345617413520813, + "learning_rate": 0.00025685950413223136, + "loss": 0.6351, + "step": 184 + }, + { + "epoch": 0.3, + "grad_norm": 0.23474591970443726, + "learning_rate": 0.00025661157024793385, + "loss": 0.6643, + "step": 185 + }, + { + "epoch": 0.3, + "grad_norm": 0.2473030984401703, + "learning_rate": 0.00025636363636363633, + "loss": 0.7663, + "step": 186 + }, + { + "epoch": 0.31, + "grad_norm": 0.2971685230731964, + "learning_rate": 0.0002561157024793388, + "loss": 0.7449, + "step": 187 + }, + { + "epoch": 0.31, + "grad_norm": 0.2745087742805481, + "learning_rate": 0.0002558677685950413, + "loss": 0.6125, + "step": 188 + }, + { + "epoch": 0.31, + "grad_norm": 0.23520545661449432, + "learning_rate": 0.0002556198347107438, + "loss": 0.573, + "step": 189 + }, + { + "epoch": 0.31, + "grad_norm": 0.2955464720726013, + "learning_rate": 0.00025537190082644627, + "loss": 0.5315, + "step": 190 + }, + { + "epoch": 0.31, + "grad_norm": 0.23987281322479248, + "learning_rate": 0.00025512396694214875, + "loss": 0.5636, + "step": 191 + }, + { + "epoch": 0.31, + "grad_norm": 0.24263744056224823, + "learning_rate": 0.00025487603305785123, + "loss": 0.6047, + "step": 192 + }, + { + "epoch": 0.32, + "grad_norm": 0.26061922311782837, + "learning_rate": 0.0002546280991735537, + "loss": 0.7812, + "step": 193 + }, + { + "epoch": 0.32, + "grad_norm": 0.2458687126636505, + "learning_rate": 0.0002543801652892562, + "loss": 0.58, + "step": 194 + }, + { + "epoch": 0.32, + "grad_norm": 0.24598994851112366, + "learning_rate": 0.00025413223140495863, + "loss": 0.7432, + "step": 195 + }, + { + "epoch": 0.32, + "grad_norm": 0.248992919921875, + "learning_rate": 0.0002538842975206611, + "loss": 0.6953, + "step": 196 + }, + { + "epoch": 0.32, + "grad_norm": 0.2518531382083893, + "learning_rate": 0.0002536363636363636, + "loss": 0.6707, + "step": 197 + }, + { + "epoch": 0.32, + "grad_norm": 0.23844210803508759, + "learning_rate": 0.0002533884297520661, + "loss": 0.6285, + "step": 198 + }, + { + "epoch": 0.33, + "grad_norm": 0.21948237717151642, + "learning_rate": 0.00025314049586776856, + "loss": 0.6859, + "step": 199 + }, + { + "epoch": 0.33, + "grad_norm": 0.2003835141658783, + "learning_rate": 0.00025289256198347105, + "loss": 0.6305, + "step": 200 + }, + { + "epoch": 0.33, + "grad_norm": 0.23421582579612732, + "learning_rate": 0.00025264462809917353, + "loss": 0.7164, + "step": 201 + }, + { + "epoch": 0.33, + "grad_norm": 0.22344104945659637, + "learning_rate": 0.000252396694214876, + "loss": 0.6498, + "step": 202 + }, + { + "epoch": 0.33, + "grad_norm": 0.17792212963104248, + "learning_rate": 0.0002521487603305785, + "loss": 0.614, + "step": 203 + }, + { + "epoch": 0.33, + "grad_norm": 0.217886820435524, + "learning_rate": 0.000251900826446281, + "loss": 0.7033, + "step": 204 + }, + { + "epoch": 0.34, + "grad_norm": 0.209726020693779, + "learning_rate": 0.00025165289256198347, + "loss": 0.5913, + "step": 205 + }, + { + "epoch": 0.34, + "grad_norm": 0.2401910424232483, + "learning_rate": 0.00025140495867768595, + "loss": 0.6405, + "step": 206 + }, + { + "epoch": 0.34, + "grad_norm": 0.21315626800060272, + "learning_rate": 0.00025115702479338843, + "loss": 0.7369, + "step": 207 + }, + { + "epoch": 0.34, + "grad_norm": 0.20102320611476898, + "learning_rate": 0.00025090909090909086, + "loss": 0.6245, + "step": 208 + }, + { + "epoch": 0.34, + "grad_norm": 0.20447981357574463, + "learning_rate": 0.00025066115702479335, + "loss": 0.5423, + "step": 209 + }, + { + "epoch": 0.34, + "grad_norm": 0.24979281425476074, + "learning_rate": 0.00025041322314049583, + "loss": 0.8078, + "step": 210 + }, + { + "epoch": 0.35, + "grad_norm": 0.20141547918319702, + "learning_rate": 0.0002501652892561983, + "loss": 0.7386, + "step": 211 + }, + { + "epoch": 0.35, + "grad_norm": 0.2538990378379822, + "learning_rate": 0.0002499173553719008, + "loss": 0.7219, + "step": 212 + }, + { + "epoch": 0.35, + "grad_norm": 0.2613961100578308, + "learning_rate": 0.0002496694214876033, + "loss": 0.7903, + "step": 213 + }, + { + "epoch": 0.35, + "grad_norm": 0.24777857959270477, + "learning_rate": 0.00024942148760330576, + "loss": 0.664, + "step": 214 + }, + { + "epoch": 0.35, + "grad_norm": 0.21958425641059875, + "learning_rate": 0.00024917355371900825, + "loss": 0.6755, + "step": 215 + }, + { + "epoch": 0.35, + "grad_norm": 0.2218528538942337, + "learning_rate": 0.00024892561983471073, + "loss": 0.5568, + "step": 216 + }, + { + "epoch": 0.36, + "grad_norm": 0.23632755875587463, + "learning_rate": 0.00024867768595041316, + "loss": 0.6858, + "step": 217 + }, + { + "epoch": 0.36, + "grad_norm": 0.2641279697418213, + "learning_rate": 0.0002484297520661157, + "loss": 0.7783, + "step": 218 + }, + { + "epoch": 0.36, + "grad_norm": 0.3147680163383484, + "learning_rate": 0.0002481818181818182, + "loss": 0.662, + "step": 219 + }, + { + "epoch": 0.36, + "grad_norm": 0.27947697043418884, + "learning_rate": 0.00024793388429752067, + "loss": 0.6477, + "step": 220 + }, + { + "epoch": 0.36, + "grad_norm": 0.2297278195619583, + "learning_rate": 0.00024768595041322315, + "loss": 0.5895, + "step": 221 + }, + { + "epoch": 0.36, + "grad_norm": 0.23085851967334747, + "learning_rate": 0.0002474380165289256, + "loss": 0.5806, + "step": 222 + }, + { + "epoch": 0.37, + "grad_norm": 0.19654251635074615, + "learning_rate": 0.00024719008264462806, + "loss": 0.5942, + "step": 223 + }, + { + "epoch": 0.37, + "grad_norm": 0.2467166632413864, + "learning_rate": 0.00024694214876033055, + "loss": 0.5059, + "step": 224 + }, + { + "epoch": 0.37, + "grad_norm": 0.22614917159080505, + "learning_rate": 0.00024669421487603303, + "loss": 0.643, + "step": 225 + }, + { + "epoch": 0.37, + "grad_norm": 0.2622920274734497, + "learning_rate": 0.0002464462809917355, + "loss": 0.6257, + "step": 226 + }, + { + "epoch": 0.37, + "grad_norm": 0.21843163669109344, + "learning_rate": 0.000246198347107438, + "loss": 0.6057, + "step": 227 + }, + { + "epoch": 0.37, + "grad_norm": 0.2294640988111496, + "learning_rate": 0.0002459504132231405, + "loss": 0.6876, + "step": 228 + }, + { + "epoch": 0.37, + "grad_norm": 0.1791463941335678, + "learning_rate": 0.00024570247933884296, + "loss": 0.5348, + "step": 229 + }, + { + "epoch": 0.38, + "grad_norm": 0.17243699729442596, + "learning_rate": 0.00024545454545454545, + "loss": 0.5966, + "step": 230 + }, + { + "epoch": 0.38, + "grad_norm": 0.22769273817539215, + "learning_rate": 0.0002452066115702479, + "loss": 0.7912, + "step": 231 + }, + { + "epoch": 0.38, + "grad_norm": 0.2325255423784256, + "learning_rate": 0.0002449586776859504, + "loss": 0.7441, + "step": 232 + }, + { + "epoch": 0.38, + "grad_norm": 0.24277740716934204, + "learning_rate": 0.0002447107438016529, + "loss": 0.6653, + "step": 233 + }, + { + "epoch": 0.38, + "grad_norm": 0.21596141159534454, + "learning_rate": 0.0002444628099173554, + "loss": 0.6668, + "step": 234 + }, + { + "epoch": 0.38, + "grad_norm": 0.20814135670661926, + "learning_rate": 0.0002442148760330578, + "loss": 0.6306, + "step": 235 + }, + { + "epoch": 0.39, + "grad_norm": 0.25570017099380493, + "learning_rate": 0.0002439669421487603, + "loss": 0.6524, + "step": 236 + }, + { + "epoch": 0.39, + "grad_norm": 0.2502390146255493, + "learning_rate": 0.00024371900826446278, + "loss": 0.6048, + "step": 237 + }, + { + "epoch": 0.39, + "grad_norm": 0.23688243329524994, + "learning_rate": 0.0002434710743801653, + "loss": 0.568, + "step": 238 + }, + { + "epoch": 0.39, + "grad_norm": 0.21041709184646606, + "learning_rate": 0.00024322314049586777, + "loss": 0.6908, + "step": 239 + }, + { + "epoch": 0.39, + "grad_norm": 0.21656759083271027, + "learning_rate": 0.00024297520661157023, + "loss": 0.4993, + "step": 240 + }, + { + "epoch": 0.39, + "grad_norm": 0.25133028626441956, + "learning_rate": 0.0002427272727272727, + "loss": 0.718, + "step": 241 + }, + { + "epoch": 0.4, + "grad_norm": 0.22228790819644928, + "learning_rate": 0.0002424793388429752, + "loss": 0.6146, + "step": 242 + }, + { + "epoch": 0.4, + "grad_norm": 0.26273205876350403, + "learning_rate": 0.00024223140495867768, + "loss": 0.7459, + "step": 243 + }, + { + "epoch": 0.4, + "grad_norm": 0.2156606763601303, + "learning_rate": 0.00024198347107438014, + "loss": 0.6692, + "step": 244 + }, + { + "epoch": 0.4, + "grad_norm": 0.2075020670890808, + "learning_rate": 0.00024173553719008262, + "loss": 0.6427, + "step": 245 + }, + { + "epoch": 0.4, + "grad_norm": 0.25821176171302795, + "learning_rate": 0.0002414876033057851, + "loss": 0.7964, + "step": 246 + }, + { + "epoch": 0.4, + "grad_norm": 0.23016126453876495, + "learning_rate": 0.0002412396694214876, + "loss": 0.536, + "step": 247 + }, + { + "epoch": 0.41, + "grad_norm": 0.23115016520023346, + "learning_rate": 0.00024099173553719004, + "loss": 0.6053, + "step": 248 + }, + { + "epoch": 0.41, + "grad_norm": 0.18249157071113586, + "learning_rate": 0.00024074380165289253, + "loss": 0.6574, + "step": 249 + }, + { + "epoch": 0.41, + "grad_norm": 0.28391778469085693, + "learning_rate": 0.000240495867768595, + "loss": 0.7152, + "step": 250 + }, + { + "epoch": 0.41, + "grad_norm": 0.2581539452075958, + "learning_rate": 0.0002402479338842975, + "loss": 0.8476, + "step": 251 + }, + { + "epoch": 0.41, + "grad_norm": 0.2304867058992386, + "learning_rate": 0.00023999999999999998, + "loss": 0.5781, + "step": 252 + }, + { + "epoch": 0.41, + "grad_norm": 0.239717036485672, + "learning_rate": 0.00023975206611570244, + "loss": 0.6543, + "step": 253 + }, + { + "epoch": 0.42, + "grad_norm": 0.22493794560432434, + "learning_rate": 0.00023950413223140495, + "loss": 0.7048, + "step": 254 + }, + { + "epoch": 0.42, + "grad_norm": 0.22085991501808167, + "learning_rate": 0.00023925619834710743, + "loss": 0.5572, + "step": 255 + }, + { + "epoch": 0.42, + "grad_norm": 0.35917988419532776, + "learning_rate": 0.0002390082644628099, + "loss": 0.8485, + "step": 256 + }, + { + "epoch": 0.42, + "grad_norm": 0.28269943594932556, + "learning_rate": 0.00023876033057851237, + "loss": 0.5732, + "step": 257 + }, + { + "epoch": 0.42, + "grad_norm": 0.26313093304634094, + "learning_rate": 0.00023851239669421485, + "loss": 0.8212, + "step": 258 + }, + { + "epoch": 0.42, + "grad_norm": 0.30286532640457153, + "learning_rate": 0.00023826446280991734, + "loss": 0.5878, + "step": 259 + }, + { + "epoch": 0.43, + "grad_norm": 0.22270837426185608, + "learning_rate": 0.00023801652892561982, + "loss": 0.6933, + "step": 260 + }, + { + "epoch": 0.43, + "grad_norm": 0.29011014103889465, + "learning_rate": 0.0002377685950413223, + "loss": 0.6188, + "step": 261 + }, + { + "epoch": 0.43, + "grad_norm": 0.2390982061624527, + "learning_rate": 0.00023752066115702476, + "loss": 0.6426, + "step": 262 + }, + { + "epoch": 0.43, + "grad_norm": 0.3416346609592438, + "learning_rate": 0.00023727272727272724, + "loss": 0.8845, + "step": 263 + }, + { + "epoch": 0.43, + "grad_norm": 0.25051388144493103, + "learning_rate": 0.00023702479338842973, + "loss": 0.7286, + "step": 264 + }, + { + "epoch": 0.43, + "grad_norm": 0.2497546523809433, + "learning_rate": 0.0002367768595041322, + "loss": 0.6027, + "step": 265 + }, + { + "epoch": 0.44, + "grad_norm": 0.23835037648677826, + "learning_rate": 0.00023652892561983467, + "loss": 0.7052, + "step": 266 + }, + { + "epoch": 0.44, + "grad_norm": 0.22467398643493652, + "learning_rate": 0.00023628099173553715, + "loss": 0.5806, + "step": 267 + }, + { + "epoch": 0.44, + "grad_norm": 0.2663390338420868, + "learning_rate": 0.00023603305785123964, + "loss": 0.6943, + "step": 268 + }, + { + "epoch": 0.44, + "grad_norm": 0.22997191548347473, + "learning_rate": 0.00023578512396694215, + "loss": 0.6411, + "step": 269 + }, + { + "epoch": 0.44, + "grad_norm": 0.23266558349132538, + "learning_rate": 0.00023553719008264463, + "loss": 0.6068, + "step": 270 + }, + { + "epoch": 0.44, + "grad_norm": 0.2304474264383316, + "learning_rate": 0.00023528925619834709, + "loss": 0.6427, + "step": 271 + }, + { + "epoch": 0.45, + "grad_norm": 0.28231826424598694, + "learning_rate": 0.00023504132231404957, + "loss": 0.8011, + "step": 272 + }, + { + "epoch": 0.45, + "grad_norm": 0.28013259172439575, + "learning_rate": 0.00023479338842975205, + "loss": 0.5988, + "step": 273 + }, + { + "epoch": 0.45, + "grad_norm": 0.22702372074127197, + "learning_rate": 0.00023454545454545454, + "loss": 0.6737, + "step": 274 + }, + { + "epoch": 0.45, + "grad_norm": 0.27958643436431885, + "learning_rate": 0.000234297520661157, + "loss": 0.6621, + "step": 275 + }, + { + "epoch": 0.45, + "grad_norm": 0.23902451992034912, + "learning_rate": 0.00023404958677685948, + "loss": 0.6525, + "step": 276 + }, + { + "epoch": 0.45, + "grad_norm": 0.2778523564338684, + "learning_rate": 0.00023380165289256196, + "loss": 0.6697, + "step": 277 + }, + { + "epoch": 0.46, + "grad_norm": 0.2382276952266693, + "learning_rate": 0.00023355371900826444, + "loss": 0.6281, + "step": 278 + }, + { + "epoch": 0.46, + "grad_norm": 0.24487091600894928, + "learning_rate": 0.00023330578512396693, + "loss": 0.6842, + "step": 279 + }, + { + "epoch": 0.46, + "grad_norm": 0.2063397765159607, + "learning_rate": 0.00023305785123966938, + "loss": 0.6554, + "step": 280 + }, + { + "epoch": 0.46, + "grad_norm": 0.21523278951644897, + "learning_rate": 0.00023280991735537187, + "loss": 0.632, + "step": 281 + }, + { + "epoch": 0.46, + "grad_norm": 0.2420080006122589, + "learning_rate": 0.00023256198347107435, + "loss": 0.6001, + "step": 282 + }, + { + "epoch": 0.46, + "grad_norm": 0.2390110194683075, + "learning_rate": 0.00023231404958677686, + "loss": 0.5648, + "step": 283 + }, + { + "epoch": 0.47, + "grad_norm": 0.24080687761306763, + "learning_rate": 0.0002320661157024793, + "loss": 0.86, + "step": 284 + }, + { + "epoch": 0.47, + "grad_norm": 0.29456445574760437, + "learning_rate": 0.0002318181818181818, + "loss": 0.7418, + "step": 285 + }, + { + "epoch": 0.47, + "grad_norm": 0.23326683044433594, + "learning_rate": 0.00023157024793388429, + "loss": 0.6967, + "step": 286 + }, + { + "epoch": 0.47, + "grad_norm": 0.20866093039512634, + "learning_rate": 0.00023132231404958677, + "loss": 0.5205, + "step": 287 + }, + { + "epoch": 0.47, + "grad_norm": 0.3158474266529083, + "learning_rate": 0.00023107438016528925, + "loss": 0.7879, + "step": 288 + }, + { + "epoch": 0.47, + "grad_norm": 0.2730140686035156, + "learning_rate": 0.0002308264462809917, + "loss": 0.7292, + "step": 289 + }, + { + "epoch": 0.47, + "grad_norm": 0.25384965538978577, + "learning_rate": 0.0002305785123966942, + "loss": 0.7258, + "step": 290 + }, + { + "epoch": 0.48, + "grad_norm": 0.20765069127082825, + "learning_rate": 0.00023033057851239668, + "loss": 0.7108, + "step": 291 + }, + { + "epoch": 0.48, + "grad_norm": 0.25662195682525635, + "learning_rate": 0.00023008264462809916, + "loss": 0.7473, + "step": 292 + }, + { + "epoch": 0.48, + "grad_norm": 0.300243616104126, + "learning_rate": 0.00022983471074380162, + "loss": 0.6902, + "step": 293 + }, + { + "epoch": 0.48, + "grad_norm": 0.23513919115066528, + "learning_rate": 0.0002295867768595041, + "loss": 0.5888, + "step": 294 + }, + { + "epoch": 0.48, + "grad_norm": 0.2077571451663971, + "learning_rate": 0.00022933884297520658, + "loss": 0.6256, + "step": 295 + }, + { + "epoch": 0.48, + "grad_norm": 0.266201376914978, + "learning_rate": 0.00022909090909090907, + "loss": 0.6913, + "step": 296 + }, + { + "epoch": 0.49, + "grad_norm": 0.2239614725112915, + "learning_rate": 0.00022884297520661152, + "loss": 0.7369, + "step": 297 + }, + { + "epoch": 0.49, + "grad_norm": 0.21509824693202972, + "learning_rate": 0.000228595041322314, + "loss": 0.4445, + "step": 298 + }, + { + "epoch": 0.49, + "grad_norm": 0.21956239640712738, + "learning_rate": 0.00022834710743801652, + "loss": 0.6732, + "step": 299 + }, + { + "epoch": 0.49, + "grad_norm": 0.18832357227802277, + "learning_rate": 0.000228099173553719, + "loss": 0.6808, + "step": 300 + }, + { + "epoch": 0.49, + "grad_norm": 0.21115505695343018, + "learning_rate": 0.0002278512396694215, + "loss": 0.5323, + "step": 301 + }, + { + "epoch": 0.49, + "grad_norm": 0.23715418577194214, + "learning_rate": 0.00022760330578512394, + "loss": 0.8333, + "step": 302 + }, + { + "epoch": 0.5, + "grad_norm": 0.29385048151016235, + "learning_rate": 0.00022735537190082643, + "loss": 0.6, + "step": 303 + }, + { + "epoch": 0.5, + "grad_norm": 0.26947689056396484, + "learning_rate": 0.0002271074380165289, + "loss": 0.8788, + "step": 304 + }, + { + "epoch": 0.5, + "grad_norm": 0.2778269946575165, + "learning_rate": 0.0002268595041322314, + "loss": 0.7073, + "step": 305 + }, + { + "epoch": 0.5, + "grad_norm": 0.20938479900360107, + "learning_rate": 0.00022661157024793385, + "loss": 0.6422, + "step": 306 + }, + { + "epoch": 0.5, + "grad_norm": 0.2777106761932373, + "learning_rate": 0.00022636363636363633, + "loss": 0.7495, + "step": 307 + }, + { + "epoch": 0.5, + "grad_norm": 0.20872819423675537, + "learning_rate": 0.00022611570247933882, + "loss": 0.6492, + "step": 308 + }, + { + "epoch": 0.51, + "grad_norm": 0.2752722501754761, + "learning_rate": 0.0002258677685950413, + "loss": 0.6014, + "step": 309 + }, + { + "epoch": 0.51, + "grad_norm": 0.24615786969661713, + "learning_rate": 0.00022561983471074378, + "loss": 0.6287, + "step": 310 + }, + { + "epoch": 0.51, + "grad_norm": 0.24146385490894318, + "learning_rate": 0.00022537190082644624, + "loss": 0.6151, + "step": 311 + }, + { + "epoch": 0.51, + "grad_norm": 0.24762235581874847, + "learning_rate": 0.00022512396694214872, + "loss": 0.6377, + "step": 312 + }, + { + "epoch": 0.51, + "grad_norm": 0.24630331993103027, + "learning_rate": 0.00022487603305785124, + "loss": 0.7255, + "step": 313 + }, + { + "epoch": 0.51, + "grad_norm": 0.2922554612159729, + "learning_rate": 0.00022462809917355372, + "loss": 0.6645, + "step": 314 + }, + { + "epoch": 0.52, + "grad_norm": 0.21686063706874847, + "learning_rate": 0.00022438016528925618, + "loss": 0.5606, + "step": 315 + }, + { + "epoch": 0.52, + "grad_norm": 0.2216208428144455, + "learning_rate": 0.00022413223140495866, + "loss": 0.5126, + "step": 316 + }, + { + "epoch": 0.52, + "grad_norm": 0.25635436177253723, + "learning_rate": 0.00022388429752066114, + "loss": 0.7387, + "step": 317 + }, + { + "epoch": 0.52, + "grad_norm": 0.2786000669002533, + "learning_rate": 0.00022363636363636363, + "loss": 0.5941, + "step": 318 + }, + { + "epoch": 0.52, + "grad_norm": 0.26092806458473206, + "learning_rate": 0.0002233884297520661, + "loss": 0.7851, + "step": 319 + }, + { + "epoch": 0.52, + "grad_norm": 0.23881889879703522, + "learning_rate": 0.00022314049586776857, + "loss": 0.598, + "step": 320 + }, + { + "epoch": 0.53, + "grad_norm": 0.23304526507854462, + "learning_rate": 0.00022289256198347105, + "loss": 0.7165, + "step": 321 + }, + { + "epoch": 0.53, + "grad_norm": 0.2340225875377655, + "learning_rate": 0.00022264462809917353, + "loss": 0.6608, + "step": 322 + }, + { + "epoch": 0.53, + "grad_norm": 0.31176140904426575, + "learning_rate": 0.00022239669421487602, + "loss": 0.6711, + "step": 323 + }, + { + "epoch": 0.53, + "grad_norm": 0.23832640051841736, + "learning_rate": 0.00022214876033057847, + "loss": 0.732, + "step": 324 + }, + { + "epoch": 0.53, + "grad_norm": 0.28845977783203125, + "learning_rate": 0.00022190082644628096, + "loss": 0.7968, + "step": 325 + }, + { + "epoch": 0.53, + "grad_norm": 0.1978536993265152, + "learning_rate": 0.00022165289256198344, + "loss": 0.6592, + "step": 326 + }, + { + "epoch": 0.54, + "grad_norm": 0.26940053701400757, + "learning_rate": 0.00022140495867768595, + "loss": 0.7953, + "step": 327 + }, + { + "epoch": 0.54, + "grad_norm": 0.20393389463424683, + "learning_rate": 0.00022115702479338844, + "loss": 0.4871, + "step": 328 + }, + { + "epoch": 0.54, + "grad_norm": 0.27152347564697266, + "learning_rate": 0.0002209090909090909, + "loss": 0.5583, + "step": 329 + }, + { + "epoch": 0.54, + "grad_norm": 0.2883144021034241, + "learning_rate": 0.00022066115702479338, + "loss": 0.6156, + "step": 330 + }, + { + "epoch": 0.54, + "grad_norm": 0.1987351030111313, + "learning_rate": 0.00022041322314049586, + "loss": 0.5196, + "step": 331 + }, + { + "epoch": 0.54, + "grad_norm": 0.2651583254337311, + "learning_rate": 0.00022016528925619834, + "loss": 0.6099, + "step": 332 + }, + { + "epoch": 0.55, + "grad_norm": 0.2574511468410492, + "learning_rate": 0.0002199173553719008, + "loss": 0.6925, + "step": 333 + }, + { + "epoch": 0.55, + "grad_norm": 0.27730292081832886, + "learning_rate": 0.00021966942148760328, + "loss": 0.6752, + "step": 334 + }, + { + "epoch": 0.55, + "grad_norm": 0.2001207172870636, + "learning_rate": 0.00021942148760330577, + "loss": 0.75, + "step": 335 + }, + { + "epoch": 0.55, + "grad_norm": 0.24222363531589508, + "learning_rate": 0.00021917355371900825, + "loss": 0.6364, + "step": 336 + }, + { + "epoch": 0.55, + "grad_norm": 0.26326724886894226, + "learning_rate": 0.0002189256198347107, + "loss": 0.673, + "step": 337 + }, + { + "epoch": 0.55, + "grad_norm": 0.2272881418466568, + "learning_rate": 0.0002186776859504132, + "loss": 0.561, + "step": 338 + }, + { + "epoch": 0.56, + "grad_norm": 0.24880024790763855, + "learning_rate": 0.00021842975206611567, + "loss": 0.5552, + "step": 339 + }, + { + "epoch": 0.56, + "grad_norm": 0.2593706548213959, + "learning_rate": 0.00021818181818181816, + "loss": 0.5417, + "step": 340 + }, + { + "epoch": 0.56, + "grad_norm": 0.19063642621040344, + "learning_rate": 0.00021793388429752067, + "loss": 0.5694, + "step": 341 + }, + { + "epoch": 0.56, + "grad_norm": 0.2146475464105606, + "learning_rate": 0.0002176859504132231, + "loss": 0.4314, + "step": 342 + }, + { + "epoch": 0.56, + "grad_norm": 0.25150927901268005, + "learning_rate": 0.0002174380165289256, + "loss": 0.631, + "step": 343 + }, + { + "epoch": 0.56, + "grad_norm": 0.2753889858722687, + "learning_rate": 0.0002171900826446281, + "loss": 0.6859, + "step": 344 + }, + { + "epoch": 0.56, + "grad_norm": 0.20773079991340637, + "learning_rate": 0.00021694214876033058, + "loss": 0.7515, + "step": 345 + }, + { + "epoch": 0.57, + "grad_norm": 0.2547062635421753, + "learning_rate": 0.00021669421487603303, + "loss": 0.7582, + "step": 346 + }, + { + "epoch": 0.57, + "grad_norm": 0.24687208235263824, + "learning_rate": 0.00021644628099173552, + "loss": 0.5865, + "step": 347 + }, + { + "epoch": 0.57, + "grad_norm": 0.24116279184818268, + "learning_rate": 0.000216198347107438, + "loss": 0.4841, + "step": 348 + }, + { + "epoch": 0.57, + "grad_norm": 0.2270282804965973, + "learning_rate": 0.00021595041322314048, + "loss": 0.5933, + "step": 349 + }, + { + "epoch": 0.57, + "grad_norm": 0.21436922252178192, + "learning_rate": 0.00021570247933884297, + "loss": 0.6959, + "step": 350 + }, + { + "epoch": 0.57, + "grad_norm": 0.25802701711654663, + "learning_rate": 0.00021545454545454542, + "loss": 0.729, + "step": 351 + }, + { + "epoch": 0.58, + "grad_norm": 0.23808260262012482, + "learning_rate": 0.0002152066115702479, + "loss": 0.6346, + "step": 352 + }, + { + "epoch": 0.58, + "grad_norm": 0.23161651194095612, + "learning_rate": 0.0002149586776859504, + "loss": 0.6459, + "step": 353 + }, + { + "epoch": 0.58, + "grad_norm": 0.2442287802696228, + "learning_rate": 0.00021471074380165287, + "loss": 0.6803, + "step": 354 + }, + { + "epoch": 0.58, + "grad_norm": 0.19150683283805847, + "learning_rate": 0.00021446280991735533, + "loss": 0.4375, + "step": 355 + }, + { + "epoch": 0.58, + "grad_norm": 0.23142127692699432, + "learning_rate": 0.00021421487603305781, + "loss": 0.5505, + "step": 356 + }, + { + "epoch": 0.58, + "grad_norm": 0.22447548806667328, + "learning_rate": 0.00021396694214876033, + "loss": 0.6368, + "step": 357 + }, + { + "epoch": 0.59, + "grad_norm": 0.25168758630752563, + "learning_rate": 0.0002137190082644628, + "loss": 0.6322, + "step": 358 + }, + { + "epoch": 0.59, + "grad_norm": 0.25538235902786255, + "learning_rate": 0.0002134710743801653, + "loss": 0.5317, + "step": 359 + }, + { + "epoch": 0.59, + "grad_norm": 0.2565425634384155, + "learning_rate": 0.00021322314049586775, + "loss": 0.6261, + "step": 360 + }, + { + "epoch": 0.59, + "grad_norm": 0.25399863719940186, + "learning_rate": 0.00021297520661157023, + "loss": 0.596, + "step": 361 + }, + { + "epoch": 0.59, + "grad_norm": 0.27143988013267517, + "learning_rate": 0.00021272727272727272, + "loss": 0.6691, + "step": 362 + }, + { + "epoch": 0.59, + "grad_norm": 0.2387736439704895, + "learning_rate": 0.0002124793388429752, + "loss": 0.5288, + "step": 363 + }, + { + "epoch": 0.6, + "grad_norm": 0.2549780607223511, + "learning_rate": 0.00021223140495867766, + "loss": 0.7455, + "step": 364 + }, + { + "epoch": 0.6, + "grad_norm": 0.2740858793258667, + "learning_rate": 0.00021198347107438014, + "loss": 0.4921, + "step": 365 + }, + { + "epoch": 0.6, + "grad_norm": 0.25273847579956055, + "learning_rate": 0.00021173553719008262, + "loss": 0.7965, + "step": 366 + }, + { + "epoch": 0.6, + "grad_norm": 0.25858959555625916, + "learning_rate": 0.0002114876033057851, + "loss": 0.7303, + "step": 367 + }, + { + "epoch": 0.6, + "grad_norm": 0.2599296271800995, + "learning_rate": 0.0002112396694214876, + "loss": 0.6342, + "step": 368 + }, + { + "epoch": 0.6, + "grad_norm": 0.21084599196910858, + "learning_rate": 0.00021099173553719005, + "loss": 0.633, + "step": 369 + }, + { + "epoch": 0.61, + "grad_norm": 0.24272632598876953, + "learning_rate": 0.00021074380165289253, + "loss": 0.6213, + "step": 370 + }, + { + "epoch": 0.61, + "grad_norm": 0.26323699951171875, + "learning_rate": 0.00021049586776859501, + "loss": 0.563, + "step": 371 + }, + { + "epoch": 0.61, + "grad_norm": 0.20646587014198303, + "learning_rate": 0.00021024793388429753, + "loss": 0.6248, + "step": 372 + }, + { + "epoch": 0.61, + "grad_norm": 0.21778297424316406, + "learning_rate": 0.00020999999999999998, + "loss": 0.7186, + "step": 373 + }, + { + "epoch": 0.61, + "grad_norm": 0.21315112709999084, + "learning_rate": 0.00020975206611570247, + "loss": 0.5961, + "step": 374 + }, + { + "epoch": 0.61, + "grad_norm": 0.20787106454372406, + "learning_rate": 0.00020950413223140495, + "loss": 0.5917, + "step": 375 + }, + { + "epoch": 0.62, + "grad_norm": 0.23541009426116943, + "learning_rate": 0.00020925619834710743, + "loss": 0.7803, + "step": 376 + }, + { + "epoch": 0.62, + "grad_norm": 0.22649626433849335, + "learning_rate": 0.00020900826446280992, + "loss": 0.5895, + "step": 377 + }, + { + "epoch": 0.62, + "grad_norm": 0.23644742369651794, + "learning_rate": 0.00020876033057851237, + "loss": 0.6656, + "step": 378 + }, + { + "epoch": 0.62, + "grad_norm": 0.22934262454509735, + "learning_rate": 0.00020851239669421486, + "loss": 0.5933, + "step": 379 + }, + { + "epoch": 0.62, + "grad_norm": 0.289989709854126, + "learning_rate": 0.00020826446280991734, + "loss": 0.6852, + "step": 380 + }, + { + "epoch": 0.62, + "grad_norm": 0.24489325284957886, + "learning_rate": 0.00020801652892561982, + "loss": 0.5546, + "step": 381 + }, + { + "epoch": 0.63, + "grad_norm": 0.27165278792381287, + "learning_rate": 0.00020776859504132228, + "loss": 0.6845, + "step": 382 + }, + { + "epoch": 0.63, + "grad_norm": 0.19467370212078094, + "learning_rate": 0.00020752066115702476, + "loss": 0.5587, + "step": 383 + }, + { + "epoch": 0.63, + "grad_norm": 0.27320200204849243, + "learning_rate": 0.00020727272727272725, + "loss": 0.7144, + "step": 384 + }, + { + "epoch": 0.63, + "grad_norm": 0.28100526332855225, + "learning_rate": 0.00020702479338842973, + "loss": 0.6914, + "step": 385 + }, + { + "epoch": 0.63, + "grad_norm": 0.3059975504875183, + "learning_rate": 0.0002067768595041322, + "loss": 0.6075, + "step": 386 + }, + { + "epoch": 0.63, + "grad_norm": 0.24904222786426544, + "learning_rate": 0.00020652892561983467, + "loss": 0.5543, + "step": 387 + }, + { + "epoch": 0.64, + "grad_norm": 0.24768255650997162, + "learning_rate": 0.00020628099173553718, + "loss": 0.607, + "step": 388 + }, + { + "epoch": 0.64, + "grad_norm": 0.25083738565444946, + "learning_rate": 0.00020603305785123967, + "loss": 0.7961, + "step": 389 + }, + { + "epoch": 0.64, + "grad_norm": 0.26338303089141846, + "learning_rate": 0.00020578512396694215, + "loss": 0.6467, + "step": 390 + }, + { + "epoch": 0.64, + "grad_norm": 0.25761598348617554, + "learning_rate": 0.0002055371900826446, + "loss": 0.5891, + "step": 391 + }, + { + "epoch": 0.64, + "grad_norm": 0.2616937756538391, + "learning_rate": 0.0002052892561983471, + "loss": 0.5706, + "step": 392 + }, + { + "epoch": 0.64, + "grad_norm": 0.18980839848518372, + "learning_rate": 0.00020504132231404957, + "loss": 0.4479, + "step": 393 + }, + { + "epoch": 0.65, + "grad_norm": 0.250431627035141, + "learning_rate": 0.00020479338842975206, + "loss": 0.6006, + "step": 394 + }, + { + "epoch": 0.65, + "grad_norm": 0.2146655172109604, + "learning_rate": 0.0002045454545454545, + "loss": 0.7113, + "step": 395 + }, + { + "epoch": 0.65, + "grad_norm": 0.2195209115743637, + "learning_rate": 0.000204297520661157, + "loss": 0.5354, + "step": 396 + }, + { + "epoch": 0.65, + "grad_norm": 0.24879257380962372, + "learning_rate": 0.00020404958677685948, + "loss": 0.5478, + "step": 397 + }, + { + "epoch": 0.65, + "grad_norm": 0.27159082889556885, + "learning_rate": 0.00020380165289256196, + "loss": 0.7681, + "step": 398 + }, + { + "epoch": 0.65, + "grad_norm": 0.20614947378635406, + "learning_rate": 0.00020355371900826445, + "loss": 0.6357, + "step": 399 + }, + { + "epoch": 0.65, + "grad_norm": 0.25690051913261414, + "learning_rate": 0.0002033057851239669, + "loss": 0.5731, + "step": 400 + }, + { + "epoch": 0.66, + "grad_norm": 0.24473583698272705, + "learning_rate": 0.0002030578512396694, + "loss": 0.6784, + "step": 401 + }, + { + "epoch": 0.66, + "grad_norm": 0.32395297288894653, + "learning_rate": 0.0002028099173553719, + "loss": 0.7118, + "step": 402 + }, + { + "epoch": 0.66, + "grad_norm": 0.2975274324417114, + "learning_rate": 0.00020256198347107438, + "loss": 0.6504, + "step": 403 + }, + { + "epoch": 0.66, + "grad_norm": 0.2652553915977478, + "learning_rate": 0.00020231404958677684, + "loss": 0.6986, + "step": 404 + }, + { + "epoch": 0.66, + "grad_norm": 0.29475778341293335, + "learning_rate": 0.00020206611570247932, + "loss": 0.6525, + "step": 405 + }, + { + "epoch": 0.66, + "grad_norm": 0.24549973011016846, + "learning_rate": 0.0002018181818181818, + "loss": 0.5408, + "step": 406 + }, + { + "epoch": 0.67, + "grad_norm": 0.2181435376405716, + "learning_rate": 0.0002015702479338843, + "loss": 0.6146, + "step": 407 + }, + { + "epoch": 0.67, + "grad_norm": 0.2682584226131439, + "learning_rate": 0.00020132231404958677, + "loss": 0.6368, + "step": 408 + }, + { + "epoch": 0.67, + "grad_norm": 0.2641114592552185, + "learning_rate": 0.00020107438016528923, + "loss": 0.51, + "step": 409 + }, + { + "epoch": 0.67, + "grad_norm": 0.27871838212013245, + "learning_rate": 0.0002008264462809917, + "loss": 0.7269, + "step": 410 + }, + { + "epoch": 0.67, + "grad_norm": 0.23890569806098938, + "learning_rate": 0.0002005785123966942, + "loss": 0.6444, + "step": 411 + }, + { + "epoch": 0.67, + "grad_norm": 0.2451583445072174, + "learning_rate": 0.00020033057851239668, + "loss": 0.5806, + "step": 412 + }, + { + "epoch": 0.68, + "grad_norm": 0.2743864953517914, + "learning_rate": 0.00020008264462809914, + "loss": 0.6305, + "step": 413 + }, + { + "epoch": 0.68, + "grad_norm": 0.2626914978027344, + "learning_rate": 0.00019983471074380162, + "loss": 0.5765, + "step": 414 + }, + { + "epoch": 0.68, + "grad_norm": 0.2874875068664551, + "learning_rate": 0.0001995867768595041, + "loss": 0.5928, + "step": 415 + }, + { + "epoch": 0.68, + "grad_norm": 0.30499163269996643, + "learning_rate": 0.00019933884297520661, + "loss": 0.6271, + "step": 416 + }, + { + "epoch": 0.68, + "grad_norm": 0.30474454164505005, + "learning_rate": 0.0001990909090909091, + "loss": 0.6755, + "step": 417 + }, + { + "epoch": 0.68, + "grad_norm": 0.1819755882024765, + "learning_rate": 0.00019884297520661155, + "loss": 0.394, + "step": 418 + }, + { + "epoch": 0.69, + "grad_norm": 0.25470343232154846, + "learning_rate": 0.00019859504132231404, + "loss": 0.7121, + "step": 419 + }, + { + "epoch": 0.69, + "grad_norm": 0.26749151945114136, + "learning_rate": 0.00019834710743801652, + "loss": 0.6487, + "step": 420 + }, + { + "epoch": 0.69, + "grad_norm": 0.20643912255764008, + "learning_rate": 0.000198099173553719, + "loss": 0.4585, + "step": 421 + }, + { + "epoch": 0.69, + "grad_norm": 0.2576930522918701, + "learning_rate": 0.00019785123966942146, + "loss": 0.5235, + "step": 422 + }, + { + "epoch": 0.69, + "grad_norm": 0.2899012863636017, + "learning_rate": 0.00019760330578512395, + "loss": 0.6292, + "step": 423 + }, + { + "epoch": 0.69, + "grad_norm": 0.2541065216064453, + "learning_rate": 0.00019735537190082643, + "loss": 0.648, + "step": 424 + }, + { + "epoch": 0.7, + "grad_norm": 0.24382047355175018, + "learning_rate": 0.0001971074380165289, + "loss": 0.5939, + "step": 425 + }, + { + "epoch": 0.7, + "grad_norm": 0.22931940853595734, + "learning_rate": 0.00019685950413223137, + "loss": 0.6812, + "step": 426 + }, + { + "epoch": 0.7, + "grad_norm": 0.2592567205429077, + "learning_rate": 0.00019661157024793385, + "loss": 0.69, + "step": 427 + }, + { + "epoch": 0.7, + "grad_norm": 0.2516980767250061, + "learning_rate": 0.00019636363636363634, + "loss": 0.5707, + "step": 428 + }, + { + "epoch": 0.7, + "grad_norm": 0.23515059053897858, + "learning_rate": 0.00019611570247933882, + "loss": 0.6739, + "step": 429 + }, + { + "epoch": 0.7, + "grad_norm": 0.24742184579372406, + "learning_rate": 0.00019586776859504133, + "loss": 0.6761, + "step": 430 + }, + { + "epoch": 0.71, + "grad_norm": 0.26232922077178955, + "learning_rate": 0.00019561983471074376, + "loss": 0.7071, + "step": 431 + }, + { + "epoch": 0.71, + "grad_norm": 0.2853042781352997, + "learning_rate": 0.00019537190082644627, + "loss": 0.7667, + "step": 432 + }, + { + "epoch": 0.71, + "grad_norm": 0.251169353723526, + "learning_rate": 0.00019512396694214875, + "loss": 0.6518, + "step": 433 + }, + { + "epoch": 0.71, + "grad_norm": 0.2321665734052658, + "learning_rate": 0.00019487603305785124, + "loss": 0.4377, + "step": 434 + }, + { + "epoch": 0.71, + "grad_norm": 0.25216928124427795, + "learning_rate": 0.0001946280991735537, + "loss": 0.7173, + "step": 435 + }, + { + "epoch": 0.71, + "grad_norm": 0.19498330354690552, + "learning_rate": 0.00019438016528925618, + "loss": 0.5584, + "step": 436 + }, + { + "epoch": 0.72, + "grad_norm": 0.32786309719085693, + "learning_rate": 0.00019413223140495866, + "loss": 0.6583, + "step": 437 + }, + { + "epoch": 0.72, + "grad_norm": 0.25834760069847107, + "learning_rate": 0.00019388429752066115, + "loss": 0.4957, + "step": 438 + }, + { + "epoch": 0.72, + "grad_norm": 0.3462083041667938, + "learning_rate": 0.00019363636363636363, + "loss": 0.5205, + "step": 439 + }, + { + "epoch": 0.72, + "grad_norm": 0.27106693387031555, + "learning_rate": 0.00019338842975206609, + "loss": 0.6803, + "step": 440 + }, + { + "epoch": 0.72, + "grad_norm": 0.28165388107299805, + "learning_rate": 0.00019314049586776857, + "loss": 0.7049, + "step": 441 + }, + { + "epoch": 0.72, + "grad_norm": 0.20732273161411285, + "learning_rate": 0.00019289256198347105, + "loss": 0.6407, + "step": 442 + }, + { + "epoch": 0.73, + "grad_norm": 0.2609116733074188, + "learning_rate": 0.00019264462809917354, + "loss": 0.5377, + "step": 443 + }, + { + "epoch": 0.73, + "grad_norm": 0.2561998963356018, + "learning_rate": 0.000192396694214876, + "loss": 0.6212, + "step": 444 + }, + { + "epoch": 0.73, + "grad_norm": 0.27699044346809387, + "learning_rate": 0.00019214876033057848, + "loss": 0.5482, + "step": 445 + }, + { + "epoch": 0.73, + "grad_norm": 0.2426328808069229, + "learning_rate": 0.000191900826446281, + "loss": 0.6444, + "step": 446 + }, + { + "epoch": 0.73, + "grad_norm": 0.26187026500701904, + "learning_rate": 0.00019165289256198347, + "loss": 0.5443, + "step": 447 + }, + { + "epoch": 0.73, + "grad_norm": 0.2719630002975464, + "learning_rate": 0.00019140495867768595, + "loss": 0.6886, + "step": 448 + }, + { + "epoch": 0.74, + "grad_norm": 0.18477971851825714, + "learning_rate": 0.0001911570247933884, + "loss": 0.5292, + "step": 449 + }, + { + "epoch": 0.74, + "grad_norm": 0.2144313007593155, + "learning_rate": 0.0001909090909090909, + "loss": 0.4613, + "step": 450 + }, + { + "epoch": 0.74, + "grad_norm": 0.2580784857273102, + "learning_rate": 0.00019066115702479338, + "loss": 0.5606, + "step": 451 + }, + { + "epoch": 0.74, + "grad_norm": 0.3073588013648987, + "learning_rate": 0.00019041322314049586, + "loss": 0.6123, + "step": 452 + }, + { + "epoch": 0.74, + "grad_norm": 0.21787844598293304, + "learning_rate": 0.00019016528925619832, + "loss": 0.5939, + "step": 453 + }, + { + "epoch": 0.74, + "grad_norm": 0.255750447511673, + "learning_rate": 0.0001899173553719008, + "loss": 0.5739, + "step": 454 + }, + { + "epoch": 0.74, + "grad_norm": 0.24147820472717285, + "learning_rate": 0.00018966942148760329, + "loss": 0.6026, + "step": 455 + }, + { + "epoch": 0.75, + "grad_norm": 0.26172590255737305, + "learning_rate": 0.00018942148760330577, + "loss": 0.5166, + "step": 456 + }, + { + "epoch": 0.75, + "grad_norm": 0.2710455358028412, + "learning_rate": 0.00018917355371900825, + "loss": 0.6429, + "step": 457 + }, + { + "epoch": 0.75, + "grad_norm": 0.1971074640750885, + "learning_rate": 0.0001889256198347107, + "loss": 0.4799, + "step": 458 + }, + { + "epoch": 0.75, + "grad_norm": 0.23394368588924408, + "learning_rate": 0.0001886776859504132, + "loss": 0.5491, + "step": 459 + }, + { + "epoch": 0.75, + "grad_norm": 0.22820048034191132, + "learning_rate": 0.0001884297520661157, + "loss": 0.5343, + "step": 460 + }, + { + "epoch": 0.75, + "grad_norm": 0.23169974982738495, + "learning_rate": 0.0001881818181818182, + "loss": 0.5852, + "step": 461 + }, + { + "epoch": 0.76, + "grad_norm": 0.24015003442764282, + "learning_rate": 0.00018793388429752064, + "loss": 0.6209, + "step": 462 + }, + { + "epoch": 0.76, + "grad_norm": 0.2230776697397232, + "learning_rate": 0.00018768595041322313, + "loss": 0.6296, + "step": 463 + }, + { + "epoch": 0.76, + "grad_norm": 0.2518354654312134, + "learning_rate": 0.0001874380165289256, + "loss": 0.6167, + "step": 464 + }, + { + "epoch": 0.76, + "grad_norm": 0.338256299495697, + "learning_rate": 0.0001871900826446281, + "loss": 0.6512, + "step": 465 + }, + { + "epoch": 0.76, + "grad_norm": 0.23796728253364563, + "learning_rate": 0.00018694214876033055, + "loss": 0.8155, + "step": 466 + }, + { + "epoch": 0.76, + "grad_norm": 0.31516361236572266, + "learning_rate": 0.00018669421487603303, + "loss": 0.8023, + "step": 467 + }, + { + "epoch": 0.77, + "grad_norm": 0.2371574491262436, + "learning_rate": 0.00018644628099173552, + "loss": 0.5613, + "step": 468 + }, + { + "epoch": 0.77, + "grad_norm": 0.2822033762931824, + "learning_rate": 0.000186198347107438, + "loss": 0.5549, + "step": 469 + }, + { + "epoch": 0.77, + "grad_norm": 0.25953295826911926, + "learning_rate": 0.00018595041322314049, + "loss": 0.6199, + "step": 470 + }, + { + "epoch": 0.77, + "grad_norm": 0.2478639930486679, + "learning_rate": 0.00018570247933884294, + "loss": 0.5806, + "step": 471 + }, + { + "epoch": 0.77, + "grad_norm": 0.2439350187778473, + "learning_rate": 0.00018545454545454543, + "loss": 0.6222, + "step": 472 + }, + { + "epoch": 0.77, + "grad_norm": 0.24993474781513214, + "learning_rate": 0.0001852066115702479, + "loss": 0.6048, + "step": 473 + }, + { + "epoch": 0.78, + "grad_norm": 0.24781496822834015, + "learning_rate": 0.00018495867768595042, + "loss": 0.5941, + "step": 474 + }, + { + "epoch": 0.78, + "grad_norm": 0.1847202032804489, + "learning_rate": 0.00018471074380165285, + "loss": 0.609, + "step": 475 + }, + { + "epoch": 0.78, + "grad_norm": 0.21596528589725494, + "learning_rate": 0.00018446280991735536, + "loss": 0.4457, + "step": 476 + }, + { + "epoch": 0.78, + "grad_norm": 0.240879625082016, + "learning_rate": 0.00018421487603305784, + "loss": 0.6118, + "step": 477 + }, + { + "epoch": 0.78, + "grad_norm": 0.2898111641407013, + "learning_rate": 0.00018396694214876033, + "loss": 0.7725, + "step": 478 + }, + { + "epoch": 0.78, + "grad_norm": 0.27428382635116577, + "learning_rate": 0.0001837190082644628, + "loss": 0.5366, + "step": 479 + }, + { + "epoch": 0.79, + "grad_norm": 0.23467296361923218, + "learning_rate": 0.00018347107438016527, + "loss": 0.6018, + "step": 480 + }, + { + "epoch": 0.79, + "grad_norm": 0.2190561592578888, + "learning_rate": 0.00018322314049586775, + "loss": 0.5249, + "step": 481 + }, + { + "epoch": 0.79, + "grad_norm": 0.2240625023841858, + "learning_rate": 0.00018297520661157024, + "loss": 0.6891, + "step": 482 + }, + { + "epoch": 0.79, + "grad_norm": 0.24726848304271698, + "learning_rate": 0.00018272727272727272, + "loss": 0.5545, + "step": 483 + }, + { + "epoch": 0.79, + "grad_norm": 0.3318251371383667, + "learning_rate": 0.00018247933884297518, + "loss": 0.4809, + "step": 484 + }, + { + "epoch": 0.79, + "grad_norm": 0.2396695613861084, + "learning_rate": 0.00018223140495867766, + "loss": 0.4942, + "step": 485 + }, + { + "epoch": 0.8, + "grad_norm": 0.25009942054748535, + "learning_rate": 0.00018198347107438014, + "loss": 0.7381, + "step": 486 + }, + { + "epoch": 0.8, + "grad_norm": 0.22655311226844788, + "learning_rate": 0.00018173553719008263, + "loss": 0.4729, + "step": 487 + }, + { + "epoch": 0.8, + "grad_norm": 0.23187695443630219, + "learning_rate": 0.0001814876033057851, + "loss": 0.5719, + "step": 488 + }, + { + "epoch": 0.8, + "grad_norm": 0.2703653573989868, + "learning_rate": 0.00018123966942148757, + "loss": 0.6031, + "step": 489 + }, + { + "epoch": 0.8, + "grad_norm": 0.2207796424627304, + "learning_rate": 0.00018099173553719008, + "loss": 0.5361, + "step": 490 + }, + { + "epoch": 0.8, + "grad_norm": 0.24914169311523438, + "learning_rate": 0.00018074380165289256, + "loss": 0.6547, + "step": 491 + }, + { + "epoch": 0.81, + "grad_norm": 0.2714746594429016, + "learning_rate": 0.00018049586776859504, + "loss": 0.5702, + "step": 492 + }, + { + "epoch": 0.81, + "grad_norm": 0.3201580047607422, + "learning_rate": 0.0001802479338842975, + "loss": 0.6119, + "step": 493 + }, + { + "epoch": 0.81, + "grad_norm": 0.2548397183418274, + "learning_rate": 0.00017999999999999998, + "loss": 0.5251, + "step": 494 + }, + { + "epoch": 0.81, + "grad_norm": 0.28669115900993347, + "learning_rate": 0.00017975206611570247, + "loss": 0.5773, + "step": 495 + }, + { + "epoch": 0.81, + "grad_norm": 0.26253971457481384, + "learning_rate": 0.00017950413223140495, + "loss": 0.6504, + "step": 496 + }, + { + "epoch": 0.81, + "grad_norm": 0.22113384306430817, + "learning_rate": 0.00017925619834710744, + "loss": 0.4741, + "step": 497 + }, + { + "epoch": 0.82, + "grad_norm": 0.261636346578598, + "learning_rate": 0.0001790082644628099, + "loss": 0.6241, + "step": 498 + }, + { + "epoch": 0.82, + "grad_norm": 0.1780402809381485, + "learning_rate": 0.00017876033057851238, + "loss": 0.5207, + "step": 499 + }, + { + "epoch": 0.82, + "grad_norm": 0.26149195432662964, + "learning_rate": 0.00017851239669421486, + "loss": 0.5872, + "step": 500 + }, + { + "epoch": 0.82, + "grad_norm": 0.26113009452819824, + "learning_rate": 0.00017826446280991734, + "loss": 0.6163, + "step": 501 + }, + { + "epoch": 0.82, + "grad_norm": 0.21397502720355988, + "learning_rate": 0.0001780165289256198, + "loss": 0.479, + "step": 502 + }, + { + "epoch": 0.82, + "grad_norm": 0.21250088512897491, + "learning_rate": 0.00017776859504132228, + "loss": 0.6978, + "step": 503 + }, + { + "epoch": 0.83, + "grad_norm": 0.2556426525115967, + "learning_rate": 0.00017752066115702477, + "loss": 0.6128, + "step": 504 + }, + { + "epoch": 0.83, + "grad_norm": 0.24139715731143951, + "learning_rate": 0.00017727272727272728, + "loss": 0.5066, + "step": 505 + }, + { + "epoch": 0.83, + "grad_norm": 0.23671215772628784, + "learning_rate": 0.00017702479338842976, + "loss": 0.5183, + "step": 506 + }, + { + "epoch": 0.83, + "grad_norm": 0.23494285345077515, + "learning_rate": 0.00017677685950413222, + "loss": 0.5181, + "step": 507 + }, + { + "epoch": 0.83, + "grad_norm": 0.2547609806060791, + "learning_rate": 0.0001765289256198347, + "loss": 0.5406, + "step": 508 + }, + { + "epoch": 0.83, + "grad_norm": 0.3042651414871216, + "learning_rate": 0.00017628099173553718, + "loss": 0.5551, + "step": 509 + }, + { + "epoch": 0.84, + "grad_norm": 0.22910748422145844, + "learning_rate": 0.00017603305785123967, + "loss": 0.6373, + "step": 510 + }, + { + "epoch": 0.84, + "grad_norm": 0.19777967035770416, + "learning_rate": 0.00017578512396694212, + "loss": 0.5471, + "step": 511 + }, + { + "epoch": 0.84, + "grad_norm": 0.31034502387046814, + "learning_rate": 0.0001755371900826446, + "loss": 0.7017, + "step": 512 + }, + { + "epoch": 0.84, + "grad_norm": 0.3504410684108734, + "learning_rate": 0.0001752892561983471, + "loss": 0.7208, + "step": 513 + }, + { + "epoch": 0.84, + "grad_norm": 0.24271292984485626, + "learning_rate": 0.00017504132231404958, + "loss": 0.5563, + "step": 514 + }, + { + "epoch": 0.84, + "grad_norm": 0.27147865295410156, + "learning_rate": 0.00017479338842975203, + "loss": 0.5869, + "step": 515 + }, + { + "epoch": 0.84, + "grad_norm": 0.2976628839969635, + "learning_rate": 0.00017454545454545452, + "loss": 0.5471, + "step": 516 + }, + { + "epoch": 0.85, + "grad_norm": 0.28489646315574646, + "learning_rate": 0.000174297520661157, + "loss": 0.6053, + "step": 517 + }, + { + "epoch": 0.85, + "grad_norm": 0.30020108819007874, + "learning_rate": 0.00017404958677685948, + "loss": 0.6178, + "step": 518 + }, + { + "epoch": 0.85, + "grad_norm": 0.23986253142356873, + "learning_rate": 0.000173801652892562, + "loss": 0.5896, + "step": 519 + }, + { + "epoch": 0.85, + "grad_norm": 0.2667832374572754, + "learning_rate": 0.00017355371900826442, + "loss": 0.5375, + "step": 520 + }, + { + "epoch": 0.85, + "grad_norm": 0.22176356613636017, + "learning_rate": 0.00017330578512396693, + "loss": 0.5723, + "step": 521 + }, + { + "epoch": 0.85, + "grad_norm": 0.263257771730423, + "learning_rate": 0.00017305785123966942, + "loss": 0.7317, + "step": 522 + }, + { + "epoch": 0.86, + "grad_norm": 0.24838753044605255, + "learning_rate": 0.0001728099173553719, + "loss": 0.5849, + "step": 523 + }, + { + "epoch": 0.86, + "grad_norm": 0.24839664995670319, + "learning_rate": 0.00017256198347107436, + "loss": 0.6678, + "step": 524 + }, + { + "epoch": 0.86, + "grad_norm": 0.2849573493003845, + "learning_rate": 0.00017231404958677684, + "loss": 0.7144, + "step": 525 + }, + { + "epoch": 0.86, + "grad_norm": 0.26900768280029297, + "learning_rate": 0.00017206611570247932, + "loss": 0.5156, + "step": 526 + }, + { + "epoch": 0.86, + "grad_norm": 0.2212425172328949, + "learning_rate": 0.0001718181818181818, + "loss": 0.4551, + "step": 527 + }, + { + "epoch": 0.86, + "grad_norm": 0.2066129595041275, + "learning_rate": 0.0001715702479338843, + "loss": 0.4193, + "step": 528 + }, + { + "epoch": 0.87, + "grad_norm": 0.2838365137577057, + "learning_rate": 0.00017132231404958675, + "loss": 0.6078, + "step": 529 + }, + { + "epoch": 0.87, + "grad_norm": 0.239679753780365, + "learning_rate": 0.00017107438016528923, + "loss": 0.616, + "step": 530 + }, + { + "epoch": 0.87, + "grad_norm": 0.23269398510456085, + "learning_rate": 0.00017082644628099172, + "loss": 0.542, + "step": 531 + }, + { + "epoch": 0.87, + "grad_norm": 0.23838558793067932, + "learning_rate": 0.0001705785123966942, + "loss": 0.5147, + "step": 532 + }, + { + "epoch": 0.87, + "grad_norm": 0.2819415330886841, + "learning_rate": 0.00017033057851239666, + "loss": 0.6437, + "step": 533 + }, + { + "epoch": 0.87, + "grad_norm": 0.243398055434227, + "learning_rate": 0.00017008264462809914, + "loss": 0.6611, + "step": 534 + }, + { + "epoch": 0.88, + "grad_norm": 0.22569122910499573, + "learning_rate": 0.00016983471074380165, + "loss": 0.3979, + "step": 535 + }, + { + "epoch": 0.88, + "grad_norm": 0.33265820145606995, + "learning_rate": 0.00016958677685950413, + "loss": 0.6005, + "step": 536 + }, + { + "epoch": 0.88, + "grad_norm": 0.26828673481941223, + "learning_rate": 0.00016933884297520662, + "loss": 0.608, + "step": 537 + }, + { + "epoch": 0.88, + "grad_norm": 0.24439513683319092, + "learning_rate": 0.00016909090909090907, + "loss": 0.5572, + "step": 538 + }, + { + "epoch": 0.88, + "grad_norm": 0.22491876780986786, + "learning_rate": 0.00016884297520661156, + "loss": 0.7226, + "step": 539 + }, + { + "epoch": 0.88, + "grad_norm": 0.24468480050563812, + "learning_rate": 0.00016859504132231404, + "loss": 0.4582, + "step": 540 + }, + { + "epoch": 0.89, + "grad_norm": 0.23392945528030396, + "learning_rate": 0.00016834710743801652, + "loss": 0.6477, + "step": 541 + }, + { + "epoch": 0.89, + "grad_norm": 0.27548858523368835, + "learning_rate": 0.00016809917355371898, + "loss": 0.5846, + "step": 542 + }, + { + "epoch": 0.89, + "grad_norm": 0.2861180603504181, + "learning_rate": 0.00016785123966942146, + "loss": 0.6412, + "step": 543 + }, + { + "epoch": 0.89, + "grad_norm": 0.24700766801834106, + "learning_rate": 0.00016760330578512395, + "loss": 0.6947, + "step": 544 + }, + { + "epoch": 0.89, + "grad_norm": 0.2600953280925751, + "learning_rate": 0.00016735537190082643, + "loss": 0.6165, + "step": 545 + }, + { + "epoch": 0.89, + "grad_norm": 0.26876646280288696, + "learning_rate": 0.00016710743801652892, + "loss": 0.6855, + "step": 546 + }, + { + "epoch": 0.9, + "grad_norm": 0.26161080598831177, + "learning_rate": 0.00016685950413223137, + "loss": 0.5066, + "step": 547 + }, + { + "epoch": 0.9, + "grad_norm": 0.25190046429634094, + "learning_rate": 0.00016661157024793386, + "loss": 0.5902, + "step": 548 + }, + { + "epoch": 0.9, + "grad_norm": 0.25269225239753723, + "learning_rate": 0.00016636363636363637, + "loss": 0.7017, + "step": 549 + }, + { + "epoch": 0.9, + "grad_norm": 0.28042706847190857, + "learning_rate": 0.00016611570247933885, + "loss": 0.6264, + "step": 550 + }, + { + "epoch": 0.9, + "grad_norm": 0.2767360508441925, + "learning_rate": 0.0001658677685950413, + "loss": 0.7562, + "step": 551 + }, + { + "epoch": 0.9, + "grad_norm": 0.2771216034889221, + "learning_rate": 0.0001656198347107438, + "loss": 0.5333, + "step": 552 + }, + { + "epoch": 0.91, + "grad_norm": 0.189210906624794, + "learning_rate": 0.00016537190082644627, + "loss": 0.5378, + "step": 553 + }, + { + "epoch": 0.91, + "grad_norm": 0.22517065703868866, + "learning_rate": 0.00016512396694214876, + "loss": 0.5292, + "step": 554 + }, + { + "epoch": 0.91, + "grad_norm": 0.2390165776014328, + "learning_rate": 0.00016487603305785121, + "loss": 0.4407, + "step": 555 + }, + { + "epoch": 0.91, + "grad_norm": 0.21548262238502502, + "learning_rate": 0.0001646280991735537, + "loss": 0.4504, + "step": 556 + }, + { + "epoch": 0.91, + "grad_norm": 0.20831167697906494, + "learning_rate": 0.00016438016528925618, + "loss": 0.6848, + "step": 557 + }, + { + "epoch": 0.91, + "grad_norm": 0.271257609128952, + "learning_rate": 0.00016413223140495866, + "loss": 0.535, + "step": 558 + }, + { + "epoch": 0.92, + "grad_norm": 0.32008254528045654, + "learning_rate": 0.00016388429752066115, + "loss": 0.5107, + "step": 559 + }, + { + "epoch": 0.92, + "grad_norm": 0.34058302640914917, + "learning_rate": 0.0001636363636363636, + "loss": 0.5708, + "step": 560 + }, + { + "epoch": 0.92, + "grad_norm": 0.28070059418678284, + "learning_rate": 0.0001633884297520661, + "loss": 0.5086, + "step": 561 + }, + { + "epoch": 0.92, + "grad_norm": 0.25487688183784485, + "learning_rate": 0.00016314049586776857, + "loss": 0.5184, + "step": 562 + }, + { + "epoch": 0.92, + "grad_norm": 0.3240332007408142, + "learning_rate": 0.00016289256198347108, + "loss": 0.6774, + "step": 563 + }, + { + "epoch": 0.92, + "grad_norm": 0.30744409561157227, + "learning_rate": 0.0001626446280991735, + "loss": 0.5314, + "step": 564 + }, + { + "epoch": 0.93, + "grad_norm": 0.25220754742622375, + "learning_rate": 0.00016239669421487602, + "loss": 0.6308, + "step": 565 + }, + { + "epoch": 0.93, + "grad_norm": 0.29116958379745483, + "learning_rate": 0.0001621487603305785, + "loss": 0.5685, + "step": 566 + }, + { + "epoch": 0.93, + "grad_norm": 0.23250073194503784, + "learning_rate": 0.000161900826446281, + "loss": 0.4318, + "step": 567 + }, + { + "epoch": 0.93, + "grad_norm": 0.2808091640472412, + "learning_rate": 0.00016165289256198347, + "loss": 0.6313, + "step": 568 + }, + { + "epoch": 0.93, + "grad_norm": 0.2711193561553955, + "learning_rate": 0.00016140495867768593, + "loss": 0.4651, + "step": 569 + }, + { + "epoch": 0.93, + "grad_norm": 0.29540935158729553, + "learning_rate": 0.00016115702479338841, + "loss": 0.6663, + "step": 570 + }, + { + "epoch": 0.93, + "grad_norm": 0.23418714106082916, + "learning_rate": 0.0001609090909090909, + "loss": 0.448, + "step": 571 + }, + { + "epoch": 0.94, + "grad_norm": 0.21675793826580048, + "learning_rate": 0.00016066115702479338, + "loss": 0.5034, + "step": 572 + }, + { + "epoch": 0.94, + "grad_norm": 0.22451865673065186, + "learning_rate": 0.00016041322314049584, + "loss": 0.4476, + "step": 573 + }, + { + "epoch": 0.94, + "grad_norm": 0.26300856471061707, + "learning_rate": 0.00016016528925619832, + "loss": 0.6646, + "step": 574 + }, + { + "epoch": 0.94, + "grad_norm": 0.3377116918563843, + "learning_rate": 0.0001599173553719008, + "loss": 0.6029, + "step": 575 + }, + { + "epoch": 0.94, + "grad_norm": 0.23391880095005035, + "learning_rate": 0.0001596694214876033, + "loss": 0.6277, + "step": 576 + }, + { + "epoch": 0.94, + "grad_norm": 0.19620922207832336, + "learning_rate": 0.0001594214876033058, + "loss": 0.4638, + "step": 577 + }, + { + "epoch": 0.95, + "grad_norm": 0.22981096804141998, + "learning_rate": 0.00015917355371900823, + "loss": 0.5826, + "step": 578 + }, + { + "epoch": 0.95, + "grad_norm": 0.34321555495262146, + "learning_rate": 0.00015892561983471074, + "loss": 0.5618, + "step": 579 + }, + { + "epoch": 0.95, + "grad_norm": 0.28461968898773193, + "learning_rate": 0.00015867768595041322, + "loss": 0.5129, + "step": 580 + }, + { + "epoch": 0.95, + "grad_norm": 0.24368269741535187, + "learning_rate": 0.0001584297520661157, + "loss": 0.5866, + "step": 581 + }, + { + "epoch": 0.95, + "grad_norm": 0.282255083322525, + "learning_rate": 0.00015818181818181816, + "loss": 0.6274, + "step": 582 + }, + { + "epoch": 0.95, + "grad_norm": 0.26298072934150696, + "learning_rate": 0.00015793388429752065, + "loss": 0.5187, + "step": 583 + }, + { + "epoch": 0.96, + "grad_norm": 0.2671455144882202, + "learning_rate": 0.00015768595041322313, + "loss": 0.6878, + "step": 584 + }, + { + "epoch": 0.96, + "grad_norm": 0.2681390643119812, + "learning_rate": 0.00015743801652892561, + "loss": 0.5469, + "step": 585 + }, + { + "epoch": 0.96, + "grad_norm": 0.38484248518943787, + "learning_rate": 0.0001571900826446281, + "loss": 0.6364, + "step": 586 + }, + { + "epoch": 0.96, + "grad_norm": 0.23353587090969086, + "learning_rate": 0.00015694214876033055, + "loss": 0.4844, + "step": 587 + }, + { + "epoch": 0.96, + "grad_norm": 0.29452502727508545, + "learning_rate": 0.00015669421487603304, + "loss": 0.5059, + "step": 588 + }, + { + "epoch": 0.96, + "grad_norm": 0.2460879236459732, + "learning_rate": 0.00015644628099173552, + "loss": 0.6495, + "step": 589 + }, + { + "epoch": 0.97, + "grad_norm": 0.30693721771240234, + "learning_rate": 0.000156198347107438, + "loss": 0.5165, + "step": 590 + }, + { + "epoch": 0.97, + "grad_norm": 0.2171495109796524, + "learning_rate": 0.00015595041322314046, + "loss": 0.6172, + "step": 591 + }, + { + "epoch": 0.97, + "grad_norm": 0.24301984906196594, + "learning_rate": 0.00015570247933884294, + "loss": 0.6786, + "step": 592 + }, + { + "epoch": 0.97, + "grad_norm": 0.2288222461938858, + "learning_rate": 0.00015545454545454546, + "loss": 0.5669, + "step": 593 + }, + { + "epoch": 0.97, + "grad_norm": 0.2407921552658081, + "learning_rate": 0.00015520661157024794, + "loss": 0.5968, + "step": 594 + }, + { + "epoch": 0.97, + "grad_norm": 0.2591527998447418, + "learning_rate": 0.0001549586776859504, + "loss": 0.544, + "step": 595 + }, + { + "epoch": 0.98, + "grad_norm": 0.25770679116249084, + "learning_rate": 0.00015471074380165288, + "loss": 0.7177, + "step": 596 + }, + { + "epoch": 0.98, + "grad_norm": 0.2528848648071289, + "learning_rate": 0.00015446280991735536, + "loss": 0.4703, + "step": 597 + }, + { + "epoch": 0.98, + "grad_norm": 0.24993537366390228, + "learning_rate": 0.00015421487603305785, + "loss": 0.6003, + "step": 598 + }, + { + "epoch": 0.98, + "grad_norm": 0.25807908177375793, + "learning_rate": 0.00015396694214876033, + "loss": 0.465, + "step": 599 + }, + { + "epoch": 0.98, + "grad_norm": 0.3142452836036682, + "learning_rate": 0.0001537190082644628, + "loss": 0.6122, + "step": 600 + }, + { + "epoch": 0.98, + "grad_norm": 0.27111849188804626, + "learning_rate": 0.00015347107438016527, + "loss": 0.5962, + "step": 601 + }, + { + "epoch": 0.99, + "grad_norm": 0.28503674268722534, + "learning_rate": 0.00015322314049586775, + "loss": 0.6667, + "step": 602 + }, + { + "epoch": 0.99, + "grad_norm": 0.27074381709098816, + "learning_rate": 0.00015297520661157024, + "loss": 0.6115, + "step": 603 + }, + { + "epoch": 0.99, + "grad_norm": 0.25918465852737427, + "learning_rate": 0.0001527272727272727, + "loss": 0.4483, + "step": 604 + }, + { + "epoch": 0.99, + "grad_norm": 0.24476633965969086, + "learning_rate": 0.00015247933884297518, + "loss": 0.6501, + "step": 605 + }, + { + "epoch": 0.99, + "grad_norm": 0.21205200254917145, + "learning_rate": 0.00015223140495867766, + "loss": 0.3914, + "step": 606 + }, + { + "epoch": 0.99, + "grad_norm": 0.25496751070022583, + "learning_rate": 0.00015198347107438017, + "loss": 0.5335, + "step": 607 + }, + { + "epoch": 1.0, + "grad_norm": 0.27991780638694763, + "learning_rate": 0.00015173553719008266, + "loss": 0.6083, + "step": 608 + }, + { + "epoch": 1.0, + "grad_norm": 0.23995639383792877, + "learning_rate": 0.0001514876033057851, + "loss": 0.55, + "step": 609 + }, + { + "epoch": 1.0, + "grad_norm": 0.2349666953086853, + "learning_rate": 0.0001512396694214876, + "loss": 0.7054, + "step": 610 + }, + { + "epoch": 1.0, + "grad_norm": 0.27498871088027954, + "learning_rate": 0.00015099173553719008, + "loss": 0.55, + "step": 611 + }, + { + "epoch": 1.0, + "grad_norm": 0.21346105635166168, + "learning_rate": 0.00015074380165289256, + "loss": 0.3467, + "step": 612 + }, + { + "epoch": 1.0, + "grad_norm": 0.2638354003429413, + "learning_rate": 0.00015049586776859502, + "loss": 0.5624, + "step": 613 + }, + { + "epoch": 1.01, + "grad_norm": 0.2751975953578949, + "learning_rate": 0.0001502479338842975, + "loss": 0.3814, + "step": 614 + }, + { + "epoch": 1.01, + "grad_norm": 0.225106880068779, + "learning_rate": 0.00015, + "loss": 0.479, + "step": 615 + }, + { + "epoch": 1.01, + "grad_norm": 0.22013232111930847, + "learning_rate": 0.00014975206611570247, + "loss": 0.5672, + "step": 616 + }, + { + "epoch": 1.01, + "grad_norm": 0.21252033114433289, + "learning_rate": 0.00014950413223140495, + "loss": 0.546, + "step": 617 + }, + { + "epoch": 1.01, + "grad_norm": 0.2847185432910919, + "learning_rate": 0.00014925619834710744, + "loss": 0.4434, + "step": 618 + }, + { + "epoch": 1.01, + "grad_norm": 0.25599631667137146, + "learning_rate": 0.0001490082644628099, + "loss": 0.4713, + "step": 619 + }, + { + "epoch": 1.02, + "grad_norm": 0.2719402611255646, + "learning_rate": 0.00014876033057851238, + "loss": 0.4475, + "step": 620 + }, + { + "epoch": 1.02, + "grad_norm": 0.26454958319664, + "learning_rate": 0.00014851239669421486, + "loss": 0.4515, + "step": 621 + }, + { + "epoch": 1.02, + "grad_norm": 0.39801672101020813, + "learning_rate": 0.00014826446280991735, + "loss": 0.4647, + "step": 622 + }, + { + "epoch": 1.02, + "grad_norm": 0.3378361463546753, + "learning_rate": 0.0001480165289256198, + "loss": 0.4414, + "step": 623 + }, + { + "epoch": 1.02, + "grad_norm": 0.3039036989212036, + "learning_rate": 0.0001477685950413223, + "loss": 0.5634, + "step": 624 + }, + { + "epoch": 1.02, + "grad_norm": 0.3506157398223877, + "learning_rate": 0.0001475206611570248, + "loss": 0.5001, + "step": 625 + }, + { + "epoch": 1.02, + "grad_norm": 0.2508845925331116, + "learning_rate": 0.00014727272727272725, + "loss": 0.3379, + "step": 626 + }, + { + "epoch": 1.03, + "grad_norm": 0.26913216710090637, + "learning_rate": 0.00014702479338842974, + "loss": 0.4575, + "step": 627 + }, + { + "epoch": 1.03, + "grad_norm": 0.329659640789032, + "learning_rate": 0.00014677685950413222, + "loss": 0.437, + "step": 628 + }, + { + "epoch": 1.03, + "grad_norm": 0.2972075343132019, + "learning_rate": 0.0001465289256198347, + "loss": 0.5048, + "step": 629 + }, + { + "epoch": 1.03, + "grad_norm": 0.3184354603290558, + "learning_rate": 0.00014628099173553716, + "loss": 0.4374, + "step": 630 + }, + { + "epoch": 1.03, + "grad_norm": 0.3377355635166168, + "learning_rate": 0.00014603305785123967, + "loss": 0.4946, + "step": 631 + }, + { + "epoch": 1.03, + "grad_norm": 0.29106801748275757, + "learning_rate": 0.00014578512396694213, + "loss": 0.5414, + "step": 632 + }, + { + "epoch": 1.04, + "grad_norm": 0.22808948159217834, + "learning_rate": 0.0001455371900826446, + "loss": 0.3739, + "step": 633 + }, + { + "epoch": 1.04, + "grad_norm": 0.27818021178245544, + "learning_rate": 0.0001452892561983471, + "loss": 0.4172, + "step": 634 + }, + { + "epoch": 1.04, + "grad_norm": 0.25634923577308655, + "learning_rate": 0.00014504132231404958, + "loss": 0.4293, + "step": 635 + }, + { + "epoch": 1.04, + "grad_norm": 0.30696937441825867, + "learning_rate": 0.00014479338842975206, + "loss": 0.4454, + "step": 636 + }, + { + "epoch": 1.04, + "grad_norm": 0.26105087995529175, + "learning_rate": 0.00014454545454545452, + "loss": 0.2978, + "step": 637 + }, + { + "epoch": 1.04, + "grad_norm": 0.3100634515285492, + "learning_rate": 0.00014429752066115703, + "loss": 0.4499, + "step": 638 + }, + { + "epoch": 1.05, + "grad_norm": 0.27640992403030396, + "learning_rate": 0.00014404958677685949, + "loss": 0.3837, + "step": 639 + }, + { + "epoch": 1.05, + "grad_norm": 0.24559038877487183, + "learning_rate": 0.00014380165289256197, + "loss": 0.3347, + "step": 640 + }, + { + "epoch": 1.05, + "grad_norm": 0.2920415699481964, + "learning_rate": 0.00014355371900826445, + "loss": 0.4333, + "step": 641 + }, + { + "epoch": 1.05, + "grad_norm": 0.3147384226322174, + "learning_rate": 0.00014330578512396694, + "loss": 0.4385, + "step": 642 + }, + { + "epoch": 1.05, + "grad_norm": 0.35469138622283936, + "learning_rate": 0.0001430578512396694, + "loss": 0.5442, + "step": 643 + }, + { + "epoch": 1.05, + "grad_norm": 0.2619563043117523, + "learning_rate": 0.00014280991735537188, + "loss": 0.3837, + "step": 644 + }, + { + "epoch": 1.06, + "grad_norm": 0.32273221015930176, + "learning_rate": 0.0001425619834710744, + "loss": 0.4946, + "step": 645 + }, + { + "epoch": 1.06, + "grad_norm": 0.2692110538482666, + "learning_rate": 0.00014231404958677684, + "loss": 0.4683, + "step": 646 + }, + { + "epoch": 1.06, + "grad_norm": 0.35255464911460876, + "learning_rate": 0.00014206611570247933, + "loss": 0.5456, + "step": 647 + }, + { + "epoch": 1.06, + "grad_norm": 0.29768630862236023, + "learning_rate": 0.0001418181818181818, + "loss": 0.3394, + "step": 648 + }, + { + "epoch": 1.06, + "grad_norm": 0.30738797783851624, + "learning_rate": 0.0001415702479338843, + "loss": 0.3583, + "step": 649 + }, + { + "epoch": 1.06, + "grad_norm": 0.33226314187049866, + "learning_rate": 0.00014132231404958675, + "loss": 0.4477, + "step": 650 + }, + { + "epoch": 1.07, + "grad_norm": 0.2842199504375458, + "learning_rate": 0.00014107438016528923, + "loss": 0.4454, + "step": 651 + }, + { + "epoch": 1.07, + "grad_norm": 0.28207266330718994, + "learning_rate": 0.00014082644628099172, + "loss": 0.3665, + "step": 652 + }, + { + "epoch": 1.07, + "grad_norm": 0.2228500097990036, + "learning_rate": 0.0001405785123966942, + "loss": 0.3446, + "step": 653 + }, + { + "epoch": 1.07, + "grad_norm": 0.2969403564929962, + "learning_rate": 0.00014033057851239669, + "loss": 0.377, + "step": 654 + }, + { + "epoch": 1.07, + "grad_norm": 0.28087565302848816, + "learning_rate": 0.00014008264462809917, + "loss": 0.3683, + "step": 655 + }, + { + "epoch": 1.07, + "grad_norm": 0.27268192172050476, + "learning_rate": 0.00013983471074380165, + "loss": 0.427, + "step": 656 + }, + { + "epoch": 1.08, + "grad_norm": 0.339070588350296, + "learning_rate": 0.0001395867768595041, + "loss": 0.4887, + "step": 657 + }, + { + "epoch": 1.08, + "grad_norm": 0.3170423209667206, + "learning_rate": 0.0001393388429752066, + "loss": 0.5097, + "step": 658 + }, + { + "epoch": 1.08, + "grad_norm": 0.3114936947822571, + "learning_rate": 0.00013909090909090908, + "loss": 0.4587, + "step": 659 + }, + { + "epoch": 1.08, + "grad_norm": 0.28112486004829407, + "learning_rate": 0.00013884297520661156, + "loss": 0.4781, + "step": 660 + }, + { + "epoch": 1.08, + "grad_norm": 0.28116974234580994, + "learning_rate": 0.00013859504132231404, + "loss": 0.3546, + "step": 661 + }, + { + "epoch": 1.08, + "grad_norm": 0.25061559677124023, + "learning_rate": 0.00013834710743801653, + "loss": 0.4512, + "step": 662 + }, + { + "epoch": 1.09, + "grad_norm": 0.29854199290275574, + "learning_rate": 0.00013809917355371898, + "loss": 0.6068, + "step": 663 + }, + { + "epoch": 1.09, + "grad_norm": 0.2901363670825958, + "learning_rate": 0.00013785123966942147, + "loss": 0.3667, + "step": 664 + }, + { + "epoch": 1.09, + "grad_norm": 0.29766595363616943, + "learning_rate": 0.00013760330578512395, + "loss": 0.5194, + "step": 665 + }, + { + "epoch": 1.09, + "grad_norm": 0.2765616476535797, + "learning_rate": 0.00013735537190082643, + "loss": 0.5079, + "step": 666 + }, + { + "epoch": 1.09, + "grad_norm": 0.27531540393829346, + "learning_rate": 0.00013710743801652892, + "loss": 0.4423, + "step": 667 + }, + { + "epoch": 1.09, + "grad_norm": 0.3063349425792694, + "learning_rate": 0.0001368595041322314, + "loss": 0.4666, + "step": 668 + }, + { + "epoch": 1.1, + "grad_norm": 0.24519848823547363, + "learning_rate": 0.00013661157024793389, + "loss": 0.2995, + "step": 669 + }, + { + "epoch": 1.1, + "grad_norm": 0.4366275370121002, + "learning_rate": 0.00013636363636363634, + "loss": 0.4961, + "step": 670 + }, + { + "epoch": 1.1, + "grad_norm": 0.28639987111091614, + "learning_rate": 0.00013611570247933883, + "loss": 0.5015, + "step": 671 + }, + { + "epoch": 1.1, + "grad_norm": 0.2763878107070923, + "learning_rate": 0.0001358677685950413, + "loss": 0.4883, + "step": 672 + }, + { + "epoch": 1.1, + "grad_norm": 0.5589582324028015, + "learning_rate": 0.0001356198347107438, + "loss": 0.5072, + "step": 673 + }, + { + "epoch": 1.1, + "grad_norm": 0.238887220621109, + "learning_rate": 0.00013537190082644625, + "loss": 0.411, + "step": 674 + }, + { + "epoch": 1.11, + "grad_norm": 0.2899521589279175, + "learning_rate": 0.00013512396694214876, + "loss": 0.3478, + "step": 675 + }, + { + "epoch": 1.11, + "grad_norm": 0.30960512161254883, + "learning_rate": 0.00013487603305785124, + "loss": 0.5058, + "step": 676 + }, + { + "epoch": 1.11, + "grad_norm": 0.33305928111076355, + "learning_rate": 0.0001346280991735537, + "loss": 0.4528, + "step": 677 + }, + { + "epoch": 1.11, + "grad_norm": 0.33324292302131653, + "learning_rate": 0.00013438016528925618, + "loss": 0.3523, + "step": 678 + }, + { + "epoch": 1.11, + "grad_norm": 0.25855520367622375, + "learning_rate": 0.00013413223140495867, + "loss": 0.4257, + "step": 679 + }, + { + "epoch": 1.11, + "grad_norm": 0.36000239849090576, + "learning_rate": 0.00013388429752066115, + "loss": 0.4963, + "step": 680 + }, + { + "epoch": 1.12, + "grad_norm": 0.30540961027145386, + "learning_rate": 0.0001336363636363636, + "loss": 0.4706, + "step": 681 + }, + { + "epoch": 1.12, + "grad_norm": 0.2791118025779724, + "learning_rate": 0.00013338842975206612, + "loss": 0.4543, + "step": 682 + }, + { + "epoch": 1.12, + "grad_norm": 0.37401753664016724, + "learning_rate": 0.00013314049586776857, + "loss": 0.5614, + "step": 683 + }, + { + "epoch": 1.12, + "grad_norm": 0.2772528827190399, + "learning_rate": 0.00013289256198347106, + "loss": 0.3881, + "step": 684 + }, + { + "epoch": 1.12, + "grad_norm": 0.29219475388526917, + "learning_rate": 0.00013264462809917354, + "loss": 0.5418, + "step": 685 + }, + { + "epoch": 1.12, + "grad_norm": 0.3255159258842468, + "learning_rate": 0.00013239669421487603, + "loss": 0.4669, + "step": 686 + }, + { + "epoch": 1.12, + "grad_norm": 0.2640572488307953, + "learning_rate": 0.0001321487603305785, + "loss": 0.4156, + "step": 687 + }, + { + "epoch": 1.13, + "grad_norm": 0.2618845999240875, + "learning_rate": 0.00013190082644628097, + "loss": 0.3537, + "step": 688 + }, + { + "epoch": 1.13, + "grad_norm": 0.27396076917648315, + "learning_rate": 0.00013165289256198348, + "loss": 0.4391, + "step": 689 + }, + { + "epoch": 1.13, + "grad_norm": 0.5098498463630676, + "learning_rate": 0.00013140495867768593, + "loss": 0.3863, + "step": 690 + }, + { + "epoch": 1.13, + "grad_norm": 0.31764644384384155, + "learning_rate": 0.00013115702479338842, + "loss": 0.3874, + "step": 691 + }, + { + "epoch": 1.13, + "grad_norm": 0.28738152980804443, + "learning_rate": 0.0001309090909090909, + "loss": 0.3209, + "step": 692 + }, + { + "epoch": 1.13, + "grad_norm": 0.32756757736206055, + "learning_rate": 0.00013066115702479338, + "loss": 0.4614, + "step": 693 + }, + { + "epoch": 1.14, + "grad_norm": 0.27650028467178345, + "learning_rate": 0.00013041322314049584, + "loss": 0.4717, + "step": 694 + }, + { + "epoch": 1.14, + "grad_norm": 0.33100056648254395, + "learning_rate": 0.00013016528925619832, + "loss": 0.4317, + "step": 695 + }, + { + "epoch": 1.14, + "grad_norm": 0.3200342357158661, + "learning_rate": 0.00012991735537190083, + "loss": 0.4494, + "step": 696 + }, + { + "epoch": 1.14, + "grad_norm": 0.29615214467048645, + "learning_rate": 0.0001296694214876033, + "loss": 0.3786, + "step": 697 + }, + { + "epoch": 1.14, + "grad_norm": 0.278094619512558, + "learning_rate": 0.00012942148760330577, + "loss": 0.4484, + "step": 698 + }, + { + "epoch": 1.14, + "grad_norm": 0.32800769805908203, + "learning_rate": 0.00012917355371900826, + "loss": 0.4635, + "step": 699 + }, + { + "epoch": 1.15, + "grad_norm": 0.3319619596004486, + "learning_rate": 0.00012892561983471074, + "loss": 0.5001, + "step": 700 + }, + { + "epoch": 1.15, + "grad_norm": 0.2818608283996582, + "learning_rate": 0.0001286776859504132, + "loss": 0.3536, + "step": 701 + }, + { + "epoch": 1.15, + "grad_norm": 0.28644126653671265, + "learning_rate": 0.00012842975206611568, + "loss": 0.4168, + "step": 702 + }, + { + "epoch": 1.15, + "grad_norm": 0.2802482545375824, + "learning_rate": 0.00012818181818181817, + "loss": 0.3918, + "step": 703 + }, + { + "epoch": 1.15, + "grad_norm": 0.21232947707176208, + "learning_rate": 0.00012793388429752065, + "loss": 0.3218, + "step": 704 + }, + { + "epoch": 1.15, + "grad_norm": 0.36512815952301025, + "learning_rate": 0.00012768595041322313, + "loss": 0.4566, + "step": 705 + }, + { + "epoch": 1.16, + "grad_norm": 0.26876160502433777, + "learning_rate": 0.00012743801652892562, + "loss": 0.4394, + "step": 706 + }, + { + "epoch": 1.16, + "grad_norm": 0.3757662773132324, + "learning_rate": 0.0001271900826446281, + "loss": 0.574, + "step": 707 + }, + { + "epoch": 1.16, + "grad_norm": 0.3161550760269165, + "learning_rate": 0.00012694214876033056, + "loss": 0.4524, + "step": 708 + }, + { + "epoch": 1.16, + "grad_norm": 0.31256961822509766, + "learning_rate": 0.00012669421487603304, + "loss": 0.4332, + "step": 709 + }, + { + "epoch": 1.16, + "grad_norm": 0.3122079074382782, + "learning_rate": 0.00012644628099173552, + "loss": 0.5669, + "step": 710 + }, + { + "epoch": 1.16, + "grad_norm": 0.33779048919677734, + "learning_rate": 0.000126198347107438, + "loss": 0.515, + "step": 711 + }, + { + "epoch": 1.17, + "grad_norm": 0.38516169786453247, + "learning_rate": 0.0001259504132231405, + "loss": 0.5502, + "step": 712 + }, + { + "epoch": 1.17, + "grad_norm": 0.2803480625152588, + "learning_rate": 0.00012570247933884297, + "loss": 0.404, + "step": 713 + }, + { + "epoch": 1.17, + "grad_norm": 0.31674399971961975, + "learning_rate": 0.00012545454545454543, + "loss": 0.4403, + "step": 714 + }, + { + "epoch": 1.17, + "grad_norm": 0.3029496669769287, + "learning_rate": 0.00012520661157024791, + "loss": 0.372, + "step": 715 + }, + { + "epoch": 1.17, + "grad_norm": 0.22542959451675415, + "learning_rate": 0.0001249586776859504, + "loss": 0.355, + "step": 716 + }, + { + "epoch": 1.17, + "grad_norm": 0.32029619812965393, + "learning_rate": 0.00012471074380165288, + "loss": 0.4845, + "step": 717 + }, + { + "epoch": 1.18, + "grad_norm": 0.34882861375808716, + "learning_rate": 0.00012446280991735537, + "loss": 0.4184, + "step": 718 + }, + { + "epoch": 1.18, + "grad_norm": 0.3319970667362213, + "learning_rate": 0.00012421487603305785, + "loss": 0.5733, + "step": 719 + }, + { + "epoch": 1.18, + "grad_norm": 0.2770652770996094, + "learning_rate": 0.00012396694214876033, + "loss": 0.4296, + "step": 720 + }, + { + "epoch": 1.18, + "grad_norm": 0.3109978437423706, + "learning_rate": 0.0001237190082644628, + "loss": 0.3757, + "step": 721 + }, + { + "epoch": 1.18, + "grad_norm": 0.23606395721435547, + "learning_rate": 0.00012347107438016527, + "loss": 0.2713, + "step": 722 + }, + { + "epoch": 1.18, + "grad_norm": 0.304574579000473, + "learning_rate": 0.00012322314049586776, + "loss": 0.4451, + "step": 723 + }, + { + "epoch": 1.19, + "grad_norm": 0.31314462423324585, + "learning_rate": 0.00012297520661157024, + "loss": 0.493, + "step": 724 + }, + { + "epoch": 1.19, + "grad_norm": 0.32014840841293335, + "learning_rate": 0.00012272727272727272, + "loss": 0.3784, + "step": 725 + }, + { + "epoch": 1.19, + "grad_norm": 0.29856279492378235, + "learning_rate": 0.0001224793388429752, + "loss": 0.581, + "step": 726 + }, + { + "epoch": 1.19, + "grad_norm": 0.30951863527297974, + "learning_rate": 0.0001222314049586777, + "loss": 0.4851, + "step": 727 + }, + { + "epoch": 1.19, + "grad_norm": 0.264663428068161, + "learning_rate": 0.00012198347107438015, + "loss": 0.431, + "step": 728 + }, + { + "epoch": 1.19, + "grad_norm": 0.3092226982116699, + "learning_rate": 0.00012173553719008264, + "loss": 0.4553, + "step": 729 + }, + { + "epoch": 1.2, + "grad_norm": 0.33568286895751953, + "learning_rate": 0.00012148760330578511, + "loss": 0.4894, + "step": 730 + }, + { + "epoch": 1.2, + "grad_norm": 0.2966444492340088, + "learning_rate": 0.0001212396694214876, + "loss": 0.3855, + "step": 731 + }, + { + "epoch": 1.2, + "grad_norm": 0.2829122841358185, + "learning_rate": 0.00012099173553719007, + "loss": 0.5328, + "step": 732 + }, + { + "epoch": 1.2, + "grad_norm": 0.31785663962364197, + "learning_rate": 0.00012074380165289255, + "loss": 0.4142, + "step": 733 + }, + { + "epoch": 1.2, + "grad_norm": 0.2983114719390869, + "learning_rate": 0.00012049586776859502, + "loss": 0.4168, + "step": 734 + }, + { + "epoch": 1.2, + "grad_norm": 0.2514868378639221, + "learning_rate": 0.0001202479338842975, + "loss": 0.4728, + "step": 735 + }, + { + "epoch": 1.21, + "grad_norm": 0.2959445118904114, + "learning_rate": 0.00011999999999999999, + "loss": 0.458, + "step": 736 + }, + { + "epoch": 1.21, + "grad_norm": 0.31830325722694397, + "learning_rate": 0.00011975206611570247, + "loss": 0.5035, + "step": 737 + }, + { + "epoch": 1.21, + "grad_norm": 0.31181418895721436, + "learning_rate": 0.00011950413223140496, + "loss": 0.3776, + "step": 738 + }, + { + "epoch": 1.21, + "grad_norm": 0.3027549684047699, + "learning_rate": 0.00011925619834710743, + "loss": 0.4483, + "step": 739 + }, + { + "epoch": 1.21, + "grad_norm": 0.28026890754699707, + "learning_rate": 0.00011900826446280991, + "loss": 0.4236, + "step": 740 + }, + { + "epoch": 1.21, + "grad_norm": 0.29137665033340454, + "learning_rate": 0.00011876033057851238, + "loss": 0.3615, + "step": 741 + }, + { + "epoch": 1.21, + "grad_norm": 0.282008558511734, + "learning_rate": 0.00011851239669421486, + "loss": 0.4335, + "step": 742 + }, + { + "epoch": 1.22, + "grad_norm": 0.297736793756485, + "learning_rate": 0.00011826446280991733, + "loss": 0.4945, + "step": 743 + }, + { + "epoch": 1.22, + "grad_norm": 0.3276868164539337, + "learning_rate": 0.00011801652892561982, + "loss": 0.5379, + "step": 744 + }, + { + "epoch": 1.22, + "grad_norm": 0.3510095179080963, + "learning_rate": 0.00011776859504132231, + "loss": 0.3589, + "step": 745 + }, + { + "epoch": 1.22, + "grad_norm": 0.29952242970466614, + "learning_rate": 0.00011752066115702478, + "loss": 0.3805, + "step": 746 + }, + { + "epoch": 1.22, + "grad_norm": 0.220473513007164, + "learning_rate": 0.00011727272727272727, + "loss": 0.3978, + "step": 747 + }, + { + "epoch": 1.22, + "grad_norm": 0.30668944120407104, + "learning_rate": 0.00011702479338842974, + "loss": 0.3577, + "step": 748 + }, + { + "epoch": 1.23, + "grad_norm": 0.3152049779891968, + "learning_rate": 0.00011677685950413222, + "loss": 0.5186, + "step": 749 + }, + { + "epoch": 1.23, + "grad_norm": 0.17376375198364258, + "learning_rate": 0.00011652892561983469, + "loss": 0.32, + "step": 750 + }, + { + "epoch": 1.23, + "grad_norm": 0.32847121357917786, + "learning_rate": 0.00011628099173553718, + "loss": 0.5403, + "step": 751 + }, + { + "epoch": 1.23, + "grad_norm": 0.28821662068367004, + "learning_rate": 0.00011603305785123965, + "loss": 0.3516, + "step": 752 + }, + { + "epoch": 1.23, + "grad_norm": 0.23324501514434814, + "learning_rate": 0.00011578512396694214, + "loss": 0.3398, + "step": 753 + }, + { + "epoch": 1.23, + "grad_norm": 0.2897385060787201, + "learning_rate": 0.00011553719008264463, + "loss": 0.3775, + "step": 754 + }, + { + "epoch": 1.24, + "grad_norm": 0.33701419830322266, + "learning_rate": 0.0001152892561983471, + "loss": 0.5225, + "step": 755 + }, + { + "epoch": 1.24, + "grad_norm": 0.3228382468223572, + "learning_rate": 0.00011504132231404958, + "loss": 0.4384, + "step": 756 + }, + { + "epoch": 1.24, + "grad_norm": 0.24733024835586548, + "learning_rate": 0.00011479338842975205, + "loss": 0.2883, + "step": 757 + }, + { + "epoch": 1.24, + "grad_norm": 0.2824367880821228, + "learning_rate": 0.00011454545454545453, + "loss": 0.3141, + "step": 758 + }, + { + "epoch": 1.24, + "grad_norm": 0.27844521403312683, + "learning_rate": 0.000114297520661157, + "loss": 0.3327, + "step": 759 + }, + { + "epoch": 1.24, + "grad_norm": 0.26114732027053833, + "learning_rate": 0.0001140495867768595, + "loss": 0.4071, + "step": 760 + }, + { + "epoch": 1.25, + "grad_norm": 0.34284186363220215, + "learning_rate": 0.00011380165289256197, + "loss": 0.4619, + "step": 761 + }, + { + "epoch": 1.25, + "grad_norm": 0.2463303506374359, + "learning_rate": 0.00011355371900826446, + "loss": 0.3038, + "step": 762 + }, + { + "epoch": 1.25, + "grad_norm": 0.26452890038490295, + "learning_rate": 0.00011330578512396693, + "loss": 0.3603, + "step": 763 + }, + { + "epoch": 1.25, + "grad_norm": 0.27888497710227966, + "learning_rate": 0.00011305785123966941, + "loss": 0.5109, + "step": 764 + }, + { + "epoch": 1.25, + "grad_norm": 0.3039766252040863, + "learning_rate": 0.00011280991735537189, + "loss": 0.5377, + "step": 765 + }, + { + "epoch": 1.25, + "grad_norm": 0.28995901346206665, + "learning_rate": 0.00011256198347107436, + "loss": 0.4797, + "step": 766 + }, + { + "epoch": 1.26, + "grad_norm": 0.3420790135860443, + "learning_rate": 0.00011231404958677686, + "loss": 0.5209, + "step": 767 + }, + { + "epoch": 1.26, + "grad_norm": 0.33119046688079834, + "learning_rate": 0.00011206611570247933, + "loss": 0.3709, + "step": 768 + }, + { + "epoch": 1.26, + "grad_norm": 0.3408135175704956, + "learning_rate": 0.00011181818181818181, + "loss": 0.4389, + "step": 769 + }, + { + "epoch": 1.26, + "grad_norm": 0.29120129346847534, + "learning_rate": 0.00011157024793388428, + "loss": 0.4327, + "step": 770 + }, + { + "epoch": 1.26, + "grad_norm": 0.32718029618263245, + "learning_rate": 0.00011132231404958677, + "loss": 0.4859, + "step": 771 + }, + { + "epoch": 1.26, + "grad_norm": 0.34422147274017334, + "learning_rate": 0.00011107438016528924, + "loss": 0.5184, + "step": 772 + }, + { + "epoch": 1.27, + "grad_norm": 0.330323189496994, + "learning_rate": 0.00011082644628099172, + "loss": 0.4322, + "step": 773 + }, + { + "epoch": 1.27, + "grad_norm": 0.3218427002429962, + "learning_rate": 0.00011057851239669422, + "loss": 0.4129, + "step": 774 + }, + { + "epoch": 1.27, + "grad_norm": 0.2976725995540619, + "learning_rate": 0.00011033057851239669, + "loss": 0.5039, + "step": 775 + }, + { + "epoch": 1.27, + "grad_norm": 0.32841789722442627, + "learning_rate": 0.00011008264462809917, + "loss": 0.4718, + "step": 776 + }, + { + "epoch": 1.27, + "grad_norm": 0.32977914810180664, + "learning_rate": 0.00010983471074380164, + "loss": 0.4248, + "step": 777 + }, + { + "epoch": 1.27, + "grad_norm": 0.2632751166820526, + "learning_rate": 0.00010958677685950413, + "loss": 0.3458, + "step": 778 + }, + { + "epoch": 1.28, + "grad_norm": 0.33028510212898254, + "learning_rate": 0.0001093388429752066, + "loss": 0.4884, + "step": 779 + }, + { + "epoch": 1.28, + "grad_norm": 0.30288752913475037, + "learning_rate": 0.00010909090909090908, + "loss": 0.3776, + "step": 780 + }, + { + "epoch": 1.28, + "grad_norm": 0.32292476296424866, + "learning_rate": 0.00010884297520661155, + "loss": 0.392, + "step": 781 + }, + { + "epoch": 1.28, + "grad_norm": 0.31956765055656433, + "learning_rate": 0.00010859504132231405, + "loss": 0.3308, + "step": 782 + }, + { + "epoch": 1.28, + "grad_norm": 0.280553936958313, + "learning_rate": 0.00010834710743801652, + "loss": 0.5806, + "step": 783 + }, + { + "epoch": 1.28, + "grad_norm": 0.35859328508377075, + "learning_rate": 0.000108099173553719, + "loss": 0.5059, + "step": 784 + }, + { + "epoch": 1.29, + "grad_norm": 0.2944432497024536, + "learning_rate": 0.00010785123966942148, + "loss": 0.5132, + "step": 785 + }, + { + "epoch": 1.29, + "grad_norm": 0.27504968643188477, + "learning_rate": 0.00010760330578512395, + "loss": 0.3741, + "step": 786 + }, + { + "epoch": 1.29, + "grad_norm": 0.29401764273643494, + "learning_rate": 0.00010735537190082644, + "loss": 0.4992, + "step": 787 + }, + { + "epoch": 1.29, + "grad_norm": 0.30569151043891907, + "learning_rate": 0.00010710743801652891, + "loss": 0.5029, + "step": 788 + }, + { + "epoch": 1.29, + "grad_norm": 0.28654801845550537, + "learning_rate": 0.0001068595041322314, + "loss": 0.4618, + "step": 789 + }, + { + "epoch": 1.29, + "grad_norm": 0.26424363255500793, + "learning_rate": 0.00010661157024793387, + "loss": 0.3929, + "step": 790 + }, + { + "epoch": 1.3, + "grad_norm": 0.28117212653160095, + "learning_rate": 0.00010636363636363636, + "loss": 0.5116, + "step": 791 + }, + { + "epoch": 1.3, + "grad_norm": 0.28402891755104065, + "learning_rate": 0.00010611570247933883, + "loss": 0.3758, + "step": 792 + }, + { + "epoch": 1.3, + "grad_norm": 0.32903602719306946, + "learning_rate": 0.00010586776859504131, + "loss": 0.3594, + "step": 793 + }, + { + "epoch": 1.3, + "grad_norm": 0.4285104274749756, + "learning_rate": 0.0001056198347107438, + "loss": 0.3007, + "step": 794 + }, + { + "epoch": 1.3, + "grad_norm": 0.27649369835853577, + "learning_rate": 0.00010537190082644627, + "loss": 0.342, + "step": 795 + }, + { + "epoch": 1.3, + "grad_norm": 0.3094039261341095, + "learning_rate": 0.00010512396694214876, + "loss": 0.4452, + "step": 796 + }, + { + "epoch": 1.3, + "grad_norm": 0.32547199726104736, + "learning_rate": 0.00010487603305785123, + "loss": 0.4274, + "step": 797 + }, + { + "epoch": 1.31, + "grad_norm": 0.30244141817092896, + "learning_rate": 0.00010462809917355372, + "loss": 0.393, + "step": 798 + }, + { + "epoch": 1.31, + "grad_norm": 0.3018583655357361, + "learning_rate": 0.00010438016528925619, + "loss": 0.4012, + "step": 799 + }, + { + "epoch": 1.31, + "grad_norm": 0.36397960782051086, + "learning_rate": 0.00010413223140495867, + "loss": 0.5231, + "step": 800 + }, + { + "epoch": 1.31, + "grad_norm": 0.3178517520427704, + "learning_rate": 0.00010388429752066114, + "loss": 0.4036, + "step": 801 + }, + { + "epoch": 1.31, + "grad_norm": 0.34640219807624817, + "learning_rate": 0.00010363636363636362, + "loss": 0.4717, + "step": 802 + }, + { + "epoch": 1.31, + "grad_norm": 0.302775114774704, + "learning_rate": 0.0001033884297520661, + "loss": 0.4207, + "step": 803 + }, + { + "epoch": 1.32, + "grad_norm": 0.30845245718955994, + "learning_rate": 0.00010314049586776859, + "loss": 0.3976, + "step": 804 + }, + { + "epoch": 1.32, + "grad_norm": 0.2689266502857208, + "learning_rate": 0.00010289256198347107, + "loss": 0.3777, + "step": 805 + }, + { + "epoch": 1.32, + "grad_norm": 0.33539149165153503, + "learning_rate": 0.00010264462809917354, + "loss": 0.3896, + "step": 806 + }, + { + "epoch": 1.32, + "grad_norm": 0.2548604905605316, + "learning_rate": 0.00010239669421487603, + "loss": 0.4026, + "step": 807 + }, + { + "epoch": 1.32, + "grad_norm": 0.5050720572471619, + "learning_rate": 0.0001021487603305785, + "loss": 0.4008, + "step": 808 + }, + { + "epoch": 1.32, + "grad_norm": 0.2518717646598816, + "learning_rate": 0.00010190082644628098, + "loss": 0.348, + "step": 809 + }, + { + "epoch": 1.33, + "grad_norm": 0.39397895336151123, + "learning_rate": 0.00010165289256198345, + "loss": 0.5369, + "step": 810 + }, + { + "epoch": 1.33, + "grad_norm": 0.3471471965312958, + "learning_rate": 0.00010140495867768595, + "loss": 0.5272, + "step": 811 + }, + { + "epoch": 1.33, + "grad_norm": 0.4147883355617523, + "learning_rate": 0.00010115702479338842, + "loss": 0.427, + "step": 812 + }, + { + "epoch": 1.33, + "grad_norm": 0.2932160794734955, + "learning_rate": 0.0001009090909090909, + "loss": 0.3274, + "step": 813 + }, + { + "epoch": 1.33, + "grad_norm": 0.28647059202194214, + "learning_rate": 0.00010066115702479339, + "loss": 0.3346, + "step": 814 + }, + { + "epoch": 1.33, + "grad_norm": 0.28154057264328003, + "learning_rate": 0.00010041322314049586, + "loss": 0.3785, + "step": 815 + }, + { + "epoch": 1.34, + "grad_norm": 0.25706711411476135, + "learning_rate": 0.00010016528925619834, + "loss": 0.3261, + "step": 816 + }, + { + "epoch": 1.34, + "grad_norm": 0.3318668603897095, + "learning_rate": 9.991735537190081e-05, + "loss": 0.4362, + "step": 817 + }, + { + "epoch": 1.34, + "grad_norm": 0.33185282349586487, + "learning_rate": 9.966942148760331e-05, + "loss": 0.5219, + "step": 818 + }, + { + "epoch": 1.34, + "grad_norm": 0.2683846056461334, + "learning_rate": 9.942148760330578e-05, + "loss": 0.3657, + "step": 819 + }, + { + "epoch": 1.34, + "grad_norm": 0.2643420100212097, + "learning_rate": 9.917355371900826e-05, + "loss": 0.4697, + "step": 820 + }, + { + "epoch": 1.34, + "grad_norm": 0.32440856099128723, + "learning_rate": 9.892561983471073e-05, + "loss": 0.5572, + "step": 821 + }, + { + "epoch": 1.35, + "grad_norm": 0.22183597087860107, + "learning_rate": 9.867768595041321e-05, + "loss": 0.3379, + "step": 822 + }, + { + "epoch": 1.35, + "grad_norm": 0.26266101002693176, + "learning_rate": 9.842975206611568e-05, + "loss": 0.439, + "step": 823 + }, + { + "epoch": 1.35, + "grad_norm": 0.2978360950946808, + "learning_rate": 9.818181818181817e-05, + "loss": 0.4654, + "step": 824 + }, + { + "epoch": 1.35, + "grad_norm": 0.2713984251022339, + "learning_rate": 9.793388429752067e-05, + "loss": 0.2983, + "step": 825 + }, + { + "epoch": 1.35, + "grad_norm": 0.2561984956264496, + "learning_rate": 9.768595041322314e-05, + "loss": 0.3381, + "step": 826 + }, + { + "epoch": 1.35, + "grad_norm": 0.2766323983669281, + "learning_rate": 9.743801652892562e-05, + "loss": 0.4167, + "step": 827 + }, + { + "epoch": 1.36, + "grad_norm": 0.33810022473335266, + "learning_rate": 9.719008264462809e-05, + "loss": 0.3793, + "step": 828 + }, + { + "epoch": 1.36, + "grad_norm": 0.3332251310348511, + "learning_rate": 9.694214876033057e-05, + "loss": 0.5517, + "step": 829 + }, + { + "epoch": 1.36, + "grad_norm": 0.2713959515094757, + "learning_rate": 9.669421487603304e-05, + "loss": 0.3583, + "step": 830 + }, + { + "epoch": 1.36, + "grad_norm": 0.2778157889842987, + "learning_rate": 9.644628099173553e-05, + "loss": 0.3089, + "step": 831 + }, + { + "epoch": 1.36, + "grad_norm": 0.33538392186164856, + "learning_rate": 9.6198347107438e-05, + "loss": 0.3776, + "step": 832 + }, + { + "epoch": 1.36, + "grad_norm": 0.32728123664855957, + "learning_rate": 9.59504132231405e-05, + "loss": 0.434, + "step": 833 + }, + { + "epoch": 1.37, + "grad_norm": 0.30630162358283997, + "learning_rate": 9.570247933884298e-05, + "loss": 0.3913, + "step": 834 + }, + { + "epoch": 1.37, + "grad_norm": 0.2960034906864166, + "learning_rate": 9.545454545454545e-05, + "loss": 0.4368, + "step": 835 + }, + { + "epoch": 1.37, + "grad_norm": 0.35711923241615295, + "learning_rate": 9.520661157024793e-05, + "loss": 0.399, + "step": 836 + }, + { + "epoch": 1.37, + "grad_norm": 0.30195897817611694, + "learning_rate": 9.49586776859504e-05, + "loss": 0.4421, + "step": 837 + }, + { + "epoch": 1.37, + "grad_norm": 0.3220643401145935, + "learning_rate": 9.471074380165288e-05, + "loss": 0.3441, + "step": 838 + }, + { + "epoch": 1.37, + "grad_norm": 0.3709239661693573, + "learning_rate": 9.446280991735535e-05, + "loss": 0.4095, + "step": 839 + }, + { + "epoch": 1.38, + "grad_norm": 0.40360063314437866, + "learning_rate": 9.421487603305785e-05, + "loss": 0.5692, + "step": 840 + }, + { + "epoch": 1.38, + "grad_norm": 0.32428041100502014, + "learning_rate": 9.396694214876032e-05, + "loss": 0.4306, + "step": 841 + }, + { + "epoch": 1.38, + "grad_norm": 0.2750518321990967, + "learning_rate": 9.37190082644628e-05, + "loss": 0.3905, + "step": 842 + }, + { + "epoch": 1.38, + "grad_norm": 0.331478476524353, + "learning_rate": 9.347107438016528e-05, + "loss": 0.6008, + "step": 843 + }, + { + "epoch": 1.38, + "grad_norm": 0.3165242671966553, + "learning_rate": 9.322314049586776e-05, + "loss": 0.4624, + "step": 844 + }, + { + "epoch": 1.38, + "grad_norm": 0.26457470655441284, + "learning_rate": 9.297520661157024e-05, + "loss": 0.4462, + "step": 845 + }, + { + "epoch": 1.39, + "grad_norm": 0.3557126522064209, + "learning_rate": 9.272727272727271e-05, + "loss": 0.5737, + "step": 846 + }, + { + "epoch": 1.39, + "grad_norm": 0.3306926488876343, + "learning_rate": 9.247933884297521e-05, + "loss": 0.4597, + "step": 847 + }, + { + "epoch": 1.39, + "grad_norm": 0.24906127154827118, + "learning_rate": 9.223140495867768e-05, + "loss": 0.378, + "step": 848 + }, + { + "epoch": 1.39, + "grad_norm": 0.29440054297447205, + "learning_rate": 9.198347107438016e-05, + "loss": 0.4562, + "step": 849 + }, + { + "epoch": 1.39, + "grad_norm": 0.34878161549568176, + "learning_rate": 9.173553719008263e-05, + "loss": 0.4546, + "step": 850 + }, + { + "epoch": 1.39, + "grad_norm": 0.3725307583808899, + "learning_rate": 9.148760330578512e-05, + "loss": 0.4119, + "step": 851 + }, + { + "epoch": 1.4, + "grad_norm": 0.30648747086524963, + "learning_rate": 9.123966942148759e-05, + "loss": 0.4428, + "step": 852 + }, + { + "epoch": 1.4, + "grad_norm": 0.2755535840988159, + "learning_rate": 9.099173553719007e-05, + "loss": 0.3592, + "step": 853 + }, + { + "epoch": 1.4, + "grad_norm": 0.2802577614784241, + "learning_rate": 9.074380165289255e-05, + "loss": 0.472, + "step": 854 + }, + { + "epoch": 1.4, + "grad_norm": 0.28871360421180725, + "learning_rate": 9.049586776859504e-05, + "loss": 0.4532, + "step": 855 + }, + { + "epoch": 1.4, + "grad_norm": 0.37071362137794495, + "learning_rate": 9.024793388429752e-05, + "loss": 0.3426, + "step": 856 + }, + { + "epoch": 1.4, + "grad_norm": 0.30081430077552795, + "learning_rate": 8.999999999999999e-05, + "loss": 0.4069, + "step": 857 + }, + { + "epoch": 1.4, + "grad_norm": 0.3186596930027008, + "learning_rate": 8.975206611570248e-05, + "loss": 0.4997, + "step": 858 + }, + { + "epoch": 1.41, + "grad_norm": 0.286479115486145, + "learning_rate": 8.950413223140495e-05, + "loss": 0.3902, + "step": 859 + }, + { + "epoch": 1.41, + "grad_norm": 0.3457258939743042, + "learning_rate": 8.925619834710743e-05, + "loss": 0.4339, + "step": 860 + }, + { + "epoch": 1.41, + "grad_norm": 0.30513113737106323, + "learning_rate": 8.90082644628099e-05, + "loss": 0.3414, + "step": 861 + }, + { + "epoch": 1.41, + "grad_norm": 0.30697953701019287, + "learning_rate": 8.876033057851238e-05, + "loss": 0.4657, + "step": 862 + }, + { + "epoch": 1.41, + "grad_norm": 0.3395203649997711, + "learning_rate": 8.851239669421488e-05, + "loss": 0.3945, + "step": 863 + }, + { + "epoch": 1.41, + "grad_norm": 0.43322789669036865, + "learning_rate": 8.826446280991735e-05, + "loss": 0.5337, + "step": 864 + }, + { + "epoch": 1.42, + "grad_norm": 0.3421814739704132, + "learning_rate": 8.801652892561983e-05, + "loss": 0.4481, + "step": 865 + }, + { + "epoch": 1.42, + "grad_norm": 0.24497461318969727, + "learning_rate": 8.77685950413223e-05, + "loss": 0.4199, + "step": 866 + }, + { + "epoch": 1.42, + "grad_norm": 0.3835270404815674, + "learning_rate": 8.752066115702479e-05, + "loss": 0.5534, + "step": 867 + }, + { + "epoch": 1.42, + "grad_norm": 0.3144569396972656, + "learning_rate": 8.727272727272726e-05, + "loss": 0.4563, + "step": 868 + }, + { + "epoch": 1.42, + "grad_norm": 0.2757865786552429, + "learning_rate": 8.702479338842974e-05, + "loss": 0.4241, + "step": 869 + }, + { + "epoch": 1.42, + "grad_norm": 0.28413090109825134, + "learning_rate": 8.677685950413221e-05, + "loss": 0.3484, + "step": 870 + }, + { + "epoch": 1.43, + "grad_norm": 0.27918362617492676, + "learning_rate": 8.652892561983471e-05, + "loss": 0.4133, + "step": 871 + }, + { + "epoch": 1.43, + "grad_norm": 0.3901917040348053, + "learning_rate": 8.628099173553718e-05, + "loss": 0.4755, + "step": 872 + }, + { + "epoch": 1.43, + "grad_norm": 0.34810692071914673, + "learning_rate": 8.603305785123966e-05, + "loss": 0.4516, + "step": 873 + }, + { + "epoch": 1.43, + "grad_norm": 0.3317393958568573, + "learning_rate": 8.578512396694215e-05, + "loss": 0.4995, + "step": 874 + }, + { + "epoch": 1.43, + "grad_norm": 0.26235052943229675, + "learning_rate": 8.553719008264462e-05, + "loss": 0.3348, + "step": 875 + }, + { + "epoch": 1.43, + "grad_norm": 0.2735447585582733, + "learning_rate": 8.52892561983471e-05, + "loss": 0.2932, + "step": 876 + }, + { + "epoch": 1.44, + "grad_norm": 0.30968329310417175, + "learning_rate": 8.504132231404957e-05, + "loss": 0.3783, + "step": 877 + }, + { + "epoch": 1.44, + "grad_norm": 0.30193984508514404, + "learning_rate": 8.479338842975207e-05, + "loss": 0.4357, + "step": 878 + }, + { + "epoch": 1.44, + "grad_norm": 0.3407258987426758, + "learning_rate": 8.454545454545454e-05, + "loss": 0.4821, + "step": 879 + }, + { + "epoch": 1.44, + "grad_norm": 0.28090009093284607, + "learning_rate": 8.429752066115702e-05, + "loss": 0.4158, + "step": 880 + }, + { + "epoch": 1.44, + "grad_norm": 0.2898884415626526, + "learning_rate": 8.404958677685949e-05, + "loss": 0.3091, + "step": 881 + }, + { + "epoch": 1.44, + "grad_norm": 0.31658637523651123, + "learning_rate": 8.380165289256197e-05, + "loss": 0.3773, + "step": 882 + }, + { + "epoch": 1.45, + "grad_norm": 0.2722189724445343, + "learning_rate": 8.355371900826446e-05, + "loss": 0.4483, + "step": 883 + }, + { + "epoch": 1.45, + "grad_norm": 0.23621954023838043, + "learning_rate": 8.330578512396693e-05, + "loss": 0.3112, + "step": 884 + }, + { + "epoch": 1.45, + "grad_norm": 0.3659461438655853, + "learning_rate": 8.305785123966942e-05, + "loss": 0.4507, + "step": 885 + }, + { + "epoch": 1.45, + "grad_norm": 0.3253099322319031, + "learning_rate": 8.28099173553719e-05, + "loss": 0.4854, + "step": 886 + }, + { + "epoch": 1.45, + "grad_norm": 0.3201637864112854, + "learning_rate": 8.256198347107438e-05, + "loss": 0.5687, + "step": 887 + }, + { + "epoch": 1.45, + "grad_norm": 0.4112270772457123, + "learning_rate": 8.231404958677685e-05, + "loss": 0.3742, + "step": 888 + }, + { + "epoch": 1.46, + "grad_norm": 0.3146194517612457, + "learning_rate": 8.206611570247933e-05, + "loss": 0.4869, + "step": 889 + }, + { + "epoch": 1.46, + "grad_norm": 0.34321263432502747, + "learning_rate": 8.18181818181818e-05, + "loss": 0.5154, + "step": 890 + }, + { + "epoch": 1.46, + "grad_norm": 0.2986968159675598, + "learning_rate": 8.157024793388429e-05, + "loss": 0.647, + "step": 891 + }, + { + "epoch": 1.46, + "grad_norm": 0.3427133858203888, + "learning_rate": 8.132231404958676e-05, + "loss": 0.3912, + "step": 892 + }, + { + "epoch": 1.46, + "grad_norm": 0.3434309661388397, + "learning_rate": 8.107438016528925e-05, + "loss": 0.51, + "step": 893 + }, + { + "epoch": 1.46, + "grad_norm": 0.32024991512298584, + "learning_rate": 8.082644628099174e-05, + "loss": 0.387, + "step": 894 + }, + { + "epoch": 1.47, + "grad_norm": 0.2961815595626831, + "learning_rate": 8.057851239669421e-05, + "loss": 0.3909, + "step": 895 + }, + { + "epoch": 1.47, + "grad_norm": 0.3219030201435089, + "learning_rate": 8.033057851239669e-05, + "loss": 0.3911, + "step": 896 + }, + { + "epoch": 1.47, + "grad_norm": 0.2776000201702118, + "learning_rate": 8.008264462809916e-05, + "loss": 0.3625, + "step": 897 + }, + { + "epoch": 1.47, + "grad_norm": 0.31484290957450867, + "learning_rate": 7.983471074380164e-05, + "loss": 0.6162, + "step": 898 + }, + { + "epoch": 1.47, + "grad_norm": 0.2789134085178375, + "learning_rate": 7.958677685950411e-05, + "loss": 0.3199, + "step": 899 + }, + { + "epoch": 1.47, + "grad_norm": 0.27821627259254456, + "learning_rate": 7.933884297520661e-05, + "loss": 0.4295, + "step": 900 + }, + { + "epoch": 1.48, + "grad_norm": 0.3022254705429077, + "learning_rate": 7.909090909090908e-05, + "loss": 0.309, + "step": 901 + }, + { + "epoch": 1.48, + "grad_norm": 0.24830293655395508, + "learning_rate": 7.884297520661157e-05, + "loss": 0.3833, + "step": 902 + }, + { + "epoch": 1.48, + "grad_norm": 0.31184327602386475, + "learning_rate": 7.859504132231405e-05, + "loss": 0.3715, + "step": 903 + }, + { + "epoch": 1.48, + "grad_norm": 0.2993053197860718, + "learning_rate": 7.834710743801652e-05, + "loss": 0.3825, + "step": 904 + }, + { + "epoch": 1.48, + "grad_norm": 0.3385005295276642, + "learning_rate": 7.8099173553719e-05, + "loss": 0.4868, + "step": 905 + }, + { + "epoch": 1.48, + "grad_norm": 0.26812323927879333, + "learning_rate": 7.785123966942147e-05, + "loss": 0.2925, + "step": 906 + }, + { + "epoch": 1.49, + "grad_norm": 0.3275848925113678, + "learning_rate": 7.760330578512397e-05, + "loss": 0.3657, + "step": 907 + }, + { + "epoch": 1.49, + "grad_norm": 0.2972089350223541, + "learning_rate": 7.735537190082644e-05, + "loss": 0.4396, + "step": 908 + }, + { + "epoch": 1.49, + "grad_norm": 0.27619728446006775, + "learning_rate": 7.710743801652892e-05, + "loss": 0.3946, + "step": 909 + }, + { + "epoch": 1.49, + "grad_norm": 0.30436667799949646, + "learning_rate": 7.68595041322314e-05, + "loss": 0.4177, + "step": 910 + }, + { + "epoch": 1.49, + "grad_norm": 0.2652393877506256, + "learning_rate": 7.661157024793388e-05, + "loss": 0.3165, + "step": 911 + }, + { + "epoch": 1.49, + "grad_norm": 0.28303712606430054, + "learning_rate": 7.636363636363635e-05, + "loss": 0.4829, + "step": 912 + }, + { + "epoch": 1.49, + "grad_norm": 0.33964964747428894, + "learning_rate": 7.611570247933883e-05, + "loss": 0.5043, + "step": 913 + }, + { + "epoch": 1.5, + "grad_norm": 0.2591302692890167, + "learning_rate": 7.586776859504133e-05, + "loss": 0.3814, + "step": 914 + }, + { + "epoch": 1.5, + "grad_norm": 0.3488747179508209, + "learning_rate": 7.56198347107438e-05, + "loss": 0.5233, + "step": 915 + }, + { + "epoch": 1.5, + "grad_norm": 0.29015597701072693, + "learning_rate": 7.537190082644628e-05, + "loss": 0.4672, + "step": 916 + }, + { + "epoch": 1.5, + "grad_norm": 0.31618839502334595, + "learning_rate": 7.512396694214875e-05, + "loss": 0.4538, + "step": 917 + }, + { + "epoch": 1.5, + "grad_norm": 0.35049545764923096, + "learning_rate": 7.487603305785124e-05, + "loss": 0.4089, + "step": 918 + }, + { + "epoch": 1.5, + "grad_norm": 0.34093132615089417, + "learning_rate": 7.462809917355372e-05, + "loss": 0.4306, + "step": 919 + }, + { + "epoch": 1.51, + "grad_norm": 0.30601584911346436, + "learning_rate": 7.438016528925619e-05, + "loss": 0.4396, + "step": 920 + }, + { + "epoch": 1.51, + "grad_norm": 0.45013612508773804, + "learning_rate": 7.413223140495867e-05, + "loss": 0.4477, + "step": 921 + }, + { + "epoch": 1.51, + "grad_norm": 0.30486834049224854, + "learning_rate": 7.388429752066116e-05, + "loss": 0.3777, + "step": 922 + }, + { + "epoch": 1.51, + "grad_norm": 0.3926061689853668, + "learning_rate": 7.363636363636363e-05, + "loss": 0.3532, + "step": 923 + }, + { + "epoch": 1.51, + "grad_norm": 0.3843371272087097, + "learning_rate": 7.338842975206611e-05, + "loss": 0.5182, + "step": 924 + }, + { + "epoch": 1.51, + "grad_norm": 0.30922451615333557, + "learning_rate": 7.314049586776858e-05, + "loss": 0.4361, + "step": 925 + }, + { + "epoch": 1.52, + "grad_norm": 0.3367323875427246, + "learning_rate": 7.289256198347106e-05, + "loss": 0.3809, + "step": 926 + }, + { + "epoch": 1.52, + "grad_norm": 0.39369019865989685, + "learning_rate": 7.264462809917355e-05, + "loss": 0.3623, + "step": 927 + }, + { + "epoch": 1.52, + "grad_norm": 0.3159162104129791, + "learning_rate": 7.239669421487603e-05, + "loss": 0.5059, + "step": 928 + }, + { + "epoch": 1.52, + "grad_norm": 0.34716740250587463, + "learning_rate": 7.214876033057851e-05, + "loss": 0.4201, + "step": 929 + }, + { + "epoch": 1.52, + "grad_norm": 0.20480923354625702, + "learning_rate": 7.190082644628098e-05, + "loss": 0.2699, + "step": 930 + }, + { + "epoch": 1.52, + "grad_norm": 0.3518913686275482, + "learning_rate": 7.165289256198347e-05, + "loss": 0.5337, + "step": 931 + }, + { + "epoch": 1.53, + "grad_norm": 0.28605952858924866, + "learning_rate": 7.140495867768594e-05, + "loss": 0.44, + "step": 932 + }, + { + "epoch": 1.53, + "grad_norm": 0.28229033946990967, + "learning_rate": 7.115702479338842e-05, + "loss": 0.3534, + "step": 933 + }, + { + "epoch": 1.53, + "grad_norm": 0.3456754684448242, + "learning_rate": 7.09090909090909e-05, + "loss": 0.3952, + "step": 934 + }, + { + "epoch": 1.53, + "grad_norm": 0.27707159519195557, + "learning_rate": 7.066115702479338e-05, + "loss": 0.3667, + "step": 935 + }, + { + "epoch": 1.53, + "grad_norm": 0.2811780273914337, + "learning_rate": 7.041322314049586e-05, + "loss": 0.3954, + "step": 936 + }, + { + "epoch": 1.53, + "grad_norm": 0.3099793493747711, + "learning_rate": 7.016528925619834e-05, + "loss": 0.441, + "step": 937 + }, + { + "epoch": 1.54, + "grad_norm": 0.4153590500354767, + "learning_rate": 6.991735537190083e-05, + "loss": 0.4462, + "step": 938 + }, + { + "epoch": 1.54, + "grad_norm": 0.2945801615715027, + "learning_rate": 6.96694214876033e-05, + "loss": 0.4535, + "step": 939 + }, + { + "epoch": 1.54, + "grad_norm": 0.2930592894554138, + "learning_rate": 6.942148760330578e-05, + "loss": 0.5566, + "step": 940 + }, + { + "epoch": 1.54, + "grad_norm": 0.3034913241863251, + "learning_rate": 6.917355371900826e-05, + "loss": 0.4695, + "step": 941 + }, + { + "epoch": 1.54, + "grad_norm": 0.3054913878440857, + "learning_rate": 6.892561983471073e-05, + "loss": 0.3921, + "step": 942 + }, + { + "epoch": 1.54, + "grad_norm": 0.3297981917858124, + "learning_rate": 6.867768595041322e-05, + "loss": 0.5057, + "step": 943 + }, + { + "epoch": 1.55, + "grad_norm": 0.23640452325344086, + "learning_rate": 6.84297520661157e-05, + "loss": 0.329, + "step": 944 + }, + { + "epoch": 1.55, + "grad_norm": 0.2970188856124878, + "learning_rate": 6.818181818181817e-05, + "loss": 0.4376, + "step": 945 + }, + { + "epoch": 1.55, + "grad_norm": 0.3243064880371094, + "learning_rate": 6.793388429752065e-05, + "loss": 0.4922, + "step": 946 + }, + { + "epoch": 1.55, + "grad_norm": 0.4473859667778015, + "learning_rate": 6.768595041322312e-05, + "loss": 0.5245, + "step": 947 + }, + { + "epoch": 1.55, + "grad_norm": 0.2901310622692108, + "learning_rate": 6.743801652892562e-05, + "loss": 0.4996, + "step": 948 + }, + { + "epoch": 1.55, + "grad_norm": 0.3633457124233246, + "learning_rate": 6.719008264462809e-05, + "loss": 0.4669, + "step": 949 + }, + { + "epoch": 1.56, + "grad_norm": 0.33570581674575806, + "learning_rate": 6.694214876033058e-05, + "loss": 0.404, + "step": 950 + }, + { + "epoch": 1.56, + "grad_norm": 0.26466354727745056, + "learning_rate": 6.669421487603306e-05, + "loss": 0.2881, + "step": 951 + }, + { + "epoch": 1.56, + "grad_norm": 0.29028353095054626, + "learning_rate": 6.644628099173553e-05, + "loss": 0.3607, + "step": 952 + }, + { + "epoch": 1.56, + "grad_norm": 0.2878669798374176, + "learning_rate": 6.619834710743801e-05, + "loss": 0.4415, + "step": 953 + }, + { + "epoch": 1.56, + "grad_norm": 0.33260804414749146, + "learning_rate": 6.595041322314048e-05, + "loss": 0.4424, + "step": 954 + }, + { + "epoch": 1.56, + "grad_norm": 0.3135119378566742, + "learning_rate": 6.570247933884297e-05, + "loss": 0.4276, + "step": 955 + }, + { + "epoch": 1.57, + "grad_norm": 0.2714795470237732, + "learning_rate": 6.545454545454545e-05, + "loss": 0.2789, + "step": 956 + }, + { + "epoch": 1.57, + "grad_norm": 0.3564438819885254, + "learning_rate": 6.520661157024792e-05, + "loss": 0.4683, + "step": 957 + }, + { + "epoch": 1.57, + "grad_norm": 0.3303399682044983, + "learning_rate": 6.495867768595042e-05, + "loss": 0.4657, + "step": 958 + }, + { + "epoch": 1.57, + "grad_norm": 0.30086350440979004, + "learning_rate": 6.471074380165289e-05, + "loss": 0.3296, + "step": 959 + }, + { + "epoch": 1.57, + "grad_norm": 0.34699100255966187, + "learning_rate": 6.446280991735537e-05, + "loss": 0.3543, + "step": 960 + }, + { + "epoch": 1.57, + "grad_norm": 0.326579213142395, + "learning_rate": 6.421487603305784e-05, + "loss": 0.4001, + "step": 961 + }, + { + "epoch": 1.58, + "grad_norm": 0.3462665379047394, + "learning_rate": 6.396694214876032e-05, + "loss": 0.3999, + "step": 962 + }, + { + "epoch": 1.58, + "grad_norm": 0.3408821225166321, + "learning_rate": 6.371900826446281e-05, + "loss": 0.3614, + "step": 963 + }, + { + "epoch": 1.58, + "grad_norm": 0.3061428666114807, + "learning_rate": 6.347107438016528e-05, + "loss": 0.4127, + "step": 964 + }, + { + "epoch": 1.58, + "grad_norm": 0.30745938420295715, + "learning_rate": 6.322314049586776e-05, + "loss": 0.3965, + "step": 965 + }, + { + "epoch": 1.58, + "grad_norm": 0.33782872557640076, + "learning_rate": 6.297520661157025e-05, + "loss": 0.5026, + "step": 966 + }, + { + "epoch": 1.58, + "grad_norm": 0.3501698076725006, + "learning_rate": 6.272727272727272e-05, + "loss": 0.4731, + "step": 967 + }, + { + "epoch": 1.58, + "grad_norm": 0.3578520119190216, + "learning_rate": 6.24793388429752e-05, + "loss": 0.4302, + "step": 968 + }, + { + "epoch": 1.59, + "grad_norm": 0.30132660269737244, + "learning_rate": 6.223140495867768e-05, + "loss": 0.3784, + "step": 969 + }, + { + "epoch": 1.59, + "grad_norm": 0.29198774695396423, + "learning_rate": 6.198347107438017e-05, + "loss": 0.396, + "step": 970 + }, + { + "epoch": 1.59, + "grad_norm": 0.3028549551963806, + "learning_rate": 6.173553719008264e-05, + "loss": 0.3531, + "step": 971 + }, + { + "epoch": 1.59, + "grad_norm": 0.3193860352039337, + "learning_rate": 6.148760330578512e-05, + "loss": 0.5261, + "step": 972 + }, + { + "epoch": 1.59, + "grad_norm": 0.330228716135025, + "learning_rate": 6.12396694214876e-05, + "loss": 0.3853, + "step": 973 + }, + { + "epoch": 1.59, + "grad_norm": 0.2856347858905792, + "learning_rate": 6.0991735537190074e-05, + "loss": 0.4543, + "step": 974 + }, + { + "epoch": 1.6, + "grad_norm": 0.3663886487483978, + "learning_rate": 6.074380165289256e-05, + "loss": 0.3821, + "step": 975 + }, + { + "epoch": 1.6, + "grad_norm": 0.3297857642173767, + "learning_rate": 6.0495867768595034e-05, + "loss": 0.4504, + "step": 976 + }, + { + "epoch": 1.6, + "grad_norm": 0.29853883385658264, + "learning_rate": 6.024793388429751e-05, + "loss": 0.3528, + "step": 977 + }, + { + "epoch": 1.6, + "grad_norm": 0.3246425986289978, + "learning_rate": 5.9999999999999995e-05, + "loss": 0.3986, + "step": 978 + }, + { + "epoch": 1.6, + "grad_norm": 0.3537238836288452, + "learning_rate": 5.975206611570248e-05, + "loss": 0.3776, + "step": 979 + }, + { + "epoch": 1.6, + "grad_norm": 0.2915757894515991, + "learning_rate": 5.9504132231404955e-05, + "loss": 0.2895, + "step": 980 + }, + { + "epoch": 1.61, + "grad_norm": 0.30707284808158875, + "learning_rate": 5.925619834710743e-05, + "loss": 0.3238, + "step": 981 + }, + { + "epoch": 1.61, + "grad_norm": 0.301845520734787, + "learning_rate": 5.900826446280991e-05, + "loss": 0.4031, + "step": 982 + }, + { + "epoch": 1.61, + "grad_norm": 0.24002347886562347, + "learning_rate": 5.876033057851239e-05, + "loss": 0.3477, + "step": 983 + }, + { + "epoch": 1.61, + "grad_norm": 0.3008634150028229, + "learning_rate": 5.851239669421487e-05, + "loss": 0.4595, + "step": 984 + }, + { + "epoch": 1.61, + "grad_norm": 0.32416027784347534, + "learning_rate": 5.8264462809917346e-05, + "loss": 0.403, + "step": 985 + }, + { + "epoch": 1.61, + "grad_norm": 0.3158760368824005, + "learning_rate": 5.801652892561982e-05, + "loss": 0.305, + "step": 986 + }, + { + "epoch": 1.62, + "grad_norm": 0.33743736147880554, + "learning_rate": 5.7768595041322313e-05, + "loss": 0.4867, + "step": 987 + }, + { + "epoch": 1.62, + "grad_norm": 0.3402981460094452, + "learning_rate": 5.752066115702479e-05, + "loss": 0.3982, + "step": 988 + }, + { + "epoch": 1.62, + "grad_norm": 0.3389660716056824, + "learning_rate": 5.727272727272727e-05, + "loss": 0.4311, + "step": 989 + }, + { + "epoch": 1.62, + "grad_norm": 0.47749587893486023, + "learning_rate": 5.702479338842975e-05, + "loss": 0.3775, + "step": 990 + }, + { + "epoch": 1.62, + "grad_norm": 0.27538084983825684, + "learning_rate": 5.677685950413223e-05, + "loss": 0.3568, + "step": 991 + }, + { + "epoch": 1.62, + "grad_norm": 0.33023789525032043, + "learning_rate": 5.6528925619834704e-05, + "loss": 0.4225, + "step": 992 + }, + { + "epoch": 1.63, + "grad_norm": 0.28135445713996887, + "learning_rate": 5.628099173553718e-05, + "loss": 0.3658, + "step": 993 + }, + { + "epoch": 1.63, + "grad_norm": 0.3511416018009186, + "learning_rate": 5.6033057851239665e-05, + "loss": 0.3928, + "step": 994 + }, + { + "epoch": 1.63, + "grad_norm": 0.2987925708293915, + "learning_rate": 5.578512396694214e-05, + "loss": 0.4015, + "step": 995 + }, + { + "epoch": 1.63, + "grad_norm": 0.3340010344982147, + "learning_rate": 5.553719008264462e-05, + "loss": 0.4566, + "step": 996 + }, + { + "epoch": 1.63, + "grad_norm": 0.23461014032363892, + "learning_rate": 5.528925619834711e-05, + "loss": 0.3556, + "step": 997 + }, + { + "epoch": 1.63, + "grad_norm": 0.3425525724887848, + "learning_rate": 5.5041322314049586e-05, + "loss": 0.3736, + "step": 998 + }, + { + "epoch": 1.64, + "grad_norm": 0.33320698142051697, + "learning_rate": 5.479338842975206e-05, + "loss": 0.3926, + "step": 999 + }, + { + "epoch": 1.64, + "grad_norm": 0.26936790347099304, + "learning_rate": 5.454545454545454e-05, + "loss": 0.3587, + "step": 1000 + }, + { + "epoch": 1.64, + "grad_norm": 0.322934091091156, + "learning_rate": 5.429752066115702e-05, + "loss": 0.3119, + "step": 1001 + }, + { + "epoch": 1.64, + "grad_norm": 0.3295484483242035, + "learning_rate": 5.40495867768595e-05, + "loss": 0.3257, + "step": 1002 + }, + { + "epoch": 1.64, + "grad_norm": 0.2893584370613098, + "learning_rate": 5.380165289256198e-05, + "loss": 0.3451, + "step": 1003 + }, + { + "epoch": 1.64, + "grad_norm": 0.3215138912200928, + "learning_rate": 5.3553719008264454e-05, + "loss": 0.4104, + "step": 1004 + }, + { + "epoch": 1.65, + "grad_norm": 0.19545914232730865, + "learning_rate": 5.330578512396694e-05, + "loss": 0.2245, + "step": 1005 + }, + { + "epoch": 1.65, + "grad_norm": 0.2952648103237152, + "learning_rate": 5.3057851239669414e-05, + "loss": 0.3393, + "step": 1006 + }, + { + "epoch": 1.65, + "grad_norm": 0.34105175733566284, + "learning_rate": 5.28099173553719e-05, + "loss": 0.519, + "step": 1007 + }, + { + "epoch": 1.65, + "grad_norm": 0.3435216546058655, + "learning_rate": 5.256198347107438e-05, + "loss": 0.4968, + "step": 1008 + }, + { + "epoch": 1.65, + "grad_norm": 0.29052355885505676, + "learning_rate": 5.231404958677686e-05, + "loss": 0.4419, + "step": 1009 + }, + { + "epoch": 1.65, + "grad_norm": 0.3326230049133301, + "learning_rate": 5.2066115702479335e-05, + "loss": 0.4461, + "step": 1010 + }, + { + "epoch": 1.66, + "grad_norm": 0.35595494508743286, + "learning_rate": 5.181818181818181e-05, + "loss": 0.4886, + "step": 1011 + }, + { + "epoch": 1.66, + "grad_norm": 0.3467525541782379, + "learning_rate": 5.1570247933884295e-05, + "loss": 0.4671, + "step": 1012 + }, + { + "epoch": 1.66, + "grad_norm": 0.29460448026657104, + "learning_rate": 5.132231404958677e-05, + "loss": 0.3872, + "step": 1013 + }, + { + "epoch": 1.66, + "grad_norm": 0.273575097322464, + "learning_rate": 5.107438016528925e-05, + "loss": 0.3603, + "step": 1014 + }, + { + "epoch": 1.66, + "grad_norm": 0.3603818416595459, + "learning_rate": 5.0826446280991726e-05, + "loss": 0.3539, + "step": 1015 + }, + { + "epoch": 1.66, + "grad_norm": 0.31469517946243286, + "learning_rate": 5.057851239669421e-05, + "loss": 0.3988, + "step": 1016 + }, + { + "epoch": 1.67, + "grad_norm": 0.3218969702720642, + "learning_rate": 5.033057851239669e-05, + "loss": 0.4366, + "step": 1017 + }, + { + "epoch": 1.67, + "grad_norm": 0.34077420830726624, + "learning_rate": 5.008264462809917e-05, + "loss": 0.4248, + "step": 1018 + }, + { + "epoch": 1.67, + "grad_norm": 0.322591096162796, + "learning_rate": 4.9834710743801654e-05, + "loss": 0.5081, + "step": 1019 + }, + { + "epoch": 1.67, + "grad_norm": 0.35607361793518066, + "learning_rate": 4.958677685950413e-05, + "loss": 0.3596, + "step": 1020 + }, + { + "epoch": 1.67, + "grad_norm": 0.2865798771381378, + "learning_rate": 4.933884297520661e-05, + "loss": 0.2703, + "step": 1021 + }, + { + "epoch": 1.67, + "grad_norm": 0.30387502908706665, + "learning_rate": 4.9090909090909084e-05, + "loss": 0.3051, + "step": 1022 + }, + { + "epoch": 1.67, + "grad_norm": 0.3474448323249817, + "learning_rate": 4.884297520661157e-05, + "loss": 0.2851, + "step": 1023 + }, + { + "epoch": 1.68, + "grad_norm": 0.3696686625480652, + "learning_rate": 4.8595041322314045e-05, + "loss": 0.4403, + "step": 1024 + }, + { + "epoch": 1.68, + "grad_norm": 0.33602291345596313, + "learning_rate": 4.834710743801652e-05, + "loss": 0.4134, + "step": 1025 + }, + { + "epoch": 1.68, + "grad_norm": 0.27331918478012085, + "learning_rate": 4.8099173553719e-05, + "loss": 0.3303, + "step": 1026 + }, + { + "epoch": 1.68, + "grad_norm": 0.3705825209617615, + "learning_rate": 4.785123966942149e-05, + "loss": 0.3411, + "step": 1027 + }, + { + "epoch": 1.68, + "grad_norm": 0.4541082978248596, + "learning_rate": 4.7603305785123966e-05, + "loss": 0.4263, + "step": 1028 + }, + { + "epoch": 1.68, + "grad_norm": 0.29885897040367126, + "learning_rate": 4.735537190082644e-05, + "loss": 0.5602, + "step": 1029 + }, + { + "epoch": 1.69, + "grad_norm": 0.35169675946235657, + "learning_rate": 4.7107438016528926e-05, + "loss": 0.4409, + "step": 1030 + }, + { + "epoch": 1.69, + "grad_norm": 0.41590291261672974, + "learning_rate": 4.68595041322314e-05, + "loss": 0.4355, + "step": 1031 + }, + { + "epoch": 1.69, + "grad_norm": 0.33613288402557373, + "learning_rate": 4.661157024793388e-05, + "loss": 0.4399, + "step": 1032 + }, + { + "epoch": 1.69, + "grad_norm": 0.3519938886165619, + "learning_rate": 4.6363636363636356e-05, + "loss": 0.4464, + "step": 1033 + }, + { + "epoch": 1.69, + "grad_norm": 0.2981269359588623, + "learning_rate": 4.611570247933884e-05, + "loss": 0.3667, + "step": 1034 + }, + { + "epoch": 1.69, + "grad_norm": 0.32030418515205383, + "learning_rate": 4.586776859504132e-05, + "loss": 0.3759, + "step": 1035 + }, + { + "epoch": 1.7, + "grad_norm": 0.39815372228622437, + "learning_rate": 4.5619834710743794e-05, + "loss": 0.3259, + "step": 1036 + }, + { + "epoch": 1.7, + "grad_norm": 0.33106112480163574, + "learning_rate": 4.537190082644628e-05, + "loss": 0.4985, + "step": 1037 + }, + { + "epoch": 1.7, + "grad_norm": 0.3748137950897217, + "learning_rate": 4.512396694214876e-05, + "loss": 0.5177, + "step": 1038 + }, + { + "epoch": 1.7, + "grad_norm": 0.31328514218330383, + "learning_rate": 4.487603305785124e-05, + "loss": 0.3406, + "step": 1039 + }, + { + "epoch": 1.7, + "grad_norm": 0.35391247272491455, + "learning_rate": 4.4628099173553715e-05, + "loss": 0.4216, + "step": 1040 + }, + { + "epoch": 1.7, + "grad_norm": 0.37352749705314636, + "learning_rate": 4.438016528925619e-05, + "loss": 0.4936, + "step": 1041 + }, + { + "epoch": 1.71, + "grad_norm": 0.20523978769779205, + "learning_rate": 4.4132231404958675e-05, + "loss": 0.2241, + "step": 1042 + }, + { + "epoch": 1.71, + "grad_norm": 0.26052072644233704, + "learning_rate": 4.388429752066115e-05, + "loss": 0.352, + "step": 1043 + }, + { + "epoch": 1.71, + "grad_norm": 0.30189159512519836, + "learning_rate": 4.363636363636363e-05, + "loss": 0.3956, + "step": 1044 + }, + { + "epoch": 1.71, + "grad_norm": 0.28206998109817505, + "learning_rate": 4.3388429752066106e-05, + "loss": 0.3073, + "step": 1045 + }, + { + "epoch": 1.71, + "grad_norm": 0.3497346341609955, + "learning_rate": 4.314049586776859e-05, + "loss": 0.4544, + "step": 1046 + }, + { + "epoch": 1.71, + "grad_norm": 0.31490492820739746, + "learning_rate": 4.289256198347107e-05, + "loss": 0.4809, + "step": 1047 + }, + { + "epoch": 1.72, + "grad_norm": 0.26548659801483154, + "learning_rate": 4.264462809917355e-05, + "loss": 0.3189, + "step": 1048 + }, + { + "epoch": 1.72, + "grad_norm": 0.40890252590179443, + "learning_rate": 4.239669421487603e-05, + "loss": 0.4825, + "step": 1049 + }, + { + "epoch": 1.72, + "grad_norm": 0.392419695854187, + "learning_rate": 4.214876033057851e-05, + "loss": 0.3518, + "step": 1050 + }, + { + "epoch": 1.72, + "grad_norm": 0.3267776370048523, + "learning_rate": 4.190082644628099e-05, + "loss": 0.5964, + "step": 1051 + }, + { + "epoch": 1.72, + "grad_norm": 0.29872927069664, + "learning_rate": 4.1652892561983464e-05, + "loss": 0.3496, + "step": 1052 + }, + { + "epoch": 1.72, + "grad_norm": 0.3140263259410858, + "learning_rate": 4.140495867768595e-05, + "loss": 0.3496, + "step": 1053 + }, + { + "epoch": 1.73, + "grad_norm": 0.35923945903778076, + "learning_rate": 4.1157024793388424e-05, + "loss": 0.4328, + "step": 1054 + }, + { + "epoch": 1.73, + "grad_norm": 0.24899311363697052, + "learning_rate": 4.09090909090909e-05, + "loss": 0.3662, + "step": 1055 + }, + { + "epoch": 1.73, + "grad_norm": 0.300325870513916, + "learning_rate": 4.066115702479338e-05, + "loss": 0.3714, + "step": 1056 + }, + { + "epoch": 1.73, + "grad_norm": 0.26927053928375244, + "learning_rate": 4.041322314049587e-05, + "loss": 0.3518, + "step": 1057 + }, + { + "epoch": 1.73, + "grad_norm": 0.28170421719551086, + "learning_rate": 4.0165289256198345e-05, + "loss": 0.4214, + "step": 1058 + }, + { + "epoch": 1.73, + "grad_norm": 0.3097275197505951, + "learning_rate": 3.991735537190082e-05, + "loss": 0.3387, + "step": 1059 + }, + { + "epoch": 1.74, + "grad_norm": 0.36259180307388306, + "learning_rate": 3.9669421487603306e-05, + "loss": 0.4968, + "step": 1060 + }, + { + "epoch": 1.74, + "grad_norm": 0.3555668592453003, + "learning_rate": 3.942148760330578e-05, + "loss": 0.4415, + "step": 1061 + }, + { + "epoch": 1.74, + "grad_norm": 0.2894740104675293, + "learning_rate": 3.917355371900826e-05, + "loss": 0.3911, + "step": 1062 + }, + { + "epoch": 1.74, + "grad_norm": 0.3361656665802002, + "learning_rate": 3.8925619834710736e-05, + "loss": 0.4286, + "step": 1063 + }, + { + "epoch": 1.74, + "grad_norm": 0.33269697427749634, + "learning_rate": 3.867768595041322e-05, + "loss": 0.5162, + "step": 1064 + }, + { + "epoch": 1.74, + "grad_norm": 0.3324260711669922, + "learning_rate": 3.84297520661157e-05, + "loss": 0.4073, + "step": 1065 + }, + { + "epoch": 1.75, + "grad_norm": 0.3037840723991394, + "learning_rate": 3.8181818181818174e-05, + "loss": 0.4084, + "step": 1066 + }, + { + "epoch": 1.75, + "grad_norm": 0.29843100905418396, + "learning_rate": 3.7933884297520664e-05, + "loss": 0.4028, + "step": 1067 + }, + { + "epoch": 1.75, + "grad_norm": 0.24433061480522156, + "learning_rate": 3.768595041322314e-05, + "loss": 0.3769, + "step": 1068 + }, + { + "epoch": 1.75, + "grad_norm": 0.31540754437446594, + "learning_rate": 3.743801652892562e-05, + "loss": 0.4006, + "step": 1069 + }, + { + "epoch": 1.75, + "grad_norm": 0.3915780186653137, + "learning_rate": 3.7190082644628094e-05, + "loss": 0.3859, + "step": 1070 + }, + { + "epoch": 1.75, + "grad_norm": 0.7843402028083801, + "learning_rate": 3.694214876033058e-05, + "loss": 0.4284, + "step": 1071 + }, + { + "epoch": 1.76, + "grad_norm": 0.3000487685203552, + "learning_rate": 3.6694214876033055e-05, + "loss": 0.6066, + "step": 1072 + }, + { + "epoch": 1.76, + "grad_norm": 0.2342897206544876, + "learning_rate": 3.644628099173553e-05, + "loss": 0.3012, + "step": 1073 + }, + { + "epoch": 1.76, + "grad_norm": 0.3100823760032654, + "learning_rate": 3.6198347107438015e-05, + "loss": 0.4236, + "step": 1074 + }, + { + "epoch": 1.76, + "grad_norm": 0.3442421853542328, + "learning_rate": 3.595041322314049e-05, + "loss": 0.4716, + "step": 1075 + }, + { + "epoch": 1.76, + "grad_norm": 0.2785506546497345, + "learning_rate": 3.570247933884297e-05, + "loss": 0.307, + "step": 1076 + }, + { + "epoch": 1.76, + "grad_norm": 0.333635151386261, + "learning_rate": 3.545454545454545e-05, + "loss": 0.4521, + "step": 1077 + }, + { + "epoch": 1.77, + "grad_norm": 0.3365010619163513, + "learning_rate": 3.520661157024793e-05, + "loss": 0.4522, + "step": 1078 + }, + { + "epoch": 1.77, + "grad_norm": 0.31510964035987854, + "learning_rate": 3.495867768595041e-05, + "loss": 0.4101, + "step": 1079 + }, + { + "epoch": 1.77, + "grad_norm": 0.2939818501472473, + "learning_rate": 3.471074380165289e-05, + "loss": 0.378, + "step": 1080 + }, + { + "epoch": 1.77, + "grad_norm": 0.33073171973228455, + "learning_rate": 3.446280991735537e-05, + "loss": 0.4319, + "step": 1081 + }, + { + "epoch": 1.77, + "grad_norm": 0.306769460439682, + "learning_rate": 3.421487603305785e-05, + "loss": 0.4584, + "step": 1082 + }, + { + "epoch": 1.77, + "grad_norm": 0.3151317536830902, + "learning_rate": 3.396694214876033e-05, + "loss": 0.3202, + "step": 1083 + }, + { + "epoch": 1.77, + "grad_norm": 0.313348650932312, + "learning_rate": 3.371900826446281e-05, + "loss": 0.4051, + "step": 1084 + }, + { + "epoch": 1.78, + "grad_norm": 0.3377431333065033, + "learning_rate": 3.347107438016529e-05, + "loss": 0.3842, + "step": 1085 + }, + { + "epoch": 1.78, + "grad_norm": 0.31378257274627686, + "learning_rate": 3.3223140495867765e-05, + "loss": 0.377, + "step": 1086 + }, + { + "epoch": 1.78, + "grad_norm": 0.31627315282821655, + "learning_rate": 3.297520661157024e-05, + "loss": 0.4278, + "step": 1087 + }, + { + "epoch": 1.78, + "grad_norm": 0.2957272529602051, + "learning_rate": 3.2727272727272725e-05, + "loss": 0.3384, + "step": 1088 + }, + { + "epoch": 1.78, + "grad_norm": 0.3261624872684479, + "learning_rate": 3.247933884297521e-05, + "loss": 0.4525, + "step": 1089 + }, + { + "epoch": 1.78, + "grad_norm": 0.28680557012557983, + "learning_rate": 3.2231404958677685e-05, + "loss": 0.3627, + "step": 1090 + }, + { + "epoch": 1.79, + "grad_norm": 0.29543063044548035, + "learning_rate": 3.198347107438016e-05, + "loss": 0.2922, + "step": 1091 + }, + { + "epoch": 1.79, + "grad_norm": 0.3554795980453491, + "learning_rate": 3.173553719008264e-05, + "loss": 0.4692, + "step": 1092 + }, + { + "epoch": 1.79, + "grad_norm": 0.28728747367858887, + "learning_rate": 3.148760330578512e-05, + "loss": 0.2595, + "step": 1093 + }, + { + "epoch": 1.79, + "grad_norm": 0.3099517524242401, + "learning_rate": 3.12396694214876e-05, + "loss": 0.3912, + "step": 1094 + }, + { + "epoch": 1.79, + "grad_norm": 0.3173176050186157, + "learning_rate": 3.099173553719008e-05, + "loss": 0.4186, + "step": 1095 + }, + { + "epoch": 1.79, + "grad_norm": 0.3445116877555847, + "learning_rate": 3.074380165289256e-05, + "loss": 0.4966, + "step": 1096 + }, + { + "epoch": 1.8, + "grad_norm": 0.32030245661735535, + "learning_rate": 3.0495867768595037e-05, + "loss": 0.4671, + "step": 1097 + }, + { + "epoch": 1.8, + "grad_norm": 0.3321797847747803, + "learning_rate": 3.0247933884297517e-05, + "loss": 0.4265, + "step": 1098 + }, + { + "epoch": 1.8, + "grad_norm": 0.36085036396980286, + "learning_rate": 2.9999999999999997e-05, + "loss": 0.4324, + "step": 1099 + }, + { + "epoch": 1.8, + "grad_norm": 0.2999497950077057, + "learning_rate": 2.9752066115702478e-05, + "loss": 0.3173, + "step": 1100 + }, + { + "epoch": 1.8, + "grad_norm": 0.31063607335090637, + "learning_rate": 2.9504132231404954e-05, + "loss": 0.4941, + "step": 1101 + }, + { + "epoch": 1.8, + "grad_norm": 0.2864468991756439, + "learning_rate": 2.9256198347107435e-05, + "loss": 0.4309, + "step": 1102 + }, + { + "epoch": 1.81, + "grad_norm": 0.2904879152774811, + "learning_rate": 2.900826446280991e-05, + "loss": 0.4782, + "step": 1103 + }, + { + "epoch": 1.81, + "grad_norm": 0.31169822812080383, + "learning_rate": 2.8760330578512395e-05, + "loss": 0.4881, + "step": 1104 + }, + { + "epoch": 1.81, + "grad_norm": 0.3462170660495758, + "learning_rate": 2.8512396694214875e-05, + "loss": 0.3551, + "step": 1105 + }, + { + "epoch": 1.81, + "grad_norm": 0.3066549301147461, + "learning_rate": 2.8264462809917352e-05, + "loss": 0.4522, + "step": 1106 + }, + { + "epoch": 1.81, + "grad_norm": 0.33785369992256165, + "learning_rate": 2.8016528925619832e-05, + "loss": 0.3763, + "step": 1107 + }, + { + "epoch": 1.81, + "grad_norm": 0.2975507378578186, + "learning_rate": 2.776859504132231e-05, + "loss": 0.3193, + "step": 1108 + }, + { + "epoch": 1.82, + "grad_norm": 0.31934845447540283, + "learning_rate": 2.7520661157024793e-05, + "loss": 0.2994, + "step": 1109 + }, + { + "epoch": 1.82, + "grad_norm": 0.29450473189353943, + "learning_rate": 2.727272727272727e-05, + "loss": 0.4279, + "step": 1110 + }, + { + "epoch": 1.82, + "grad_norm": 0.3054717779159546, + "learning_rate": 2.702479338842975e-05, + "loss": 0.4687, + "step": 1111 + }, + { + "epoch": 1.82, + "grad_norm": 0.32938167452812195, + "learning_rate": 2.6776859504132227e-05, + "loss": 0.4815, + "step": 1112 + }, + { + "epoch": 1.82, + "grad_norm": 0.2678495943546295, + "learning_rate": 2.6528925619834707e-05, + "loss": 0.3116, + "step": 1113 + }, + { + "epoch": 1.82, + "grad_norm": 0.26357004046440125, + "learning_rate": 2.628099173553719e-05, + "loss": 0.3286, + "step": 1114 + }, + { + "epoch": 1.83, + "grad_norm": 0.3359578251838684, + "learning_rate": 2.6033057851239667e-05, + "loss": 0.4137, + "step": 1115 + }, + { + "epoch": 1.83, + "grad_norm": 0.3395717442035675, + "learning_rate": 2.5785123966942148e-05, + "loss": 0.3812, + "step": 1116 + }, + { + "epoch": 1.83, + "grad_norm": 0.29891693592071533, + "learning_rate": 2.5537190082644625e-05, + "loss": 0.2989, + "step": 1117 + }, + { + "epoch": 1.83, + "grad_norm": 0.402649462223053, + "learning_rate": 2.5289256198347105e-05, + "loss": 0.4333, + "step": 1118 + }, + { + "epoch": 1.83, + "grad_norm": 0.3397662341594696, + "learning_rate": 2.5041322314049585e-05, + "loss": 0.4188, + "step": 1119 + }, + { + "epoch": 1.83, + "grad_norm": 0.33743607997894287, + "learning_rate": 2.4793388429752065e-05, + "loss": 0.5309, + "step": 1120 + }, + { + "epoch": 1.84, + "grad_norm": 0.3248274624347687, + "learning_rate": 2.4545454545454542e-05, + "loss": 0.3905, + "step": 1121 + }, + { + "epoch": 1.84, + "grad_norm": 0.3567257821559906, + "learning_rate": 2.4297520661157022e-05, + "loss": 0.4107, + "step": 1122 + }, + { + "epoch": 1.84, + "grad_norm": 0.4383893311023712, + "learning_rate": 2.40495867768595e-05, + "loss": 0.5024, + "step": 1123 + }, + { + "epoch": 1.84, + "grad_norm": 0.2777807414531708, + "learning_rate": 2.3801652892561983e-05, + "loss": 0.3289, + "step": 1124 + }, + { + "epoch": 1.84, + "grad_norm": 0.3409118950366974, + "learning_rate": 2.3553719008264463e-05, + "loss": 0.5199, + "step": 1125 + }, + { + "epoch": 1.84, + "grad_norm": 0.3060845732688904, + "learning_rate": 2.330578512396694e-05, + "loss": 0.412, + "step": 1126 + }, + { + "epoch": 1.85, + "grad_norm": 0.3366425335407257, + "learning_rate": 2.305785123966942e-05, + "loss": 0.484, + "step": 1127 + }, + { + "epoch": 1.85, + "grad_norm": 0.36060798168182373, + "learning_rate": 2.2809917355371897e-05, + "loss": 0.5543, + "step": 1128 + }, + { + "epoch": 1.85, + "grad_norm": 0.25729015469551086, + "learning_rate": 2.256198347107438e-05, + "loss": 0.2763, + "step": 1129 + }, + { + "epoch": 1.85, + "grad_norm": 0.2890430688858032, + "learning_rate": 2.2314049586776857e-05, + "loss": 0.3762, + "step": 1130 + }, + { + "epoch": 1.85, + "grad_norm": 0.31579041481018066, + "learning_rate": 2.2066115702479338e-05, + "loss": 0.396, + "step": 1131 + }, + { + "epoch": 1.85, + "grad_norm": 0.3136342763900757, + "learning_rate": 2.1818181818181814e-05, + "loss": 0.4134, + "step": 1132 + }, + { + "epoch": 1.86, + "grad_norm": 0.37239784002304077, + "learning_rate": 2.1570247933884295e-05, + "loss": 0.4666, + "step": 1133 + }, + { + "epoch": 1.86, + "grad_norm": 0.2847795784473419, + "learning_rate": 2.1322314049586775e-05, + "loss": 0.3481, + "step": 1134 + }, + { + "epoch": 1.86, + "grad_norm": 0.27870920300483704, + "learning_rate": 2.1074380165289255e-05, + "loss": 0.2669, + "step": 1135 + }, + { + "epoch": 1.86, + "grad_norm": 0.2700231969356537, + "learning_rate": 2.0826446280991732e-05, + "loss": 0.2798, + "step": 1136 + }, + { + "epoch": 1.86, + "grad_norm": 0.3257925510406494, + "learning_rate": 2.0578512396694212e-05, + "loss": 0.4931, + "step": 1137 + }, + { + "epoch": 1.86, + "grad_norm": 0.2964242994785309, + "learning_rate": 2.033057851239669e-05, + "loss": 0.429, + "step": 1138 + }, + { + "epoch": 1.86, + "grad_norm": 0.32561832666397095, + "learning_rate": 2.0082644628099173e-05, + "loss": 0.3467, + "step": 1139 + }, + { + "epoch": 1.87, + "grad_norm": 0.27957382798194885, + "learning_rate": 1.9834710743801653e-05, + "loss": 0.2686, + "step": 1140 + }, + { + "epoch": 1.87, + "grad_norm": 0.3476884663105011, + "learning_rate": 1.958677685950413e-05, + "loss": 0.4814, + "step": 1141 + }, + { + "epoch": 1.87, + "grad_norm": 0.2950107753276825, + "learning_rate": 1.933884297520661e-05, + "loss": 0.3578, + "step": 1142 + }, + { + "epoch": 1.87, + "grad_norm": 0.30689096450805664, + "learning_rate": 1.9090909090909087e-05, + "loss": 0.3725, + "step": 1143 + }, + { + "epoch": 1.87, + "grad_norm": 0.430915504693985, + "learning_rate": 1.884297520661157e-05, + "loss": 0.4766, + "step": 1144 + }, + { + "epoch": 1.87, + "grad_norm": 0.3086168169975281, + "learning_rate": 1.8595041322314047e-05, + "loss": 0.5506, + "step": 1145 + }, + { + "epoch": 1.88, + "grad_norm": 0.3441203534603119, + "learning_rate": 1.8347107438016527e-05, + "loss": 0.4251, + "step": 1146 + }, + { + "epoch": 1.88, + "grad_norm": 0.2828252613544464, + "learning_rate": 1.8099173553719008e-05, + "loss": 0.3, + "step": 1147 + }, + { + "epoch": 1.88, + "grad_norm": 0.33563023805618286, + "learning_rate": 1.7851239669421485e-05, + "loss": 0.4082, + "step": 1148 + }, + { + "epoch": 1.88, + "grad_norm": 0.33100175857543945, + "learning_rate": 1.7603305785123965e-05, + "loss": 0.5853, + "step": 1149 + }, + { + "epoch": 1.88, + "grad_norm": 0.3554556369781494, + "learning_rate": 1.7355371900826445e-05, + "loss": 0.5677, + "step": 1150 + }, + { + "epoch": 1.88, + "grad_norm": 0.32995131611824036, + "learning_rate": 1.7107438016528925e-05, + "loss": 0.3315, + "step": 1151 + }, + { + "epoch": 1.89, + "grad_norm": 0.3160393238067627, + "learning_rate": 1.6859504132231405e-05, + "loss": 0.3632, + "step": 1152 + }, + { + "epoch": 1.89, + "grad_norm": 0.3632807433605194, + "learning_rate": 1.6611570247933882e-05, + "loss": 0.4053, + "step": 1153 + }, + { + "epoch": 1.89, + "grad_norm": 0.2931605279445648, + "learning_rate": 1.6363636363636363e-05, + "loss": 0.358, + "step": 1154 + }, + { + "epoch": 1.89, + "grad_norm": 0.32687610387802124, + "learning_rate": 1.6115702479338843e-05, + "loss": 0.4584, + "step": 1155 + }, + { + "epoch": 1.89, + "grad_norm": 0.3283078074455261, + "learning_rate": 1.586776859504132e-05, + "loss": 0.3818, + "step": 1156 + }, + { + "epoch": 1.89, + "grad_norm": 0.31993189454078674, + "learning_rate": 1.56198347107438e-05, + "loss": 0.3714, + "step": 1157 + }, + { + "epoch": 1.9, + "grad_norm": 0.2674204409122467, + "learning_rate": 1.537190082644628e-05, + "loss": 0.3943, + "step": 1158 + }, + { + "epoch": 1.9, + "grad_norm": 0.3968242406845093, + "learning_rate": 1.5123966942148759e-05, + "loss": 0.4465, + "step": 1159 + }, + { + "epoch": 1.9, + "grad_norm": 0.2870213985443115, + "learning_rate": 1.4876033057851239e-05, + "loss": 0.3616, + "step": 1160 + }, + { + "epoch": 1.9, + "grad_norm": 0.29502633213996887, + "learning_rate": 1.4628099173553717e-05, + "loss": 0.4112, + "step": 1161 + }, + { + "epoch": 1.9, + "grad_norm": 0.36414459347724915, + "learning_rate": 1.4380165289256198e-05, + "loss": 0.3928, + "step": 1162 + }, + { + "epoch": 1.9, + "grad_norm": 0.274940550327301, + "learning_rate": 1.4132231404958676e-05, + "loss": 0.3971, + "step": 1163 + }, + { + "epoch": 1.91, + "grad_norm": 0.3382115364074707, + "learning_rate": 1.3884297520661155e-05, + "loss": 0.339, + "step": 1164 + }, + { + "epoch": 1.91, + "grad_norm": 0.32059189677238464, + "learning_rate": 1.3636363636363635e-05, + "loss": 0.4632, + "step": 1165 + }, + { + "epoch": 1.91, + "grad_norm": 0.40788954496383667, + "learning_rate": 1.3388429752066113e-05, + "loss": 0.4729, + "step": 1166 + }, + { + "epoch": 1.91, + "grad_norm": 0.4415609836578369, + "learning_rate": 1.3140495867768595e-05, + "loss": 0.4311, + "step": 1167 + }, + { + "epoch": 1.91, + "grad_norm": 0.29439279437065125, + "learning_rate": 1.2892561983471074e-05, + "loss": 0.3428, + "step": 1168 + }, + { + "epoch": 1.91, + "grad_norm": 0.38421952724456787, + "learning_rate": 1.2644628099173552e-05, + "loss": 0.5504, + "step": 1169 + }, + { + "epoch": 1.92, + "grad_norm": 0.2757047116756439, + "learning_rate": 1.2396694214876033e-05, + "loss": 0.3488, + "step": 1170 + }, + { + "epoch": 1.92, + "grad_norm": 0.27029332518577576, + "learning_rate": 1.2148760330578511e-05, + "loss": 0.3922, + "step": 1171 + }, + { + "epoch": 1.92, + "grad_norm": 0.29828086495399475, + "learning_rate": 1.1900826446280991e-05, + "loss": 0.3484, + "step": 1172 + }, + { + "epoch": 1.92, + "grad_norm": 0.3248095214366913, + "learning_rate": 1.165289256198347e-05, + "loss": 0.4166, + "step": 1173 + }, + { + "epoch": 1.92, + "grad_norm": 0.3183375895023346, + "learning_rate": 1.1404958677685948e-05, + "loss": 0.4207, + "step": 1174 + }, + { + "epoch": 1.92, + "grad_norm": 0.38209760189056396, + "learning_rate": 1.1157024793388429e-05, + "loss": 0.4136, + "step": 1175 + }, + { + "epoch": 1.93, + "grad_norm": 0.31191781163215637, + "learning_rate": 1.0909090909090907e-05, + "loss": 0.3821, + "step": 1176 + }, + { + "epoch": 1.93, + "grad_norm": 0.3147072494029999, + "learning_rate": 1.0661157024793387e-05, + "loss": 0.2973, + "step": 1177 + }, + { + "epoch": 1.93, + "grad_norm": 0.346629798412323, + "learning_rate": 1.0413223140495866e-05, + "loss": 0.5924, + "step": 1178 + }, + { + "epoch": 1.93, + "grad_norm": 0.30329591035842896, + "learning_rate": 1.0165289256198345e-05, + "loss": 0.4802, + "step": 1179 + }, + { + "epoch": 1.93, + "grad_norm": 0.3608144521713257, + "learning_rate": 9.917355371900826e-06, + "loss": 0.4187, + "step": 1180 + }, + { + "epoch": 1.93, + "grad_norm": 0.3330174684524536, + "learning_rate": 9.669421487603305e-06, + "loss": 0.4585, + "step": 1181 + }, + { + "epoch": 1.94, + "grad_norm": 0.2880091071128845, + "learning_rate": 9.421487603305785e-06, + "loss": 0.3926, + "step": 1182 + }, + { + "epoch": 1.94, + "grad_norm": 0.2711026668548584, + "learning_rate": 9.173553719008264e-06, + "loss": 0.3128, + "step": 1183 + }, + { + "epoch": 1.94, + "grad_norm": 0.3472573161125183, + "learning_rate": 8.925619834710742e-06, + "loss": 0.3626, + "step": 1184 + }, + { + "epoch": 1.94, + "grad_norm": 0.29903772473335266, + "learning_rate": 8.677685950413222e-06, + "loss": 0.3778, + "step": 1185 + }, + { + "epoch": 1.94, + "grad_norm": 0.309654176235199, + "learning_rate": 8.429752066115703e-06, + "loss": 0.3965, + "step": 1186 + }, + { + "epoch": 1.94, + "grad_norm": 0.3163444399833679, + "learning_rate": 8.181818181818181e-06, + "loss": 0.3207, + "step": 1187 + }, + { + "epoch": 1.95, + "grad_norm": 0.3754628300666809, + "learning_rate": 7.93388429752066e-06, + "loss": 0.3954, + "step": 1188 + }, + { + "epoch": 1.95, + "grad_norm": 0.2967177629470825, + "learning_rate": 7.68595041322314e-06, + "loss": 0.4092, + "step": 1189 + }, + { + "epoch": 1.95, + "grad_norm": 0.37930914759635925, + "learning_rate": 7.438016528925619e-06, + "loss": 0.5038, + "step": 1190 + }, + { + "epoch": 1.95, + "grad_norm": 0.31978312134742737, + "learning_rate": 7.190082644628099e-06, + "loss": 0.3039, + "step": 1191 + }, + { + "epoch": 1.95, + "grad_norm": 0.34556475281715393, + "learning_rate": 6.942148760330577e-06, + "loss": 0.3749, + "step": 1192 + }, + { + "epoch": 1.95, + "grad_norm": 0.33958449959754944, + "learning_rate": 6.694214876033057e-06, + "loss": 0.4974, + "step": 1193 + }, + { + "epoch": 1.95, + "grad_norm": 0.34213709831237793, + "learning_rate": 6.446280991735537e-06, + "loss": 0.4874, + "step": 1194 + }, + { + "epoch": 1.96, + "grad_norm": 0.3194979131221771, + "learning_rate": 6.198347107438016e-06, + "loss": 0.4415, + "step": 1195 + }, + { + "epoch": 1.96, + "grad_norm": 0.3170003890991211, + "learning_rate": 5.950413223140496e-06, + "loss": 0.299, + "step": 1196 + }, + { + "epoch": 1.96, + "grad_norm": 0.35796797275543213, + "learning_rate": 5.702479338842974e-06, + "loss": 0.4516, + "step": 1197 + }, + { + "epoch": 1.96, + "grad_norm": 0.36410433053970337, + "learning_rate": 5.454545454545454e-06, + "loss": 0.3137, + "step": 1198 + }, + { + "epoch": 1.96, + "grad_norm": 0.27563753724098206, + "learning_rate": 5.206611570247933e-06, + "loss": 0.3465, + "step": 1199 + }, + { + "epoch": 1.96, + "grad_norm": 0.3430056869983673, + "learning_rate": 4.958677685950413e-06, + "loss": 0.5325, + "step": 1200 + }, + { + "epoch": 1.97, + "grad_norm": 0.3032241463661194, + "learning_rate": 4.710743801652893e-06, + "loss": 0.3802, + "step": 1201 + }, + { + "epoch": 1.97, + "grad_norm": 0.3008878231048584, + "learning_rate": 4.462809917355371e-06, + "loss": 0.3674, + "step": 1202 + }, + { + "epoch": 1.97, + "grad_norm": 0.34465453028678894, + "learning_rate": 4.214876033057851e-06, + "loss": 0.3465, + "step": 1203 + }, + { + "epoch": 1.97, + "grad_norm": 0.3217530846595764, + "learning_rate": 3.96694214876033e-06, + "loss": 0.395, + "step": 1204 + }, + { + "epoch": 1.97, + "grad_norm": 0.3256390690803528, + "learning_rate": 3.7190082644628097e-06, + "loss": 0.2928, + "step": 1205 + }, + { + "epoch": 1.97, + "grad_norm": 0.404376357793808, + "learning_rate": 3.4710743801652887e-06, + "loss": 0.5579, + "step": 1206 + }, + { + "epoch": 1.98, + "grad_norm": 0.2786218822002411, + "learning_rate": 3.2231404958677685e-06, + "loss": 0.3842, + "step": 1207 + }, + { + "epoch": 1.98, + "grad_norm": 0.339501291513443, + "learning_rate": 2.975206611570248e-06, + "loss": 0.4061, + "step": 1208 + }, + { + "epoch": 1.98, + "grad_norm": 0.3386409878730774, + "learning_rate": 2.727272727272727e-06, + "loss": 0.3452, + "step": 1209 + }, + { + "epoch": 1.98, + "grad_norm": 0.36449265480041504, + "learning_rate": 2.4793388429752066e-06, + "loss": 0.3769, + "step": 1210 + }, + { + "epoch": 1.98, + "grad_norm": 0.3336932361125946, + "learning_rate": 2.2314049586776856e-06, + "loss": 0.4361, + "step": 1211 + }, + { + "epoch": 1.98, + "grad_norm": 0.28075236082077026, + "learning_rate": 1.983471074380165e-06, + "loss": 0.3614, + "step": 1212 + }, + { + "epoch": 1.99, + "grad_norm": 0.31337854266166687, + "learning_rate": 1.7355371900826443e-06, + "loss": 0.37, + "step": 1213 + }, + { + "epoch": 1.99, + "grad_norm": 0.3034374415874481, + "learning_rate": 1.487603305785124e-06, + "loss": 0.274, + "step": 1214 + }, + { + "epoch": 1.99, + "grad_norm": 0.3485061526298523, + "learning_rate": 1.2396694214876033e-06, + "loss": 0.425, + "step": 1215 + }, + { + "epoch": 1.99, + "grad_norm": 0.24720066785812378, + "learning_rate": 9.917355371900825e-07, + "loss": 0.307, + "step": 1216 + }, + { + "epoch": 1.99, + "grad_norm": 0.2727121412754059, + "learning_rate": 7.43801652892562e-07, + "loss": 0.2991, + "step": 1217 + }, + { + "epoch": 1.99, + "grad_norm": 0.33211690187454224, + "learning_rate": 4.958677685950412e-07, + "loss": 0.5309, + "step": 1218 + }, + { + "epoch": 2.0, + "grad_norm": 0.328895628452301, + "learning_rate": 2.479338842975206e-07, + "loss": 0.3547, + "step": 1219 + }, + { + "epoch": 2.0, + "grad_norm": 0.2642543315887451, + "learning_rate": 0.0, + "loss": 0.3047, + "step": 1220 + } + ], + "logging_steps": 1, + "max_steps": 1220, + "num_input_tokens_seen": 0, + "num_train_epochs": 2, + "save_steps": 610, + "total_flos": 2.626577866972938e+18, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}