diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,14811 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.9114583333333334, + "eval_steps": 300, + "global_step": 2100, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00043402777777777775, + "grad_norm": 22.039037704467773, + "learning_rate": 7.194244604316547e-08, + "loss": 0.7083, + "step": 1 + }, + { + "epoch": 0.0008680555555555555, + "grad_norm": 20.530189514160156, + "learning_rate": 1.4388489208633095e-07, + "loss": 0.7177, + "step": 2 + }, + { + "epoch": 0.0013020833333333333, + "grad_norm": 23.356374740600586, + "learning_rate": 2.1582733812949643e-07, + "loss": 0.7513, + "step": 3 + }, + { + "epoch": 0.001736111111111111, + "grad_norm": 20.72550392150879, + "learning_rate": 2.877697841726619e-07, + "loss": 0.7145, + "step": 4 + }, + { + "epoch": 0.002170138888888889, + "grad_norm": 22.307851791381836, + "learning_rate": 3.5971223021582736e-07, + "loss": 0.6842, + "step": 5 + }, + { + "epoch": 0.0026041666666666665, + "grad_norm": 26.96861457824707, + "learning_rate": 4.3165467625899287e-07, + "loss": 0.7214, + "step": 6 + }, + { + "epoch": 0.0030381944444444445, + "grad_norm": 22.72673988342285, + "learning_rate": 5.035971223021583e-07, + "loss": 0.6912, + "step": 7 + }, + { + "epoch": 0.003472222222222222, + "grad_norm": 20.26327896118164, + "learning_rate": 5.755395683453238e-07, + "loss": 0.6848, + "step": 8 + }, + { + "epoch": 0.00390625, + "grad_norm": 19.785690307617188, + "learning_rate": 6.474820143884893e-07, + "loss": 0.6813, + "step": 9 + }, + { + "epoch": 0.004340277777777778, + "grad_norm": 20.36345863342285, + "learning_rate": 7.194244604316547e-07, + "loss": 0.6662, + "step": 10 + }, + { + "epoch": 0.004774305555555556, + "grad_norm": 22.651525497436523, + "learning_rate": 7.913669064748202e-07, + "loss": 0.6969, + "step": 11 + }, + { + "epoch": 0.005208333333333333, + "grad_norm": 24.08732795715332, + "learning_rate": 8.633093525179857e-07, + "loss": 0.6428, + "step": 12 + }, + { + "epoch": 0.005642361111111111, + "grad_norm": 24.18958854675293, + "learning_rate": 9.352517985611512e-07, + "loss": 0.5645, + "step": 13 + }, + { + "epoch": 0.006076388888888889, + "grad_norm": 23.265363693237305, + "learning_rate": 1.0071942446043167e-06, + "loss": 0.5913, + "step": 14 + }, + { + "epoch": 0.006510416666666667, + "grad_norm": 21.222742080688477, + "learning_rate": 1.079136690647482e-06, + "loss": 0.5673, + "step": 15 + }, + { + "epoch": 0.006944444444444444, + "grad_norm": 19.037031173706055, + "learning_rate": 1.1510791366906476e-06, + "loss": 0.529, + "step": 16 + }, + { + "epoch": 0.007378472222222222, + "grad_norm": 25.137651443481445, + "learning_rate": 1.2230215827338131e-06, + "loss": 0.3661, + "step": 17 + }, + { + "epoch": 0.0078125, + "grad_norm": 27.268735885620117, + "learning_rate": 1.2949640287769785e-06, + "loss": 0.3817, + "step": 18 + }, + { + "epoch": 0.008246527777777778, + "grad_norm": 26.401870727539062, + "learning_rate": 1.366906474820144e-06, + "loss": 0.36, + "step": 19 + }, + { + "epoch": 0.008680555555555556, + "grad_norm": 19.024900436401367, + "learning_rate": 1.4388489208633094e-06, + "loss": 0.3818, + "step": 20 + }, + { + "epoch": 0.009114583333333334, + "grad_norm": 15.296829223632812, + "learning_rate": 1.510791366906475e-06, + "loss": 0.3154, + "step": 21 + }, + { + "epoch": 0.009548611111111112, + "grad_norm": 21.349294662475586, + "learning_rate": 1.5827338129496403e-06, + "loss": 0.2649, + "step": 22 + }, + { + "epoch": 0.009982638888888888, + "grad_norm": 42.53651428222656, + "learning_rate": 1.654676258992806e-06, + "loss": 0.2867, + "step": 23 + }, + { + "epoch": 0.010416666666666666, + "grad_norm": 13.64511775970459, + "learning_rate": 1.7266187050359715e-06, + "loss": 0.2534, + "step": 24 + }, + { + "epoch": 0.010850694444444444, + "grad_norm": 20.433380126953125, + "learning_rate": 1.7985611510791368e-06, + "loss": 0.2721, + "step": 25 + }, + { + "epoch": 0.011284722222222222, + "grad_norm": 12.288137435913086, + "learning_rate": 1.8705035971223024e-06, + "loss": 0.2361, + "step": 26 + }, + { + "epoch": 0.01171875, + "grad_norm": 14.382365226745605, + "learning_rate": 1.942446043165468e-06, + "loss": 0.2462, + "step": 27 + }, + { + "epoch": 0.012152777777777778, + "grad_norm": 9.596738815307617, + "learning_rate": 2.0143884892086333e-06, + "loss": 0.1907, + "step": 28 + }, + { + "epoch": 0.012586805555555556, + "grad_norm": 7.768435478210449, + "learning_rate": 2.0863309352517987e-06, + "loss": 0.1846, + "step": 29 + }, + { + "epoch": 0.013020833333333334, + "grad_norm": 10.87830924987793, + "learning_rate": 2.158273381294964e-06, + "loss": 0.2154, + "step": 30 + }, + { + "epoch": 0.013454861111111112, + "grad_norm": 11.848028182983398, + "learning_rate": 2.23021582733813e-06, + "loss": 0.2569, + "step": 31 + }, + { + "epoch": 0.013888888888888888, + "grad_norm": 9.48557186126709, + "learning_rate": 2.302158273381295e-06, + "loss": 0.2271, + "step": 32 + }, + { + "epoch": 0.014322916666666666, + "grad_norm": 13.119791984558105, + "learning_rate": 2.3741007194244605e-06, + "loss": 0.1902, + "step": 33 + }, + { + "epoch": 0.014756944444444444, + "grad_norm": 8.415437698364258, + "learning_rate": 2.4460431654676263e-06, + "loss": 0.2007, + "step": 34 + }, + { + "epoch": 0.015190972222222222, + "grad_norm": 10.289463996887207, + "learning_rate": 2.5179856115107916e-06, + "loss": 0.1879, + "step": 35 + }, + { + "epoch": 0.015625, + "grad_norm": 14.64406681060791, + "learning_rate": 2.589928057553957e-06, + "loss": 0.173, + "step": 36 + }, + { + "epoch": 0.016059027777777776, + "grad_norm": 12.37926959991455, + "learning_rate": 2.6618705035971228e-06, + "loss": 0.2205, + "step": 37 + }, + { + "epoch": 0.016493055555555556, + "grad_norm": 13.952500343322754, + "learning_rate": 2.733812949640288e-06, + "loss": 0.2058, + "step": 38 + }, + { + "epoch": 0.016927083333333332, + "grad_norm": 10.09529972076416, + "learning_rate": 2.805755395683453e-06, + "loss": 0.1783, + "step": 39 + }, + { + "epoch": 0.017361111111111112, + "grad_norm": 11.618066787719727, + "learning_rate": 2.877697841726619e-06, + "loss": 0.1718, + "step": 40 + }, + { + "epoch": 0.017795138888888888, + "grad_norm": 5.552605628967285, + "learning_rate": 2.949640287769784e-06, + "loss": 0.1269, + "step": 41 + }, + { + "epoch": 0.018229166666666668, + "grad_norm": 6.753513336181641, + "learning_rate": 3.02158273381295e-06, + "loss": 0.1408, + "step": 42 + }, + { + "epoch": 0.018663194444444444, + "grad_norm": 13.211586952209473, + "learning_rate": 3.0935251798561158e-06, + "loss": 0.1528, + "step": 43 + }, + { + "epoch": 0.019097222222222224, + "grad_norm": 7.235265731811523, + "learning_rate": 3.1654676258992807e-06, + "loss": 0.1649, + "step": 44 + }, + { + "epoch": 0.01953125, + "grad_norm": 7.747194766998291, + "learning_rate": 3.237410071942446e-06, + "loss": 0.1485, + "step": 45 + }, + { + "epoch": 0.019965277777777776, + "grad_norm": 7.433028697967529, + "learning_rate": 3.309352517985612e-06, + "loss": 0.1665, + "step": 46 + }, + { + "epoch": 0.020399305555555556, + "grad_norm": 10.274096488952637, + "learning_rate": 3.381294964028777e-06, + "loss": 0.1545, + "step": 47 + }, + { + "epoch": 0.020833333333333332, + "grad_norm": 13.211413383483887, + "learning_rate": 3.453237410071943e-06, + "loss": 0.1682, + "step": 48 + }, + { + "epoch": 0.021267361111111112, + "grad_norm": 6.6468706130981445, + "learning_rate": 3.525179856115108e-06, + "loss": 0.1291, + "step": 49 + }, + { + "epoch": 0.021701388888888888, + "grad_norm": 8.226445198059082, + "learning_rate": 3.5971223021582737e-06, + "loss": 0.2014, + "step": 50 + }, + { + "epoch": 0.022135416666666668, + "grad_norm": 6.410370826721191, + "learning_rate": 3.669064748201439e-06, + "loss": 0.1378, + "step": 51 + }, + { + "epoch": 0.022569444444444444, + "grad_norm": 4.667335033416748, + "learning_rate": 3.741007194244605e-06, + "loss": 0.1412, + "step": 52 + }, + { + "epoch": 0.023003472222222224, + "grad_norm": 4.757472515106201, + "learning_rate": 3.81294964028777e-06, + "loss": 0.1376, + "step": 53 + }, + { + "epoch": 0.0234375, + "grad_norm": 3.980797529220581, + "learning_rate": 3.884892086330936e-06, + "loss": 0.1506, + "step": 54 + }, + { + "epoch": 0.023871527777777776, + "grad_norm": 5.135439872741699, + "learning_rate": 3.956834532374101e-06, + "loss": 0.1208, + "step": 55 + }, + { + "epoch": 0.024305555555555556, + "grad_norm": 4.567761421203613, + "learning_rate": 4.028776978417267e-06, + "loss": 0.1458, + "step": 56 + }, + { + "epoch": 0.024739583333333332, + "grad_norm": 5.0912699699401855, + "learning_rate": 4.100719424460432e-06, + "loss": 0.1469, + "step": 57 + }, + { + "epoch": 0.025173611111111112, + "grad_norm": 3.5036520957946777, + "learning_rate": 4.172661870503597e-06, + "loss": 0.1111, + "step": 58 + }, + { + "epoch": 0.025607638888888888, + "grad_norm": 3.6658434867858887, + "learning_rate": 4.244604316546763e-06, + "loss": 0.1222, + "step": 59 + }, + { + "epoch": 0.026041666666666668, + "grad_norm": 3.44124436378479, + "learning_rate": 4.316546762589928e-06, + "loss": 0.1143, + "step": 60 + }, + { + "epoch": 0.026475694444444444, + "grad_norm": 4.265533447265625, + "learning_rate": 4.388489208633094e-06, + "loss": 0.1016, + "step": 61 + }, + { + "epoch": 0.026909722222222224, + "grad_norm": 3.329289674758911, + "learning_rate": 4.46043165467626e-06, + "loss": 0.1105, + "step": 62 + }, + { + "epoch": 0.02734375, + "grad_norm": 6.9250383377075195, + "learning_rate": 4.5323741007194245e-06, + "loss": 0.2056, + "step": 63 + }, + { + "epoch": 0.027777777777777776, + "grad_norm": 29.69718360900879, + "learning_rate": 4.60431654676259e-06, + "loss": 0.1643, + "step": 64 + }, + { + "epoch": 0.028211805555555556, + "grad_norm": 3.4973223209381104, + "learning_rate": 4.676258992805755e-06, + "loss": 0.1296, + "step": 65 + }, + { + "epoch": 0.028645833333333332, + "grad_norm": 3.7894699573516846, + "learning_rate": 4.748201438848921e-06, + "loss": 0.1521, + "step": 66 + }, + { + "epoch": 0.029079861111111112, + "grad_norm": 4.099551200866699, + "learning_rate": 4.820143884892087e-06, + "loss": 0.11, + "step": 67 + }, + { + "epoch": 0.029513888888888888, + "grad_norm": 4.133640766143799, + "learning_rate": 4.892086330935253e-06, + "loss": 0.1544, + "step": 68 + }, + { + "epoch": 0.029947916666666668, + "grad_norm": 3.045886754989624, + "learning_rate": 4.9640287769784175e-06, + "loss": 0.1247, + "step": 69 + }, + { + "epoch": 0.030381944444444444, + "grad_norm": 4.6925950050354, + "learning_rate": 5.035971223021583e-06, + "loss": 0.1302, + "step": 70 + }, + { + "epoch": 0.030815972222222224, + "grad_norm": 4.342012882232666, + "learning_rate": 5.107913669064749e-06, + "loss": 0.1378, + "step": 71 + }, + { + "epoch": 0.03125, + "grad_norm": 3.7377989292144775, + "learning_rate": 5.179856115107914e-06, + "loss": 0.1376, + "step": 72 + }, + { + "epoch": 0.031684027777777776, + "grad_norm": 2.9397761821746826, + "learning_rate": 5.251798561151079e-06, + "loss": 0.1279, + "step": 73 + }, + { + "epoch": 0.03211805555555555, + "grad_norm": 2.6648385524749756, + "learning_rate": 5.3237410071942456e-06, + "loss": 0.125, + "step": 74 + }, + { + "epoch": 0.032552083333333336, + "grad_norm": 4.317229747772217, + "learning_rate": 5.3956834532374105e-06, + "loss": 0.1038, + "step": 75 + }, + { + "epoch": 0.03298611111111111, + "grad_norm": 3.513620376586914, + "learning_rate": 5.467625899280576e-06, + "loss": 0.0993, + "step": 76 + }, + { + "epoch": 0.03342013888888889, + "grad_norm": 3.500981330871582, + "learning_rate": 5.539568345323741e-06, + "loss": 0.122, + "step": 77 + }, + { + "epoch": 0.033854166666666664, + "grad_norm": 2.3485305309295654, + "learning_rate": 5.611510791366906e-06, + "loss": 0.0902, + "step": 78 + }, + { + "epoch": 0.03428819444444445, + "grad_norm": 2.7842323780059814, + "learning_rate": 5.683453237410073e-06, + "loss": 0.1088, + "step": 79 + }, + { + "epoch": 0.034722222222222224, + "grad_norm": 3.6362524032592773, + "learning_rate": 5.755395683453238e-06, + "loss": 0.1331, + "step": 80 + }, + { + "epoch": 0.03515625, + "grad_norm": 2.647966146469116, + "learning_rate": 5.8273381294964035e-06, + "loss": 0.1081, + "step": 81 + }, + { + "epoch": 0.035590277777777776, + "grad_norm": 4.212615013122559, + "learning_rate": 5.899280575539568e-06, + "loss": 0.1277, + "step": 82 + }, + { + "epoch": 0.03602430555555555, + "grad_norm": 2.0377724170684814, + "learning_rate": 5.971223021582734e-06, + "loss": 0.1064, + "step": 83 + }, + { + "epoch": 0.036458333333333336, + "grad_norm": 2.544142484664917, + "learning_rate": 6.0431654676259e-06, + "loss": 0.102, + "step": 84 + }, + { + "epoch": 0.03689236111111111, + "grad_norm": 2.4017419815063477, + "learning_rate": 6.115107913669065e-06, + "loss": 0.0823, + "step": 85 + }, + { + "epoch": 0.03732638888888889, + "grad_norm": 3.2943034172058105, + "learning_rate": 6.1870503597122315e-06, + "loss": 0.1098, + "step": 86 + }, + { + "epoch": 0.037760416666666664, + "grad_norm": 3.2607340812683105, + "learning_rate": 6.2589928057553964e-06, + "loss": 0.1552, + "step": 87 + }, + { + "epoch": 0.03819444444444445, + "grad_norm": 2.7917239665985107, + "learning_rate": 6.330935251798561e-06, + "loss": 0.1047, + "step": 88 + }, + { + "epoch": 0.038628472222222224, + "grad_norm": 3.072331190109253, + "learning_rate": 6.402877697841727e-06, + "loss": 0.0831, + "step": 89 + }, + { + "epoch": 0.0390625, + "grad_norm": 1.7378712892532349, + "learning_rate": 6.474820143884892e-06, + "loss": 0.113, + "step": 90 + }, + { + "epoch": 0.039496527777777776, + "grad_norm": 2.934422492980957, + "learning_rate": 6.546762589928059e-06, + "loss": 0.1121, + "step": 91 + }, + { + "epoch": 0.03993055555555555, + "grad_norm": 2.3598244190216064, + "learning_rate": 6.618705035971224e-06, + "loss": 0.0975, + "step": 92 + }, + { + "epoch": 0.040364583333333336, + "grad_norm": 3.3509647846221924, + "learning_rate": 6.6906474820143886e-06, + "loss": 0.1271, + "step": 93 + }, + { + "epoch": 0.04079861111111111, + "grad_norm": 3.8913393020629883, + "learning_rate": 6.762589928057554e-06, + "loss": 0.1008, + "step": 94 + }, + { + "epoch": 0.04123263888888889, + "grad_norm": 2.031092882156372, + "learning_rate": 6.834532374100719e-06, + "loss": 0.0998, + "step": 95 + }, + { + "epoch": 0.041666666666666664, + "grad_norm": 1.9853342771530151, + "learning_rate": 6.906474820143886e-06, + "loss": 0.1243, + "step": 96 + }, + { + "epoch": 0.04210069444444445, + "grad_norm": 1.718274712562561, + "learning_rate": 6.978417266187051e-06, + "loss": 0.109, + "step": 97 + }, + { + "epoch": 0.042534722222222224, + "grad_norm": 1.5930973291397095, + "learning_rate": 7.050359712230216e-06, + "loss": 0.1049, + "step": 98 + }, + { + "epoch": 0.04296875, + "grad_norm": 2.425053119659424, + "learning_rate": 7.122302158273382e-06, + "loss": 0.0991, + "step": 99 + }, + { + "epoch": 0.043402777777777776, + "grad_norm": 2.481341600418091, + "learning_rate": 7.194244604316547e-06, + "loss": 0.0843, + "step": 100 + }, + { + "epoch": 0.04383680555555555, + "grad_norm": 2.6645309925079346, + "learning_rate": 7.266187050359713e-06, + "loss": 0.098, + "step": 101 + }, + { + "epoch": 0.044270833333333336, + "grad_norm": 4.1276960372924805, + "learning_rate": 7.338129496402878e-06, + "loss": 0.1159, + "step": 102 + }, + { + "epoch": 0.04470486111111111, + "grad_norm": 4.709148406982422, + "learning_rate": 7.410071942446043e-06, + "loss": 0.0773, + "step": 103 + }, + { + "epoch": 0.04513888888888889, + "grad_norm": 6.506033897399902, + "learning_rate": 7.48201438848921e-06, + "loss": 0.1495, + "step": 104 + }, + { + "epoch": 0.045572916666666664, + "grad_norm": 2.1495561599731445, + "learning_rate": 7.5539568345323745e-06, + "loss": 0.1304, + "step": 105 + }, + { + "epoch": 0.04600694444444445, + "grad_norm": 2.843578577041626, + "learning_rate": 7.62589928057554e-06, + "loss": 0.1134, + "step": 106 + }, + { + "epoch": 0.046440972222222224, + "grad_norm": 2.714841842651367, + "learning_rate": 7.697841726618706e-06, + "loss": 0.0984, + "step": 107 + }, + { + "epoch": 0.046875, + "grad_norm": 2.3955094814300537, + "learning_rate": 7.769784172661872e-06, + "loss": 0.0714, + "step": 108 + }, + { + "epoch": 0.047309027777777776, + "grad_norm": 2.931508779525757, + "learning_rate": 7.841726618705036e-06, + "loss": 0.0862, + "step": 109 + }, + { + "epoch": 0.04774305555555555, + "grad_norm": 2.2287561893463135, + "learning_rate": 7.913669064748202e-06, + "loss": 0.0688, + "step": 110 + }, + { + "epoch": 0.048177083333333336, + "grad_norm": 4.007569313049316, + "learning_rate": 7.985611510791367e-06, + "loss": 0.1335, + "step": 111 + }, + { + "epoch": 0.04861111111111111, + "grad_norm": 2.308885097503662, + "learning_rate": 8.057553956834533e-06, + "loss": 0.0981, + "step": 112 + }, + { + "epoch": 0.04904513888888889, + "grad_norm": 3.60953950881958, + "learning_rate": 8.129496402877699e-06, + "loss": 0.0906, + "step": 113 + }, + { + "epoch": 0.049479166666666664, + "grad_norm": 3.5016674995422363, + "learning_rate": 8.201438848920865e-06, + "loss": 0.0946, + "step": 114 + }, + { + "epoch": 0.04991319444444445, + "grad_norm": 2.3107264041900635, + "learning_rate": 8.273381294964029e-06, + "loss": 0.1004, + "step": 115 + }, + { + "epoch": 0.050347222222222224, + "grad_norm": 2.496023654937744, + "learning_rate": 8.345323741007195e-06, + "loss": 0.1074, + "step": 116 + }, + { + "epoch": 0.05078125, + "grad_norm": 2.5439746379852295, + "learning_rate": 8.41726618705036e-06, + "loss": 0.1125, + "step": 117 + }, + { + "epoch": 0.051215277777777776, + "grad_norm": 2.115340232849121, + "learning_rate": 8.489208633093526e-06, + "loss": 0.0943, + "step": 118 + }, + { + "epoch": 0.05164930555555555, + "grad_norm": 1.8092048168182373, + "learning_rate": 8.561151079136692e-06, + "loss": 0.1043, + "step": 119 + }, + { + "epoch": 0.052083333333333336, + "grad_norm": 2.728424072265625, + "learning_rate": 8.633093525179856e-06, + "loss": 0.1212, + "step": 120 + }, + { + "epoch": 0.05251736111111111, + "grad_norm": 1.6297051906585693, + "learning_rate": 8.705035971223022e-06, + "loss": 0.0968, + "step": 121 + }, + { + "epoch": 0.05295138888888889, + "grad_norm": 1.4963327646255493, + "learning_rate": 8.776978417266188e-06, + "loss": 0.0692, + "step": 122 + }, + { + "epoch": 0.053385416666666664, + "grad_norm": 1.9528976678848267, + "learning_rate": 8.848920863309353e-06, + "loss": 0.0871, + "step": 123 + }, + { + "epoch": 0.05381944444444445, + "grad_norm": 3.2168524265289307, + "learning_rate": 8.92086330935252e-06, + "loss": 0.1252, + "step": 124 + }, + { + "epoch": 0.054253472222222224, + "grad_norm": 3.5068187713623047, + "learning_rate": 8.992805755395683e-06, + "loss": 0.1394, + "step": 125 + }, + { + "epoch": 0.0546875, + "grad_norm": 2.7131741046905518, + "learning_rate": 9.064748201438849e-06, + "loss": 0.1179, + "step": 126 + }, + { + "epoch": 0.055121527777777776, + "grad_norm": 2.037222146987915, + "learning_rate": 9.136690647482015e-06, + "loss": 0.0866, + "step": 127 + }, + { + "epoch": 0.05555555555555555, + "grad_norm": 1.1361604928970337, + "learning_rate": 9.20863309352518e-06, + "loss": 0.0829, + "step": 128 + }, + { + "epoch": 0.055989583333333336, + "grad_norm": 1.3983852863311768, + "learning_rate": 9.280575539568346e-06, + "loss": 0.1227, + "step": 129 + }, + { + "epoch": 0.05642361111111111, + "grad_norm": 1.1151443719863892, + "learning_rate": 9.35251798561151e-06, + "loss": 0.0963, + "step": 130 + }, + { + "epoch": 0.05685763888888889, + "grad_norm": 1.3063839673995972, + "learning_rate": 9.424460431654678e-06, + "loss": 0.0969, + "step": 131 + }, + { + "epoch": 0.057291666666666664, + "grad_norm": 1.619152307510376, + "learning_rate": 9.496402877697842e-06, + "loss": 0.1083, + "step": 132 + }, + { + "epoch": 0.05772569444444445, + "grad_norm": 1.5226192474365234, + "learning_rate": 9.568345323741008e-06, + "loss": 0.0849, + "step": 133 + }, + { + "epoch": 0.058159722222222224, + "grad_norm": 1.5232475996017456, + "learning_rate": 9.640287769784174e-06, + "loss": 0.0649, + "step": 134 + }, + { + "epoch": 0.05859375, + "grad_norm": 2.6182069778442383, + "learning_rate": 9.712230215827338e-06, + "loss": 0.0981, + "step": 135 + }, + { + "epoch": 0.059027777777777776, + "grad_norm": 2.3720853328704834, + "learning_rate": 9.784172661870505e-06, + "loss": 0.1011, + "step": 136 + }, + { + "epoch": 0.05946180555555555, + "grad_norm": 1.811148762702942, + "learning_rate": 9.85611510791367e-06, + "loss": 0.0983, + "step": 137 + }, + { + "epoch": 0.059895833333333336, + "grad_norm": 1.452284574508667, + "learning_rate": 9.928057553956835e-06, + "loss": 0.0656, + "step": 138 + }, + { + "epoch": 0.06032986111111111, + "grad_norm": 1.6373428106307983, + "learning_rate": 1e-05, + "loss": 0.1291, + "step": 139 + }, + { + "epoch": 0.06076388888888889, + "grad_norm": 1.5119938850402832, + "learning_rate": 9.999998764567487e-06, + "loss": 0.106, + "step": 140 + }, + { + "epoch": 0.061197916666666664, + "grad_norm": 1.1721326112747192, + "learning_rate": 9.999995058270555e-06, + "loss": 0.1004, + "step": 141 + }, + { + "epoch": 0.06163194444444445, + "grad_norm": 1.0162428617477417, + "learning_rate": 9.999988881111037e-06, + "loss": 0.0773, + "step": 142 + }, + { + "epoch": 0.062065972222222224, + "grad_norm": 4.524620532989502, + "learning_rate": 9.999980233091988e-06, + "loss": 0.0876, + "step": 143 + }, + { + "epoch": 0.0625, + "grad_norm": 2.312443733215332, + "learning_rate": 9.999969114217678e-06, + "loss": 0.1297, + "step": 144 + }, + { + "epoch": 0.06293402777777778, + "grad_norm": 2.083632707595825, + "learning_rate": 9.999955524493603e-06, + "loss": 0.1167, + "step": 145 + }, + { + "epoch": 0.06336805555555555, + "grad_norm": 1.0585899353027344, + "learning_rate": 9.99993946392648e-06, + "loss": 0.0778, + "step": 146 + }, + { + "epoch": 0.06380208333333333, + "grad_norm": 1.1303775310516357, + "learning_rate": 9.999920932524242e-06, + "loss": 0.1124, + "step": 147 + }, + { + "epoch": 0.0642361111111111, + "grad_norm": 1.113276720046997, + "learning_rate": 9.999899930296052e-06, + "loss": 0.1029, + "step": 148 + }, + { + "epoch": 0.0646701388888889, + "grad_norm": 1.1997543573379517, + "learning_rate": 9.999876457252281e-06, + "loss": 0.0998, + "step": 149 + }, + { + "epoch": 0.06510416666666667, + "grad_norm": 1.2771512269973755, + "learning_rate": 9.999850513404539e-06, + "loss": 0.109, + "step": 150 + }, + { + "epoch": 0.06553819444444445, + "grad_norm": 1.8055669069290161, + "learning_rate": 9.99982209876564e-06, + "loss": 0.1004, + "step": 151 + }, + { + "epoch": 0.06597222222222222, + "grad_norm": 1.3983486890792847, + "learning_rate": 9.999791213349624e-06, + "loss": 0.0636, + "step": 152 + }, + { + "epoch": 0.06640625, + "grad_norm": 1.934438705444336, + "learning_rate": 9.999757857171758e-06, + "loss": 0.1031, + "step": 153 + }, + { + "epoch": 0.06684027777777778, + "grad_norm": 1.6015806198120117, + "learning_rate": 9.999722030248526e-06, + "loss": 0.0763, + "step": 154 + }, + { + "epoch": 0.06727430555555555, + "grad_norm": 1.175926923751831, + "learning_rate": 9.999683732597631e-06, + "loss": 0.0813, + "step": 155 + }, + { + "epoch": 0.06770833333333333, + "grad_norm": 1.0643317699432373, + "learning_rate": 9.999642964238e-06, + "loss": 0.081, + "step": 156 + }, + { + "epoch": 0.0681423611111111, + "grad_norm": 1.2169849872589111, + "learning_rate": 9.999599725189777e-06, + "loss": 0.0734, + "step": 157 + }, + { + "epoch": 0.0685763888888889, + "grad_norm": 1.030631422996521, + "learning_rate": 9.99955401547433e-06, + "loss": 0.066, + "step": 158 + }, + { + "epoch": 0.06901041666666667, + "grad_norm": 1.4374827146530151, + "learning_rate": 9.999505835114252e-06, + "loss": 0.0832, + "step": 159 + }, + { + "epoch": 0.06944444444444445, + "grad_norm": 1.1652729511260986, + "learning_rate": 9.999455184133346e-06, + "loss": 0.0626, + "step": 160 + }, + { + "epoch": 0.06987847222222222, + "grad_norm": 2.172323226928711, + "learning_rate": 9.999402062556648e-06, + "loss": 0.1019, + "step": 161 + }, + { + "epoch": 0.0703125, + "grad_norm": 1.4461246728897095, + "learning_rate": 9.999346470410405e-06, + "loss": 0.0822, + "step": 162 + }, + { + "epoch": 0.07074652777777778, + "grad_norm": 1.3454128503799438, + "learning_rate": 9.999288407722089e-06, + "loss": 0.0975, + "step": 163 + }, + { + "epoch": 0.07118055555555555, + "grad_norm": 1.2311937808990479, + "learning_rate": 9.9992278745204e-06, + "loss": 0.0967, + "step": 164 + }, + { + "epoch": 0.07161458333333333, + "grad_norm": 0.9970878958702087, + "learning_rate": 9.999164870835241e-06, + "loss": 0.0988, + "step": 165 + }, + { + "epoch": 0.0720486111111111, + "grad_norm": 1.034321904182434, + "learning_rate": 9.999099396697755e-06, + "loss": 0.1032, + "step": 166 + }, + { + "epoch": 0.0724826388888889, + "grad_norm": 1.2639926671981812, + "learning_rate": 9.999031452140295e-06, + "loss": 0.0922, + "step": 167 + }, + { + "epoch": 0.07291666666666667, + "grad_norm": 0.9578790068626404, + "learning_rate": 9.998961037196437e-06, + "loss": 0.0644, + "step": 168 + }, + { + "epoch": 0.07335069444444445, + "grad_norm": 1.2277770042419434, + "learning_rate": 9.998888151900979e-06, + "loss": 0.0536, + "step": 169 + }, + { + "epoch": 0.07378472222222222, + "grad_norm": 2.1698157787323, + "learning_rate": 9.998812796289939e-06, + "loss": 0.1146, + "step": 170 + }, + { + "epoch": 0.07421875, + "grad_norm": 1.36558198928833, + "learning_rate": 9.998734970400554e-06, + "loss": 0.0612, + "step": 171 + }, + { + "epoch": 0.07465277777777778, + "grad_norm": 1.664151906967163, + "learning_rate": 9.998654674271285e-06, + "loss": 0.0615, + "step": 172 + }, + { + "epoch": 0.07508680555555555, + "grad_norm": 1.5223487615585327, + "learning_rate": 9.998571907941811e-06, + "loss": 0.0874, + "step": 173 + }, + { + "epoch": 0.07552083333333333, + "grad_norm": 1.261687994003296, + "learning_rate": 9.998486671453034e-06, + "loss": 0.0705, + "step": 174 + }, + { + "epoch": 0.0759548611111111, + "grad_norm": 1.260436773300171, + "learning_rate": 9.998398964847075e-06, + "loss": 0.0715, + "step": 175 + }, + { + "epoch": 0.0763888888888889, + "grad_norm": 0.9473410248756409, + "learning_rate": 9.998308788167277e-06, + "loss": 0.0807, + "step": 176 + }, + { + "epoch": 0.07682291666666667, + "grad_norm": 1.0099543333053589, + "learning_rate": 9.998216141458203e-06, + "loss": 0.0871, + "step": 177 + }, + { + "epoch": 0.07725694444444445, + "grad_norm": 0.9015682935714722, + "learning_rate": 9.998121024765635e-06, + "loss": 0.0744, + "step": 178 + }, + { + "epoch": 0.07769097222222222, + "grad_norm": 1.6776366233825684, + "learning_rate": 9.998023438136578e-06, + "loss": 0.0897, + "step": 179 + }, + { + "epoch": 0.078125, + "grad_norm": 1.5866495370864868, + "learning_rate": 9.997923381619257e-06, + "loss": 0.0935, + "step": 180 + }, + { + "epoch": 0.07855902777777778, + "grad_norm": 1.2076671123504639, + "learning_rate": 9.997820855263117e-06, + "loss": 0.089, + "step": 181 + }, + { + "epoch": 0.07899305555555555, + "grad_norm": 1.3449783325195312, + "learning_rate": 9.997715859118822e-06, + "loss": 0.0826, + "step": 182 + }, + { + "epoch": 0.07942708333333333, + "grad_norm": 1.6158901453018188, + "learning_rate": 9.99760839323826e-06, + "loss": 0.1154, + "step": 183 + }, + { + "epoch": 0.0798611111111111, + "grad_norm": 1.1235649585723877, + "learning_rate": 9.997498457674539e-06, + "loss": 0.0851, + "step": 184 + }, + { + "epoch": 0.0802951388888889, + "grad_norm": 0.9119734168052673, + "learning_rate": 9.997386052481985e-06, + "loss": 0.1149, + "step": 185 + }, + { + "epoch": 0.08072916666666667, + "grad_norm": 1.1313296556472778, + "learning_rate": 9.997271177716146e-06, + "loss": 0.073, + "step": 186 + }, + { + "epoch": 0.08116319444444445, + "grad_norm": 0.9562469720840454, + "learning_rate": 9.997153833433789e-06, + "loss": 0.0891, + "step": 187 + }, + { + "epoch": 0.08159722222222222, + "grad_norm": 0.916240930557251, + "learning_rate": 9.9970340196929e-06, + "loss": 0.1016, + "step": 188 + }, + { + "epoch": 0.08203125, + "grad_norm": 1.28097665309906, + "learning_rate": 9.996911736552694e-06, + "loss": 0.0879, + "step": 189 + }, + { + "epoch": 0.08246527777777778, + "grad_norm": 1.1469060182571411, + "learning_rate": 9.996786984073594e-06, + "loss": 0.0563, + "step": 190 + }, + { + "epoch": 0.08289930555555555, + "grad_norm": 3.393587112426758, + "learning_rate": 9.996659762317254e-06, + "loss": 0.1082, + "step": 191 + }, + { + "epoch": 0.08333333333333333, + "grad_norm": 3.045605421066284, + "learning_rate": 9.99653007134654e-06, + "loss": 0.0754, + "step": 192 + }, + { + "epoch": 0.0837673611111111, + "grad_norm": 1.3723241090774536, + "learning_rate": 9.996397911225544e-06, + "loss": 0.067, + "step": 193 + }, + { + "epoch": 0.0842013888888889, + "grad_norm": 1.394571304321289, + "learning_rate": 9.996263282019574e-06, + "loss": 0.0915, + "step": 194 + }, + { + "epoch": 0.08463541666666667, + "grad_norm": 1.0956934690475464, + "learning_rate": 9.996126183795162e-06, + "loss": 0.0637, + "step": 195 + }, + { + "epoch": 0.08506944444444445, + "grad_norm": 0.8720288872718811, + "learning_rate": 9.995986616620057e-06, + "loss": 0.0809, + "step": 196 + }, + { + "epoch": 0.08550347222222222, + "grad_norm": 1.0597174167633057, + "learning_rate": 9.995844580563232e-06, + "loss": 0.0856, + "step": 197 + }, + { + "epoch": 0.0859375, + "grad_norm": 1.317156195640564, + "learning_rate": 9.995700075694872e-06, + "loss": 0.0994, + "step": 198 + }, + { + "epoch": 0.08637152777777778, + "grad_norm": 1.026953935623169, + "learning_rate": 9.995553102086392e-06, + "loss": 0.0891, + "step": 199 + }, + { + "epoch": 0.08680555555555555, + "grad_norm": 1.7246663570404053, + "learning_rate": 9.995403659810422e-06, + "loss": 0.1019, + "step": 200 + }, + { + "epoch": 0.08723958333333333, + "grad_norm": 0.7544384598731995, + "learning_rate": 9.995251748940811e-06, + "loss": 0.0577, + "step": 201 + }, + { + "epoch": 0.0876736111111111, + "grad_norm": 0.9751232266426086, + "learning_rate": 9.995097369552628e-06, + "loss": 0.0864, + "step": 202 + }, + { + "epoch": 0.0881076388888889, + "grad_norm": 1.1838496923446655, + "learning_rate": 9.994940521722167e-06, + "loss": 0.0692, + "step": 203 + }, + { + "epoch": 0.08854166666666667, + "grad_norm": 1.2722941637039185, + "learning_rate": 9.994781205526935e-06, + "loss": 0.0854, + "step": 204 + }, + { + "epoch": 0.08897569444444445, + "grad_norm": 1.445184350013733, + "learning_rate": 9.994619421045661e-06, + "loss": 0.1057, + "step": 205 + }, + { + "epoch": 0.08940972222222222, + "grad_norm": 0.7705734968185425, + "learning_rate": 9.994455168358297e-06, + "loss": 0.0696, + "step": 206 + }, + { + "epoch": 0.08984375, + "grad_norm": 0.8373009562492371, + "learning_rate": 9.994288447546013e-06, + "loss": 0.0837, + "step": 207 + }, + { + "epoch": 0.09027777777777778, + "grad_norm": 0.6120887994766235, + "learning_rate": 9.994119258691192e-06, + "loss": 0.0622, + "step": 208 + }, + { + "epoch": 0.09071180555555555, + "grad_norm": 0.9396751523017883, + "learning_rate": 9.993947601877451e-06, + "loss": 0.078, + "step": 209 + }, + { + "epoch": 0.09114583333333333, + "grad_norm": 1.0563628673553467, + "learning_rate": 9.993773477189613e-06, + "loss": 0.0684, + "step": 210 + }, + { + "epoch": 0.0915798611111111, + "grad_norm": 1.0072035789489746, + "learning_rate": 9.993596884713724e-06, + "loss": 0.0708, + "step": 211 + }, + { + "epoch": 0.0920138888888889, + "grad_norm": 1.5425307750701904, + "learning_rate": 9.993417824537055e-06, + "loss": 0.0955, + "step": 212 + }, + { + "epoch": 0.09244791666666667, + "grad_norm": 1.181329369544983, + "learning_rate": 9.993236296748093e-06, + "loss": 0.0776, + "step": 213 + }, + { + "epoch": 0.09288194444444445, + "grad_norm": 0.8895711898803711, + "learning_rate": 9.99305230143654e-06, + "loss": 0.0695, + "step": 214 + }, + { + "epoch": 0.09331597222222222, + "grad_norm": 1.0264010429382324, + "learning_rate": 9.992865838693328e-06, + "loss": 0.091, + "step": 215 + }, + { + "epoch": 0.09375, + "grad_norm": 0.741313099861145, + "learning_rate": 9.992676908610595e-06, + "loss": 0.0735, + "step": 216 + }, + { + "epoch": 0.09418402777777778, + "grad_norm": 0.715937614440918, + "learning_rate": 9.99248551128171e-06, + "loss": 0.0908, + "step": 217 + }, + { + "epoch": 0.09461805555555555, + "grad_norm": 0.7246452569961548, + "learning_rate": 9.992291646801253e-06, + "loss": 0.0793, + "step": 218 + }, + { + "epoch": 0.09505208333333333, + "grad_norm": 0.8316558003425598, + "learning_rate": 9.99209531526503e-06, + "loss": 0.0854, + "step": 219 + }, + { + "epoch": 0.0954861111111111, + "grad_norm": 0.8730294704437256, + "learning_rate": 9.991896516770058e-06, + "loss": 0.0746, + "step": 220 + }, + { + "epoch": 0.0959201388888889, + "grad_norm": 0.8386092185974121, + "learning_rate": 9.991695251414584e-06, + "loss": 0.0619, + "step": 221 + }, + { + "epoch": 0.09635416666666667, + "grad_norm": 1.0405433177947998, + "learning_rate": 9.991491519298063e-06, + "loss": 0.0892, + "step": 222 + }, + { + "epoch": 0.09678819444444445, + "grad_norm": 1.1907081604003906, + "learning_rate": 9.991285320521176e-06, + "loss": 0.0581, + "step": 223 + }, + { + "epoch": 0.09722222222222222, + "grad_norm": 2.635026693344116, + "learning_rate": 9.99107665518582e-06, + "loss": 0.0631, + "step": 224 + }, + { + "epoch": 0.09765625, + "grad_norm": 1.6267889738082886, + "learning_rate": 9.990865523395113e-06, + "loss": 0.0853, + "step": 225 + }, + { + "epoch": 0.09809027777777778, + "grad_norm": 0.743120551109314, + "learning_rate": 9.990651925253389e-06, + "loss": 0.0472, + "step": 226 + }, + { + "epoch": 0.09852430555555555, + "grad_norm": 1.1194217205047607, + "learning_rate": 9.990435860866204e-06, + "loss": 0.102, + "step": 227 + }, + { + "epoch": 0.09895833333333333, + "grad_norm": 0.7726638913154602, + "learning_rate": 9.99021733034033e-06, + "loss": 0.0681, + "step": 228 + }, + { + "epoch": 0.0993923611111111, + "grad_norm": 1.217424750328064, + "learning_rate": 9.98999633378376e-06, + "loss": 0.0882, + "step": 229 + }, + { + "epoch": 0.0998263888888889, + "grad_norm": 0.8821778297424316, + "learning_rate": 9.989772871305704e-06, + "loss": 0.0991, + "step": 230 + }, + { + "epoch": 0.10026041666666667, + "grad_norm": 1.0021941661834717, + "learning_rate": 9.98954694301659e-06, + "loss": 0.0786, + "step": 231 + }, + { + "epoch": 0.10069444444444445, + "grad_norm": 0.8431205749511719, + "learning_rate": 9.989318549028068e-06, + "loss": 0.0872, + "step": 232 + }, + { + "epoch": 0.10112847222222222, + "grad_norm": 0.9266021847724915, + "learning_rate": 9.989087689453001e-06, + "loss": 0.0735, + "step": 233 + }, + { + "epoch": 0.1015625, + "grad_norm": 1.1363204717636108, + "learning_rate": 9.988854364405477e-06, + "loss": 0.0777, + "step": 234 + }, + { + "epoch": 0.10199652777777778, + "grad_norm": 0.9753544330596924, + "learning_rate": 9.988618574000796e-06, + "loss": 0.069, + "step": 235 + }, + { + "epoch": 0.10243055555555555, + "grad_norm": 1.2155951261520386, + "learning_rate": 9.988380318355483e-06, + "loss": 0.0821, + "step": 236 + }, + { + "epoch": 0.10286458333333333, + "grad_norm": 0.957694947719574, + "learning_rate": 9.988139597587274e-06, + "loss": 0.0602, + "step": 237 + }, + { + "epoch": 0.1032986111111111, + "grad_norm": 0.958967387676239, + "learning_rate": 9.987896411815127e-06, + "loss": 0.0708, + "step": 238 + }, + { + "epoch": 0.1037326388888889, + "grad_norm": 1.111096978187561, + "learning_rate": 9.987650761159218e-06, + "loss": 0.0829, + "step": 239 + }, + { + "epoch": 0.10416666666666667, + "grad_norm": 0.9012123942375183, + "learning_rate": 9.987402645740943e-06, + "loss": 0.0839, + "step": 240 + }, + { + "epoch": 0.10460069444444445, + "grad_norm": 0.6766245365142822, + "learning_rate": 9.987152065682913e-06, + "loss": 0.0814, + "step": 241 + }, + { + "epoch": 0.10503472222222222, + "grad_norm": 0.7312656044960022, + "learning_rate": 9.986899021108957e-06, + "loss": 0.0674, + "step": 242 + }, + { + "epoch": 0.10546875, + "grad_norm": 1.022851824760437, + "learning_rate": 9.986643512144122e-06, + "loss": 0.0824, + "step": 243 + }, + { + "epoch": 0.10590277777777778, + "grad_norm": 0.8258888125419617, + "learning_rate": 9.986385538914676e-06, + "loss": 0.0866, + "step": 244 + }, + { + "epoch": 0.10633680555555555, + "grad_norm": 0.804434597492218, + "learning_rate": 9.986125101548101e-06, + "loss": 0.0868, + "step": 245 + }, + { + "epoch": 0.10677083333333333, + "grad_norm": 0.7631927132606506, + "learning_rate": 9.9858622001731e-06, + "loss": 0.0798, + "step": 246 + }, + { + "epoch": 0.1072048611111111, + "grad_norm": 0.703942596912384, + "learning_rate": 9.985596834919587e-06, + "loss": 0.0598, + "step": 247 + }, + { + "epoch": 0.1076388888888889, + "grad_norm": 0.9682087898254395, + "learning_rate": 9.985329005918702e-06, + "loss": 0.0817, + "step": 248 + }, + { + "epoch": 0.10807291666666667, + "grad_norm": 1.005095362663269, + "learning_rate": 9.9850587133028e-06, + "loss": 0.0734, + "step": 249 + }, + { + "epoch": 0.10850694444444445, + "grad_norm": 1.0828065872192383, + "learning_rate": 9.98478595720545e-06, + "loss": 0.0788, + "step": 250 + }, + { + "epoch": 0.10894097222222222, + "grad_norm": 1.217239499092102, + "learning_rate": 9.984510737761441e-06, + "loss": 0.1023, + "step": 251 + }, + { + "epoch": 0.109375, + "grad_norm": 0.8663279414176941, + "learning_rate": 9.98423305510678e-06, + "loss": 0.0839, + "step": 252 + }, + { + "epoch": 0.10980902777777778, + "grad_norm": 0.8627954125404358, + "learning_rate": 9.98395290937869e-06, + "loss": 0.0661, + "step": 253 + }, + { + "epoch": 0.11024305555555555, + "grad_norm": 1.0426639318466187, + "learning_rate": 9.98367030071561e-06, + "loss": 0.0921, + "step": 254 + }, + { + "epoch": 0.11067708333333333, + "grad_norm": 0.7946957945823669, + "learning_rate": 9.983385229257201e-06, + "loss": 0.0455, + "step": 255 + }, + { + "epoch": 0.1111111111111111, + "grad_norm": 0.8853251338005066, + "learning_rate": 9.983097695144332e-06, + "loss": 0.0662, + "step": 256 + }, + { + "epoch": 0.1115451388888889, + "grad_norm": 0.9303094744682312, + "learning_rate": 9.982807698519101e-06, + "loss": 0.0709, + "step": 257 + }, + { + "epoch": 0.11197916666666667, + "grad_norm": 2.260749101638794, + "learning_rate": 9.982515239524811e-06, + "loss": 0.08, + "step": 258 + }, + { + "epoch": 0.11241319444444445, + "grad_norm": 1.252661108970642, + "learning_rate": 9.98222031830599e-06, + "loss": 0.0714, + "step": 259 + }, + { + "epoch": 0.11284722222222222, + "grad_norm": 0.9974795579910278, + "learning_rate": 9.981922935008381e-06, + "loss": 0.0602, + "step": 260 + }, + { + "epoch": 0.11328125, + "grad_norm": 1.5989861488342285, + "learning_rate": 9.98162308977894e-06, + "loss": 0.0771, + "step": 261 + }, + { + "epoch": 0.11371527777777778, + "grad_norm": 2.777109384536743, + "learning_rate": 9.981320782765847e-06, + "loss": 0.1554, + "step": 262 + }, + { + "epoch": 0.11414930555555555, + "grad_norm": 1.3210549354553223, + "learning_rate": 9.98101601411849e-06, + "loss": 0.0628, + "step": 263 + }, + { + "epoch": 0.11458333333333333, + "grad_norm": 1.3631166219711304, + "learning_rate": 9.980708783987476e-06, + "loss": 0.0743, + "step": 264 + }, + { + "epoch": 0.1150173611111111, + "grad_norm": 1.32863450050354, + "learning_rate": 9.980399092524635e-06, + "loss": 0.1134, + "step": 265 + }, + { + "epoch": 0.1154513888888889, + "grad_norm": 0.6396360993385315, + "learning_rate": 9.980086939883004e-06, + "loss": 0.0687, + "step": 266 + }, + { + "epoch": 0.11588541666666667, + "grad_norm": 0.9112362861633301, + "learning_rate": 9.979772326216842e-06, + "loss": 0.0576, + "step": 267 + }, + { + "epoch": 0.11631944444444445, + "grad_norm": 0.8859538435935974, + "learning_rate": 9.979455251681622e-06, + "loss": 0.0751, + "step": 268 + }, + { + "epoch": 0.11675347222222222, + "grad_norm": 1.0131158828735352, + "learning_rate": 9.979135716434036e-06, + "loss": 0.0616, + "step": 269 + }, + { + "epoch": 0.1171875, + "grad_norm": 0.7546622157096863, + "learning_rate": 9.978813720631986e-06, + "loss": 0.0537, + "step": 270 + }, + { + "epoch": 0.11762152777777778, + "grad_norm": 1.2106804847717285, + "learning_rate": 9.978489264434597e-06, + "loss": 0.076, + "step": 271 + }, + { + "epoch": 0.11805555555555555, + "grad_norm": 0.7226711511611938, + "learning_rate": 9.978162348002205e-06, + "loss": 0.0513, + "step": 272 + }, + { + "epoch": 0.11848958333333333, + "grad_norm": 1.3641725778579712, + "learning_rate": 9.977832971496362e-06, + "loss": 0.084, + "step": 273 + }, + { + "epoch": 0.1189236111111111, + "grad_norm": 1.5535808801651, + "learning_rate": 9.977501135079839e-06, + "loss": 0.097, + "step": 274 + }, + { + "epoch": 0.1193576388888889, + "grad_norm": 0.8176865577697754, + "learning_rate": 9.977166838916619e-06, + "loss": 0.0646, + "step": 275 + }, + { + "epoch": 0.11979166666666667, + "grad_norm": 1.3481011390686035, + "learning_rate": 9.976830083171903e-06, + "loss": 0.0823, + "step": 276 + }, + { + "epoch": 0.12022569444444445, + "grad_norm": 1.4268255233764648, + "learning_rate": 9.976490868012108e-06, + "loss": 0.101, + "step": 277 + }, + { + "epoch": 0.12065972222222222, + "grad_norm": 1.1602445840835571, + "learning_rate": 9.976149193604862e-06, + "loss": 0.0882, + "step": 278 + }, + { + "epoch": 0.12109375, + "grad_norm": 0.6257598996162415, + "learning_rate": 9.975805060119013e-06, + "loss": 0.0558, + "step": 279 + }, + { + "epoch": 0.12152777777777778, + "grad_norm": 0.7104304432868958, + "learning_rate": 9.975458467724624e-06, + "loss": 0.079, + "step": 280 + }, + { + "epoch": 0.12196180555555555, + "grad_norm": 0.691241443157196, + "learning_rate": 9.975109416592967e-06, + "loss": 0.0705, + "step": 281 + }, + { + "epoch": 0.12239583333333333, + "grad_norm": 0.8947821855545044, + "learning_rate": 9.974757906896539e-06, + "loss": 0.0751, + "step": 282 + }, + { + "epoch": 0.1228298611111111, + "grad_norm": 0.8744166493415833, + "learning_rate": 9.974403938809044e-06, + "loss": 0.1026, + "step": 283 + }, + { + "epoch": 0.1232638888888889, + "grad_norm": 1.5625134706497192, + "learning_rate": 9.974047512505401e-06, + "loss": 0.0952, + "step": 284 + }, + { + "epoch": 0.12369791666666667, + "grad_norm": 1.6969810724258423, + "learning_rate": 9.97368862816175e-06, + "loss": 0.1144, + "step": 285 + }, + { + "epoch": 0.12413194444444445, + "grad_norm": 1.002626895904541, + "learning_rate": 9.973327285955442e-06, + "loss": 0.068, + "step": 286 + }, + { + "epoch": 0.12456597222222222, + "grad_norm": 0.8302503228187561, + "learning_rate": 9.972963486065042e-06, + "loss": 0.0987, + "step": 287 + }, + { + "epoch": 0.125, + "grad_norm": 0.6188612580299377, + "learning_rate": 9.972597228670327e-06, + "loss": 0.0643, + "step": 288 + }, + { + "epoch": 0.1254340277777778, + "grad_norm": 0.9741168022155762, + "learning_rate": 9.972228513952295e-06, + "loss": 0.0579, + "step": 289 + }, + { + "epoch": 0.12586805555555555, + "grad_norm": 1.0956751108169556, + "learning_rate": 9.971857342093154e-06, + "loss": 0.0529, + "step": 290 + }, + { + "epoch": 0.12630208333333334, + "grad_norm": 1.3885653018951416, + "learning_rate": 9.971483713276326e-06, + "loss": 0.0843, + "step": 291 + }, + { + "epoch": 0.1267361111111111, + "grad_norm": 0.7927285432815552, + "learning_rate": 9.971107627686452e-06, + "loss": 0.0587, + "step": 292 + }, + { + "epoch": 0.1271701388888889, + "grad_norm": 0.9241272807121277, + "learning_rate": 9.970729085509379e-06, + "loss": 0.0531, + "step": 293 + }, + { + "epoch": 0.12760416666666666, + "grad_norm": 1.4300391674041748, + "learning_rate": 9.970348086932174e-06, + "loss": 0.1115, + "step": 294 + }, + { + "epoch": 0.12803819444444445, + "grad_norm": 0.7712482213973999, + "learning_rate": 9.969964632143114e-06, + "loss": 0.0469, + "step": 295 + }, + { + "epoch": 0.1284722222222222, + "grad_norm": 0.7699704766273499, + "learning_rate": 9.969578721331696e-06, + "loss": 0.0685, + "step": 296 + }, + { + "epoch": 0.12890625, + "grad_norm": 0.6662436127662659, + "learning_rate": 9.969190354688625e-06, + "loss": 0.0694, + "step": 297 + }, + { + "epoch": 0.1293402777777778, + "grad_norm": 0.7108180522918701, + "learning_rate": 9.968799532405823e-06, + "loss": 0.0681, + "step": 298 + }, + { + "epoch": 0.12977430555555555, + "grad_norm": 1.0249931812286377, + "learning_rate": 9.968406254676418e-06, + "loss": 0.0977, + "step": 299 + }, + { + "epoch": 0.13020833333333334, + "grad_norm": 0.7156983017921448, + "learning_rate": 9.968010521694762e-06, + "loss": 0.0662, + "step": 300 + }, + { + "epoch": 0.13020833333333334, + "eval_layout_rm_rank_test_accuracy": 0.9643132637732938, + "eval_layout_rm_rank_test_chosen_scores": 0.6622011065483093, + "eval_layout_rm_rank_test_loss": 0.07006000727415085, + "eval_layout_rm_rank_test_rejected_scores": 0.12240368872880936, + "eval_layout_rm_rank_test_runtime": 1177.9284, + "eval_layout_rm_rank_test_samples_per_second": 20.91, + "eval_layout_rm_rank_test_steps_per_second": 0.654, + "step": 300 + }, + { + "epoch": 0.1306423611111111, + "grad_norm": 0.8029769659042358, + "learning_rate": 9.967612333656417e-06, + "loss": 0.0798, + "step": 301 + }, + { + "epoch": 0.1310763888888889, + "grad_norm": 0.6051777601242065, + "learning_rate": 9.967211690758154e-06, + "loss": 0.038, + "step": 302 + }, + { + "epoch": 0.13151041666666666, + "grad_norm": 0.7547597289085388, + "learning_rate": 9.966808593197959e-06, + "loss": 0.0705, + "step": 303 + }, + { + "epoch": 0.13194444444444445, + "grad_norm": 1.097530484199524, + "learning_rate": 9.966403041175034e-06, + "loss": 0.0771, + "step": 304 + }, + { + "epoch": 0.1323784722222222, + "grad_norm": 1.216100811958313, + "learning_rate": 9.965995034889792e-06, + "loss": 0.0835, + "step": 305 + }, + { + "epoch": 0.1328125, + "grad_norm": 1.5321711301803589, + "learning_rate": 9.965584574543857e-06, + "loss": 0.0677, + "step": 306 + }, + { + "epoch": 0.1332465277777778, + "grad_norm": 1.0033652782440186, + "learning_rate": 9.96517166034007e-06, + "loss": 0.0707, + "step": 307 + }, + { + "epoch": 0.13368055555555555, + "grad_norm": 0.8167241215705872, + "learning_rate": 9.964756292482479e-06, + "loss": 0.0805, + "step": 308 + }, + { + "epoch": 0.13411458333333334, + "grad_norm": 0.8531818389892578, + "learning_rate": 9.964338471176349e-06, + "loss": 0.0721, + "step": 309 + }, + { + "epoch": 0.1345486111111111, + "grad_norm": 0.6103355884552002, + "learning_rate": 9.963918196628155e-06, + "loss": 0.049, + "step": 310 + }, + { + "epoch": 0.1349826388888889, + "grad_norm": 1.079995036125183, + "learning_rate": 9.963495469045587e-06, + "loss": 0.0943, + "step": 311 + }, + { + "epoch": 0.13541666666666666, + "grad_norm": 0.9208278656005859, + "learning_rate": 9.963070288637547e-06, + "loss": 0.0703, + "step": 312 + }, + { + "epoch": 0.13585069444444445, + "grad_norm": 0.7477347254753113, + "learning_rate": 9.962642655614143e-06, + "loss": 0.0477, + "step": 313 + }, + { + "epoch": 0.1362847222222222, + "grad_norm": 1.0080677270889282, + "learning_rate": 9.962212570186702e-06, + "loss": 0.0738, + "step": 314 + }, + { + "epoch": 0.13671875, + "grad_norm": 1.0213791131973267, + "learning_rate": 9.961780032567763e-06, + "loss": 0.0788, + "step": 315 + }, + { + "epoch": 0.1371527777777778, + "grad_norm": 1.3381445407867432, + "learning_rate": 9.961345042971071e-06, + "loss": 0.0824, + "step": 316 + }, + { + "epoch": 0.13758680555555555, + "grad_norm": 1.2868865728378296, + "learning_rate": 9.960907601611587e-06, + "loss": 0.099, + "step": 317 + }, + { + "epoch": 0.13802083333333334, + "grad_norm": 1.0978431701660156, + "learning_rate": 9.960467708705484e-06, + "loss": 0.0673, + "step": 318 + }, + { + "epoch": 0.1384548611111111, + "grad_norm": 0.8579413890838623, + "learning_rate": 9.960025364470145e-06, + "loss": 0.0579, + "step": 319 + }, + { + "epoch": 0.1388888888888889, + "grad_norm": 1.3059312105178833, + "learning_rate": 9.959580569124164e-06, + "loss": 0.1266, + "step": 320 + }, + { + "epoch": 0.13932291666666666, + "grad_norm": 0.8588614463806152, + "learning_rate": 9.959133322887345e-06, + "loss": 0.0716, + "step": 321 + }, + { + "epoch": 0.13975694444444445, + "grad_norm": 0.669397234916687, + "learning_rate": 9.958683625980708e-06, + "loss": 0.0721, + "step": 322 + }, + { + "epoch": 0.1401909722222222, + "grad_norm": 0.7610234022140503, + "learning_rate": 9.958231478626477e-06, + "loss": 0.0773, + "step": 323 + }, + { + "epoch": 0.140625, + "grad_norm": 0.6288635730743408, + "learning_rate": 9.957776881048097e-06, + "loss": 0.0824, + "step": 324 + }, + { + "epoch": 0.1410590277777778, + "grad_norm": 0.8414584994316101, + "learning_rate": 9.957319833470214e-06, + "loss": 0.0883, + "step": 325 + }, + { + "epoch": 0.14149305555555555, + "grad_norm": 0.6574875712394714, + "learning_rate": 9.95686033611869e-06, + "loss": 0.0778, + "step": 326 + }, + { + "epoch": 0.14192708333333334, + "grad_norm": 0.8761506676673889, + "learning_rate": 9.956398389220594e-06, + "loss": 0.0641, + "step": 327 + }, + { + "epoch": 0.1423611111111111, + "grad_norm": 0.7599236369132996, + "learning_rate": 9.95593399300421e-06, + "loss": 0.0696, + "step": 328 + }, + { + "epoch": 0.1427951388888889, + "grad_norm": 0.8212679624557495, + "learning_rate": 9.95546714769903e-06, + "loss": 0.0717, + "step": 329 + }, + { + "epoch": 0.14322916666666666, + "grad_norm": 0.7221580147743225, + "learning_rate": 9.954997853535754e-06, + "loss": 0.0418, + "step": 330 + }, + { + "epoch": 0.14366319444444445, + "grad_norm": 0.7911623120307922, + "learning_rate": 9.954526110746297e-06, + "loss": 0.0755, + "step": 331 + }, + { + "epoch": 0.1440972222222222, + "grad_norm": 0.7278416156768799, + "learning_rate": 9.95405191956378e-06, + "loss": 0.0622, + "step": 332 + }, + { + "epoch": 0.14453125, + "grad_norm": 0.8105444312095642, + "learning_rate": 9.953575280222537e-06, + "loss": 0.0554, + "step": 333 + }, + { + "epoch": 0.1449652777777778, + "grad_norm": 1.0872361660003662, + "learning_rate": 9.953096192958109e-06, + "loss": 0.0672, + "step": 334 + }, + { + "epoch": 0.14539930555555555, + "grad_norm": 1.1132135391235352, + "learning_rate": 9.952614658007247e-06, + "loss": 0.0759, + "step": 335 + }, + { + "epoch": 0.14583333333333334, + "grad_norm": 0.6996872425079346, + "learning_rate": 9.952130675607917e-06, + "loss": 0.057, + "step": 336 + }, + { + "epoch": 0.1462673611111111, + "grad_norm": 0.7257365584373474, + "learning_rate": 9.951644245999285e-06, + "loss": 0.0601, + "step": 337 + }, + { + "epoch": 0.1467013888888889, + "grad_norm": 1.0367202758789062, + "learning_rate": 9.951155369421733e-06, + "loss": 0.0853, + "step": 338 + }, + { + "epoch": 0.14713541666666666, + "grad_norm": 1.2991832494735718, + "learning_rate": 9.950664046116852e-06, + "loss": 0.0716, + "step": 339 + }, + { + "epoch": 0.14756944444444445, + "grad_norm": 0.9172611832618713, + "learning_rate": 9.95017027632744e-06, + "loss": 0.0646, + "step": 340 + }, + { + "epoch": 0.1480034722222222, + "grad_norm": 0.8245013952255249, + "learning_rate": 9.949674060297504e-06, + "loss": 0.0854, + "step": 341 + }, + { + "epoch": 0.1484375, + "grad_norm": 0.7438561320304871, + "learning_rate": 9.949175398272259e-06, + "loss": 0.0536, + "step": 342 + }, + { + "epoch": 0.1488715277777778, + "grad_norm": 0.6299500465393066, + "learning_rate": 9.948674290498134e-06, + "loss": 0.0459, + "step": 343 + }, + { + "epoch": 0.14930555555555555, + "grad_norm": 0.7602421641349792, + "learning_rate": 9.948170737222763e-06, + "loss": 0.078, + "step": 344 + }, + { + "epoch": 0.14973958333333334, + "grad_norm": 0.6525736451148987, + "learning_rate": 9.947664738694985e-06, + "loss": 0.0601, + "step": 345 + }, + { + "epoch": 0.1501736111111111, + "grad_norm": 1.00484299659729, + "learning_rate": 9.947156295164853e-06, + "loss": 0.0675, + "step": 346 + }, + { + "epoch": 0.1506076388888889, + "grad_norm": 0.8091078996658325, + "learning_rate": 9.946645406883625e-06, + "loss": 0.051, + "step": 347 + }, + { + "epoch": 0.15104166666666666, + "grad_norm": 0.7195174098014832, + "learning_rate": 9.946132074103769e-06, + "loss": 0.0546, + "step": 348 + }, + { + "epoch": 0.15147569444444445, + "grad_norm": 2.212528944015503, + "learning_rate": 9.94561629707896e-06, + "loss": 0.105, + "step": 349 + }, + { + "epoch": 0.1519097222222222, + "grad_norm": 1.7493406534194946, + "learning_rate": 9.94509807606408e-06, + "loss": 0.1233, + "step": 350 + }, + { + "epoch": 0.15234375, + "grad_norm": 1.3774362802505493, + "learning_rate": 9.944577411315222e-06, + "loss": 0.0697, + "step": 351 + }, + { + "epoch": 0.1527777777777778, + "grad_norm": 0.7310093641281128, + "learning_rate": 9.944054303089682e-06, + "loss": 0.0711, + "step": 352 + }, + { + "epoch": 0.15321180555555555, + "grad_norm": 0.8248127698898315, + "learning_rate": 9.94352875164597e-06, + "loss": 0.0725, + "step": 353 + }, + { + "epoch": 0.15364583333333334, + "grad_norm": 0.822748601436615, + "learning_rate": 9.943000757243795e-06, + "loss": 0.0709, + "step": 354 + }, + { + "epoch": 0.1540798611111111, + "grad_norm": 0.6451482772827148, + "learning_rate": 9.94247032014408e-06, + "loss": 0.0572, + "step": 355 + }, + { + "epoch": 0.1545138888888889, + "grad_norm": 0.6533085703849792, + "learning_rate": 9.941937440608951e-06, + "loss": 0.0712, + "step": 356 + }, + { + "epoch": 0.15494791666666666, + "grad_norm": 0.6614604592323303, + "learning_rate": 9.941402118901743e-06, + "loss": 0.0603, + "step": 357 + }, + { + "epoch": 0.15538194444444445, + "grad_norm": 1.1059565544128418, + "learning_rate": 9.940864355286999e-06, + "loss": 0.0697, + "step": 358 + }, + { + "epoch": 0.1558159722222222, + "grad_norm": 0.7619580626487732, + "learning_rate": 9.940324150030468e-06, + "loss": 0.0615, + "step": 359 + }, + { + "epoch": 0.15625, + "grad_norm": 1.6438034772872925, + "learning_rate": 9.9397815033991e-06, + "loss": 0.0975, + "step": 360 + }, + { + "epoch": 0.1566840277777778, + "grad_norm": 0.9065143465995789, + "learning_rate": 9.939236415661063e-06, + "loss": 0.0597, + "step": 361 + }, + { + "epoch": 0.15711805555555555, + "grad_norm": 0.8576701283454895, + "learning_rate": 9.938688887085717e-06, + "loss": 0.0656, + "step": 362 + }, + { + "epoch": 0.15755208333333334, + "grad_norm": 0.8027536273002625, + "learning_rate": 9.938138917943643e-06, + "loss": 0.0557, + "step": 363 + }, + { + "epoch": 0.1579861111111111, + "grad_norm": 0.8124869465827942, + "learning_rate": 9.937586508506617e-06, + "loss": 0.0781, + "step": 364 + }, + { + "epoch": 0.1584201388888889, + "grad_norm": 0.7721472382545471, + "learning_rate": 9.937031659047627e-06, + "loss": 0.0674, + "step": 365 + }, + { + "epoch": 0.15885416666666666, + "grad_norm": 0.7477177381515503, + "learning_rate": 9.936474369840862e-06, + "loss": 0.0634, + "step": 366 + }, + { + "epoch": 0.15928819444444445, + "grad_norm": 1.119555115699768, + "learning_rate": 9.93591464116172e-06, + "loss": 0.0763, + "step": 367 + }, + { + "epoch": 0.1597222222222222, + "grad_norm": 1.1497366428375244, + "learning_rate": 9.935352473286808e-06, + "loss": 0.0625, + "step": 368 + }, + { + "epoch": 0.16015625, + "grad_norm": 1.4596697092056274, + "learning_rate": 9.934787866493929e-06, + "loss": 0.09, + "step": 369 + }, + { + "epoch": 0.1605902777777778, + "grad_norm": 0.7065611481666565, + "learning_rate": 9.934220821062097e-06, + "loss": 0.0558, + "step": 370 + }, + { + "epoch": 0.16102430555555555, + "grad_norm": 0.850710391998291, + "learning_rate": 9.933651337271532e-06, + "loss": 0.0661, + "step": 371 + }, + { + "epoch": 0.16145833333333334, + "grad_norm": 0.637713611125946, + "learning_rate": 9.933079415403657e-06, + "loss": 0.0591, + "step": 372 + }, + { + "epoch": 0.1618923611111111, + "grad_norm": 0.6183542609214783, + "learning_rate": 9.932505055741103e-06, + "loss": 0.0535, + "step": 373 + }, + { + "epoch": 0.1623263888888889, + "grad_norm": 0.5929973125457764, + "learning_rate": 9.931928258567697e-06, + "loss": 0.0536, + "step": 374 + }, + { + "epoch": 0.16276041666666666, + "grad_norm": 0.8810651302337646, + "learning_rate": 9.931349024168483e-06, + "loss": 0.0658, + "step": 375 + }, + { + "epoch": 0.16319444444444445, + "grad_norm": 1.0665572881698608, + "learning_rate": 9.930767352829698e-06, + "loss": 0.0669, + "step": 376 + }, + { + "epoch": 0.1636284722222222, + "grad_norm": 0.6635595560073853, + "learning_rate": 9.930183244838792e-06, + "loss": 0.0575, + "step": 377 + }, + { + "epoch": 0.1640625, + "grad_norm": 1.1162075996398926, + "learning_rate": 9.929596700484412e-06, + "loss": 0.0806, + "step": 378 + }, + { + "epoch": 0.1644965277777778, + "grad_norm": 0.7197816371917725, + "learning_rate": 9.929007720056417e-06, + "loss": 0.0418, + "step": 379 + }, + { + "epoch": 0.16493055555555555, + "grad_norm": 1.1181401014328003, + "learning_rate": 9.92841630384586e-06, + "loss": 0.0798, + "step": 380 + }, + { + "epoch": 0.16536458333333334, + "grad_norm": 0.9076238870620728, + "learning_rate": 9.927822452145007e-06, + "loss": 0.0785, + "step": 381 + }, + { + "epoch": 0.1657986111111111, + "grad_norm": 0.8867773413658142, + "learning_rate": 9.92722616524732e-06, + "loss": 0.0604, + "step": 382 + }, + { + "epoch": 0.1662326388888889, + "grad_norm": 0.8934979438781738, + "learning_rate": 9.926627443447472e-06, + "loss": 0.0479, + "step": 383 + }, + { + "epoch": 0.16666666666666666, + "grad_norm": 0.801541268825531, + "learning_rate": 9.926026287041331e-06, + "loss": 0.0488, + "step": 384 + }, + { + "epoch": 0.16710069444444445, + "grad_norm": 0.7244776487350464, + "learning_rate": 9.925422696325976e-06, + "loss": 0.0615, + "step": 385 + }, + { + "epoch": 0.1675347222222222, + "grad_norm": 1.1914883852005005, + "learning_rate": 9.924816671599682e-06, + "loss": 0.0755, + "step": 386 + }, + { + "epoch": 0.16796875, + "grad_norm": 1.536291480064392, + "learning_rate": 9.924208213161932e-06, + "loss": 0.0713, + "step": 387 + }, + { + "epoch": 0.1684027777777778, + "grad_norm": 1.358291506767273, + "learning_rate": 9.92359732131341e-06, + "loss": 0.0888, + "step": 388 + }, + { + "epoch": 0.16883680555555555, + "grad_norm": 0.885543942451477, + "learning_rate": 9.922983996355999e-06, + "loss": 0.0556, + "step": 389 + }, + { + "epoch": 0.16927083333333334, + "grad_norm": 0.765248715877533, + "learning_rate": 9.922368238592792e-06, + "loss": 0.0585, + "step": 390 + }, + { + "epoch": 0.1697048611111111, + "grad_norm": 0.8303397297859192, + "learning_rate": 9.921750048328078e-06, + "loss": 0.0751, + "step": 391 + }, + { + "epoch": 0.1701388888888889, + "grad_norm": 0.8853304386138916, + "learning_rate": 9.92112942586735e-06, + "loss": 0.0598, + "step": 392 + }, + { + "epoch": 0.17057291666666666, + "grad_norm": 0.7674742341041565, + "learning_rate": 9.920506371517303e-06, + "loss": 0.0663, + "step": 393 + }, + { + "epoch": 0.17100694444444445, + "grad_norm": 0.9201746582984924, + "learning_rate": 9.919880885585833e-06, + "loss": 0.0674, + "step": 394 + }, + { + "epoch": 0.1714409722222222, + "grad_norm": 0.9129164814949036, + "learning_rate": 9.919252968382039e-06, + "loss": 0.0605, + "step": 395 + }, + { + "epoch": 0.171875, + "grad_norm": 0.8458436727523804, + "learning_rate": 9.918622620216218e-06, + "loss": 0.0425, + "step": 396 + }, + { + "epoch": 0.1723090277777778, + "grad_norm": 0.9405080080032349, + "learning_rate": 9.917989841399875e-06, + "loss": 0.0839, + "step": 397 + }, + { + "epoch": 0.17274305555555555, + "grad_norm": 0.874566376209259, + "learning_rate": 9.917354632245712e-06, + "loss": 0.0624, + "step": 398 + }, + { + "epoch": 0.17317708333333334, + "grad_norm": 0.4679570198059082, + "learning_rate": 9.91671699306763e-06, + "loss": 0.035, + "step": 399 + }, + { + "epoch": 0.1736111111111111, + "grad_norm": 0.9329303503036499, + "learning_rate": 9.916076924180734e-06, + "loss": 0.0934, + "step": 400 + }, + { + "epoch": 0.1740451388888889, + "grad_norm": 0.6457216143608093, + "learning_rate": 9.915434425901328e-06, + "loss": 0.0524, + "step": 401 + }, + { + "epoch": 0.17447916666666666, + "grad_norm": 0.5407356023788452, + "learning_rate": 9.91478949854692e-06, + "loss": 0.053, + "step": 402 + }, + { + "epoch": 0.17491319444444445, + "grad_norm": 0.6888991594314575, + "learning_rate": 9.914142142436212e-06, + "loss": 0.0707, + "step": 403 + }, + { + "epoch": 0.1753472222222222, + "grad_norm": 0.6760328412055969, + "learning_rate": 9.913492357889112e-06, + "loss": 0.0853, + "step": 404 + }, + { + "epoch": 0.17578125, + "grad_norm": 0.6837098598480225, + "learning_rate": 9.912840145226725e-06, + "loss": 0.0515, + "step": 405 + }, + { + "epoch": 0.1762152777777778, + "grad_norm": 0.7048181295394897, + "learning_rate": 9.912185504771359e-06, + "loss": 0.0474, + "step": 406 + }, + { + "epoch": 0.17664930555555555, + "grad_norm": 1.214665412902832, + "learning_rate": 9.911528436846519e-06, + "loss": 0.0874, + "step": 407 + }, + { + "epoch": 0.17708333333333334, + "grad_norm": 0.8326371908187866, + "learning_rate": 9.910868941776906e-06, + "loss": 0.0717, + "step": 408 + }, + { + "epoch": 0.1775173611111111, + "grad_norm": 0.818688690662384, + "learning_rate": 9.91020701988843e-06, + "loss": 0.0906, + "step": 409 + }, + { + "epoch": 0.1779513888888889, + "grad_norm": 1.0963165760040283, + "learning_rate": 9.909542671508192e-06, + "loss": 0.0805, + "step": 410 + }, + { + "epoch": 0.17838541666666666, + "grad_norm": 0.6087559461593628, + "learning_rate": 9.908875896964497e-06, + "loss": 0.0529, + "step": 411 + }, + { + "epoch": 0.17881944444444445, + "grad_norm": 0.5906699895858765, + "learning_rate": 9.908206696586846e-06, + "loss": 0.0787, + "step": 412 + }, + { + "epoch": 0.1792534722222222, + "grad_norm": 0.5716590285301208, + "learning_rate": 9.90753507070594e-06, + "loss": 0.0671, + "step": 413 + }, + { + "epoch": 0.1796875, + "grad_norm": 0.7733680605888367, + "learning_rate": 9.906861019653677e-06, + "loss": 0.0495, + "step": 414 + }, + { + "epoch": 0.1801215277777778, + "grad_norm": 0.8408774733543396, + "learning_rate": 9.906184543763158e-06, + "loss": 0.0848, + "step": 415 + }, + { + "epoch": 0.18055555555555555, + "grad_norm": 1.0400339365005493, + "learning_rate": 9.905505643368675e-06, + "loss": 0.0825, + "step": 416 + }, + { + "epoch": 0.18098958333333334, + "grad_norm": 0.5606048107147217, + "learning_rate": 9.904824318805725e-06, + "loss": 0.0548, + "step": 417 + }, + { + "epoch": 0.1814236111111111, + "grad_norm": 0.7534106373786926, + "learning_rate": 9.904140570410997e-06, + "loss": 0.0719, + "step": 418 + }, + { + "epoch": 0.1818576388888889, + "grad_norm": 0.6680476069450378, + "learning_rate": 9.903454398522388e-06, + "loss": 0.0634, + "step": 419 + }, + { + "epoch": 0.18229166666666666, + "grad_norm": 0.5825947523117065, + "learning_rate": 9.90276580347898e-06, + "loss": 0.0447, + "step": 420 + }, + { + "epoch": 0.18272569444444445, + "grad_norm": 0.7220215201377869, + "learning_rate": 9.902074785621058e-06, + "loss": 0.0438, + "step": 421 + }, + { + "epoch": 0.1831597222222222, + "grad_norm": 1.1537821292877197, + "learning_rate": 9.901381345290107e-06, + "loss": 0.0707, + "step": 422 + }, + { + "epoch": 0.18359375, + "grad_norm": 0.9954478740692139, + "learning_rate": 9.900685482828803e-06, + "loss": 0.0788, + "step": 423 + }, + { + "epoch": 0.1840277777777778, + "grad_norm": 1.56073796749115, + "learning_rate": 9.899987198581028e-06, + "loss": 0.0803, + "step": 424 + }, + { + "epoch": 0.18446180555555555, + "grad_norm": 0.9214962720870972, + "learning_rate": 9.899286492891849e-06, + "loss": 0.0592, + "step": 425 + }, + { + "epoch": 0.18489583333333334, + "grad_norm": 0.5767590403556824, + "learning_rate": 9.898583366107539e-06, + "loss": 0.0487, + "step": 426 + }, + { + "epoch": 0.1853298611111111, + "grad_norm": 0.8960171937942505, + "learning_rate": 9.897877818575563e-06, + "loss": 0.0545, + "step": 427 + }, + { + "epoch": 0.1857638888888889, + "grad_norm": 0.9727705121040344, + "learning_rate": 9.897169850644585e-06, + "loss": 0.0686, + "step": 428 + }, + { + "epoch": 0.18619791666666666, + "grad_norm": 1.2093175649642944, + "learning_rate": 9.896459462664465e-06, + "loss": 0.0697, + "step": 429 + }, + { + "epoch": 0.18663194444444445, + "grad_norm": 0.9476456046104431, + "learning_rate": 9.895746654986253e-06, + "loss": 0.0638, + "step": 430 + }, + { + "epoch": 0.1870659722222222, + "grad_norm": 0.5582579970359802, + "learning_rate": 9.895031427962205e-06, + "loss": 0.0528, + "step": 431 + }, + { + "epoch": 0.1875, + "grad_norm": 0.9498739838600159, + "learning_rate": 9.894313781945762e-06, + "loss": 0.0452, + "step": 432 + }, + { + "epoch": 0.1879340277777778, + "grad_norm": 2.0486114025115967, + "learning_rate": 9.893593717291568e-06, + "loss": 0.0941, + "step": 433 + }, + { + "epoch": 0.18836805555555555, + "grad_norm": 1.371767520904541, + "learning_rate": 9.892871234355457e-06, + "loss": 0.0568, + "step": 434 + }, + { + "epoch": 0.18880208333333334, + "grad_norm": 1.096907615661621, + "learning_rate": 9.892146333494463e-06, + "loss": 0.0472, + "step": 435 + }, + { + "epoch": 0.1892361111111111, + "grad_norm": 1.0589361190795898, + "learning_rate": 9.891419015066812e-06, + "loss": 0.0644, + "step": 436 + }, + { + "epoch": 0.1896701388888889, + "grad_norm": 0.9310447573661804, + "learning_rate": 9.890689279431924e-06, + "loss": 0.0709, + "step": 437 + }, + { + "epoch": 0.19010416666666666, + "grad_norm": 0.8208966851234436, + "learning_rate": 9.889957126950416e-06, + "loss": 0.0561, + "step": 438 + }, + { + "epoch": 0.19053819444444445, + "grad_norm": 1.5516587495803833, + "learning_rate": 9.889222557984098e-06, + "loss": 0.0617, + "step": 439 + }, + { + "epoch": 0.1909722222222222, + "grad_norm": 1.1894527673721313, + "learning_rate": 9.888485572895971e-06, + "loss": 0.0771, + "step": 440 + }, + { + "epoch": 0.19140625, + "grad_norm": 1.1664812564849854, + "learning_rate": 9.88774617205024e-06, + "loss": 0.0504, + "step": 441 + }, + { + "epoch": 0.1918402777777778, + "grad_norm": 0.7609825730323792, + "learning_rate": 9.88700435581229e-06, + "loss": 0.0606, + "step": 442 + }, + { + "epoch": 0.19227430555555555, + "grad_norm": 0.94313645362854, + "learning_rate": 9.88626012454871e-06, + "loss": 0.0642, + "step": 443 + }, + { + "epoch": 0.19270833333333334, + "grad_norm": 0.6481614112854004, + "learning_rate": 9.885513478627277e-06, + "loss": 0.0571, + "step": 444 + }, + { + "epoch": 0.1931423611111111, + "grad_norm": 0.8185029029846191, + "learning_rate": 9.884764418416965e-06, + "loss": 0.0529, + "step": 445 + }, + { + "epoch": 0.1935763888888889, + "grad_norm": 0.844165563583374, + "learning_rate": 9.884012944287939e-06, + "loss": 0.0875, + "step": 446 + }, + { + "epoch": 0.19401041666666666, + "grad_norm": 0.7095887660980225, + "learning_rate": 9.883259056611559e-06, + "loss": 0.0592, + "step": 447 + }, + { + "epoch": 0.19444444444444445, + "grad_norm": 0.6745681762695312, + "learning_rate": 9.882502755760371e-06, + "loss": 0.0615, + "step": 448 + }, + { + "epoch": 0.1948784722222222, + "grad_norm": 0.7381845116615295, + "learning_rate": 9.881744042108123e-06, + "loss": 0.0607, + "step": 449 + }, + { + "epoch": 0.1953125, + "grad_norm": 0.8988702893257141, + "learning_rate": 9.880982916029749e-06, + "loss": 0.0835, + "step": 450 + }, + { + "epoch": 0.1957465277777778, + "grad_norm": 0.8817563056945801, + "learning_rate": 9.880219377901377e-06, + "loss": 0.1022, + "step": 451 + }, + { + "epoch": 0.19618055555555555, + "grad_norm": 0.8826650381088257, + "learning_rate": 9.879453428100327e-06, + "loss": 0.0735, + "step": 452 + }, + { + "epoch": 0.19661458333333334, + "grad_norm": 0.6274738907814026, + "learning_rate": 9.87868506700511e-06, + "loss": 0.0611, + "step": 453 + }, + { + "epoch": 0.1970486111111111, + "grad_norm": 0.752539336681366, + "learning_rate": 9.877914294995432e-06, + "loss": 0.0621, + "step": 454 + }, + { + "epoch": 0.1974826388888889, + "grad_norm": 0.9866121411323547, + "learning_rate": 9.877141112452185e-06, + "loss": 0.0611, + "step": 455 + }, + { + "epoch": 0.19791666666666666, + "grad_norm": 0.6701814532279968, + "learning_rate": 9.876365519757457e-06, + "loss": 0.0699, + "step": 456 + }, + { + "epoch": 0.19835069444444445, + "grad_norm": 1.2707716226577759, + "learning_rate": 9.875587517294522e-06, + "loss": 0.0788, + "step": 457 + }, + { + "epoch": 0.1987847222222222, + "grad_norm": 1.0906691551208496, + "learning_rate": 9.87480710544785e-06, + "loss": 0.0987, + "step": 458 + }, + { + "epoch": 0.19921875, + "grad_norm": 1.0719773769378662, + "learning_rate": 9.874024284603099e-06, + "loss": 0.0619, + "step": 459 + }, + { + "epoch": 0.1996527777777778, + "grad_norm": 0.8737601041793823, + "learning_rate": 9.873239055147119e-06, + "loss": 0.0734, + "step": 460 + }, + { + "epoch": 0.20008680555555555, + "grad_norm": 0.6847056150436401, + "learning_rate": 9.872451417467947e-06, + "loss": 0.0614, + "step": 461 + }, + { + "epoch": 0.20052083333333334, + "grad_norm": 0.7287288308143616, + "learning_rate": 9.871661371954814e-06, + "loss": 0.0705, + "step": 462 + }, + { + "epoch": 0.2009548611111111, + "grad_norm": 0.7544885873794556, + "learning_rate": 9.87086891899814e-06, + "loss": 0.0804, + "step": 463 + }, + { + "epoch": 0.2013888888888889, + "grad_norm": 0.6416462063789368, + "learning_rate": 9.870074058989532e-06, + "loss": 0.0696, + "step": 464 + }, + { + "epoch": 0.20182291666666666, + "grad_norm": 0.6606134176254272, + "learning_rate": 9.869276792321786e-06, + "loss": 0.0498, + "step": 465 + }, + { + "epoch": 0.20225694444444445, + "grad_norm": 0.7388238906860352, + "learning_rate": 9.868477119388897e-06, + "loss": 0.0707, + "step": 466 + }, + { + "epoch": 0.2026909722222222, + "grad_norm": 0.8424966335296631, + "learning_rate": 9.867675040586035e-06, + "loss": 0.0956, + "step": 467 + }, + { + "epoch": 0.203125, + "grad_norm": 0.7446620464324951, + "learning_rate": 9.866870556309568e-06, + "loss": 0.0682, + "step": 468 + }, + { + "epoch": 0.2035590277777778, + "grad_norm": 0.6781134009361267, + "learning_rate": 9.86606366695705e-06, + "loss": 0.0595, + "step": 469 + }, + { + "epoch": 0.20399305555555555, + "grad_norm": 0.5591951012611389, + "learning_rate": 9.865254372927224e-06, + "loss": 0.0516, + "step": 470 + }, + { + "epoch": 0.20442708333333334, + "grad_norm": 0.81165611743927, + "learning_rate": 9.864442674620023e-06, + "loss": 0.0763, + "step": 471 + }, + { + "epoch": 0.2048611111111111, + "grad_norm": 0.6503605246543884, + "learning_rate": 9.863628572436565e-06, + "loss": 0.0426, + "step": 472 + }, + { + "epoch": 0.2052951388888889, + "grad_norm": 0.7912452220916748, + "learning_rate": 9.862812066779157e-06, + "loss": 0.0745, + "step": 473 + }, + { + "epoch": 0.20572916666666666, + "grad_norm": 0.590604841709137, + "learning_rate": 9.861993158051295e-06, + "loss": 0.0609, + "step": 474 + }, + { + "epoch": 0.20616319444444445, + "grad_norm": 1.0321898460388184, + "learning_rate": 9.86117184665766e-06, + "loss": 0.0697, + "step": 475 + }, + { + "epoch": 0.2065972222222222, + "grad_norm": 1.022867202758789, + "learning_rate": 9.860348133004123e-06, + "loss": 0.0532, + "step": 476 + }, + { + "epoch": 0.20703125, + "grad_norm": 0.6199460029602051, + "learning_rate": 9.859522017497742e-06, + "loss": 0.0623, + "step": 477 + }, + { + "epoch": 0.2074652777777778, + "grad_norm": 1.0487356185913086, + "learning_rate": 9.858693500546758e-06, + "loss": 0.0826, + "step": 478 + }, + { + "epoch": 0.20789930555555555, + "grad_norm": 0.42486223578453064, + "learning_rate": 9.857862582560606e-06, + "loss": 0.0333, + "step": 479 + }, + { + "epoch": 0.20833333333333334, + "grad_norm": 1.3734253644943237, + "learning_rate": 9.857029263949898e-06, + "loss": 0.1035, + "step": 480 + }, + { + "epoch": 0.2087673611111111, + "grad_norm": 0.8876743912696838, + "learning_rate": 9.856193545126444e-06, + "loss": 0.0786, + "step": 481 + }, + { + "epoch": 0.2092013888888889, + "grad_norm": 0.6525356769561768, + "learning_rate": 9.855355426503228e-06, + "loss": 0.0602, + "step": 482 + }, + { + "epoch": 0.20963541666666666, + "grad_norm": 1.0160382986068726, + "learning_rate": 9.854514908494427e-06, + "loss": 0.0816, + "step": 483 + }, + { + "epoch": 0.21006944444444445, + "grad_norm": 1.0310873985290527, + "learning_rate": 9.853671991515405e-06, + "loss": 0.0613, + "step": 484 + }, + { + "epoch": 0.2105034722222222, + "grad_norm": 0.7252808809280396, + "learning_rate": 9.852826675982705e-06, + "loss": 0.0674, + "step": 485 + }, + { + "epoch": 0.2109375, + "grad_norm": 0.7665201425552368, + "learning_rate": 9.851978962314063e-06, + "loss": 0.0509, + "step": 486 + }, + { + "epoch": 0.2113715277777778, + "grad_norm": 0.6184194684028625, + "learning_rate": 9.851128850928394e-06, + "loss": 0.0782, + "step": 487 + }, + { + "epoch": 0.21180555555555555, + "grad_norm": 0.6223265528678894, + "learning_rate": 9.850276342245797e-06, + "loss": 0.0454, + "step": 488 + }, + { + "epoch": 0.21223958333333334, + "grad_norm": 0.7626925110816956, + "learning_rate": 9.849421436687565e-06, + "loss": 0.0725, + "step": 489 + }, + { + "epoch": 0.2126736111111111, + "grad_norm": 0.8450047373771667, + "learning_rate": 9.848564134676166e-06, + "loss": 0.0682, + "step": 490 + }, + { + "epoch": 0.2131076388888889, + "grad_norm": 0.7744449377059937, + "learning_rate": 9.847704436635255e-06, + "loss": 0.0645, + "step": 491 + }, + { + "epoch": 0.21354166666666666, + "grad_norm": 1.0327328443527222, + "learning_rate": 9.846842342989675e-06, + "loss": 0.0837, + "step": 492 + }, + { + "epoch": 0.21397569444444445, + "grad_norm": 0.8731493949890137, + "learning_rate": 9.845977854165443e-06, + "loss": 0.0634, + "step": 493 + }, + { + "epoch": 0.2144097222222222, + "grad_norm": 0.5188009738922119, + "learning_rate": 9.845110970589773e-06, + "loss": 0.0391, + "step": 494 + }, + { + "epoch": 0.21484375, + "grad_norm": 0.7408533692359924, + "learning_rate": 9.84424169269105e-06, + "loss": 0.0556, + "step": 495 + }, + { + "epoch": 0.2152777777777778, + "grad_norm": 0.6349946856498718, + "learning_rate": 9.843370020898852e-06, + "loss": 0.0576, + "step": 496 + }, + { + "epoch": 0.21571180555555555, + "grad_norm": 0.8657709956169128, + "learning_rate": 9.842495955643932e-06, + "loss": 0.0552, + "step": 497 + }, + { + "epoch": 0.21614583333333334, + "grad_norm": 0.7064958214759827, + "learning_rate": 9.841619497358232e-06, + "loss": 0.0662, + "step": 498 + }, + { + "epoch": 0.2165798611111111, + "grad_norm": 0.9118136763572693, + "learning_rate": 9.840740646474873e-06, + "loss": 0.054, + "step": 499 + }, + { + "epoch": 0.2170138888888889, + "grad_norm": 0.577598512172699, + "learning_rate": 9.839859403428159e-06, + "loss": 0.0557, + "step": 500 + }, + { + "epoch": 0.21744791666666666, + "grad_norm": 0.7629770040512085, + "learning_rate": 9.838975768653579e-06, + "loss": 0.0657, + "step": 501 + }, + { + "epoch": 0.21788194444444445, + "grad_norm": 0.7026235461235046, + "learning_rate": 9.838089742587796e-06, + "loss": 0.0614, + "step": 502 + }, + { + "epoch": 0.2183159722222222, + "grad_norm": 0.8833339214324951, + "learning_rate": 9.837201325668667e-06, + "loss": 0.0532, + "step": 503 + }, + { + "epoch": 0.21875, + "grad_norm": 0.8416234254837036, + "learning_rate": 9.836310518335219e-06, + "loss": 0.0568, + "step": 504 + }, + { + "epoch": 0.2191840277777778, + "grad_norm": 1.0332963466644287, + "learning_rate": 9.835417321027664e-06, + "loss": 0.0687, + "step": 505 + }, + { + "epoch": 0.21961805555555555, + "grad_norm": 0.796916127204895, + "learning_rate": 9.8345217341874e-06, + "loss": 0.0631, + "step": 506 + }, + { + "epoch": 0.22005208333333334, + "grad_norm": 0.6922689080238342, + "learning_rate": 9.833623758257003e-06, + "loss": 0.0611, + "step": 507 + }, + { + "epoch": 0.2204861111111111, + "grad_norm": 0.9117850661277771, + "learning_rate": 9.832723393680222e-06, + "loss": 0.0769, + "step": 508 + }, + { + "epoch": 0.2209201388888889, + "grad_norm": 1.138522982597351, + "learning_rate": 9.831820640901997e-06, + "loss": 0.0677, + "step": 509 + }, + { + "epoch": 0.22135416666666666, + "grad_norm": 0.6167424917221069, + "learning_rate": 9.830915500368444e-06, + "loss": 0.0613, + "step": 510 + }, + { + "epoch": 0.22178819444444445, + "grad_norm": 0.8744767904281616, + "learning_rate": 9.830007972526858e-06, + "loss": 0.0858, + "step": 511 + }, + { + "epoch": 0.2222222222222222, + "grad_norm": 0.7937278151512146, + "learning_rate": 9.829098057825716e-06, + "loss": 0.0742, + "step": 512 + }, + { + "epoch": 0.22265625, + "grad_norm": 0.8917018175125122, + "learning_rate": 9.828185756714672e-06, + "loss": 0.0886, + "step": 513 + }, + { + "epoch": 0.2230902777777778, + "grad_norm": 0.6010713577270508, + "learning_rate": 9.827271069644562e-06, + "loss": 0.046, + "step": 514 + }, + { + "epoch": 0.22352430555555555, + "grad_norm": 0.7526761293411255, + "learning_rate": 9.826353997067398e-06, + "loss": 0.0556, + "step": 515 + }, + { + "epoch": 0.22395833333333334, + "grad_norm": 0.6621930599212646, + "learning_rate": 9.825434539436374e-06, + "loss": 0.0491, + "step": 516 + }, + { + "epoch": 0.2243923611111111, + "grad_norm": 1.0901848077774048, + "learning_rate": 9.82451269720586e-06, + "loss": 0.0706, + "step": 517 + }, + { + "epoch": 0.2248263888888889, + "grad_norm": 1.0902998447418213, + "learning_rate": 9.823588470831405e-06, + "loss": 0.0616, + "step": 518 + }, + { + "epoch": 0.22526041666666666, + "grad_norm": 1.2565516233444214, + "learning_rate": 9.822661860769738e-06, + "loss": 0.0952, + "step": 519 + }, + { + "epoch": 0.22569444444444445, + "grad_norm": 1.2477786540985107, + "learning_rate": 9.821732867478764e-06, + "loss": 0.0915, + "step": 520 + }, + { + "epoch": 0.2261284722222222, + "grad_norm": 1.1893221139907837, + "learning_rate": 9.820801491417569e-06, + "loss": 0.0878, + "step": 521 + }, + { + "epoch": 0.2265625, + "grad_norm": 0.8113741874694824, + "learning_rate": 9.81986773304641e-06, + "loss": 0.0625, + "step": 522 + }, + { + "epoch": 0.2269965277777778, + "grad_norm": 0.84547358751297, + "learning_rate": 9.818931592826727e-06, + "loss": 0.0617, + "step": 523 + }, + { + "epoch": 0.22743055555555555, + "grad_norm": 0.6920998692512512, + "learning_rate": 9.817993071221135e-06, + "loss": 0.0717, + "step": 524 + }, + { + "epoch": 0.22786458333333334, + "grad_norm": 0.9595234990119934, + "learning_rate": 9.817052168693427e-06, + "loss": 0.0931, + "step": 525 + }, + { + "epoch": 0.2282986111111111, + "grad_norm": 0.5811662673950195, + "learning_rate": 9.81610888570857e-06, + "loss": 0.0681, + "step": 526 + }, + { + "epoch": 0.2287326388888889, + "grad_norm": 0.8212871551513672, + "learning_rate": 9.81516322273271e-06, + "loss": 0.08, + "step": 527 + }, + { + "epoch": 0.22916666666666666, + "grad_norm": 0.5696889758110046, + "learning_rate": 9.814215180233167e-06, + "loss": 0.0671, + "step": 528 + }, + { + "epoch": 0.22960069444444445, + "grad_norm": 0.5984359383583069, + "learning_rate": 9.81326475867844e-06, + "loss": 0.0694, + "step": 529 + }, + { + "epoch": 0.2300347222222222, + "grad_norm": 0.6086364388465881, + "learning_rate": 9.812311958538202e-06, + "loss": 0.0545, + "step": 530 + }, + { + "epoch": 0.23046875, + "grad_norm": 0.8695059418678284, + "learning_rate": 9.811356780283297e-06, + "loss": 0.085, + "step": 531 + }, + { + "epoch": 0.2309027777777778, + "grad_norm": 0.7423482537269592, + "learning_rate": 9.810399224385752e-06, + "loss": 0.0552, + "step": 532 + }, + { + "epoch": 0.23133680555555555, + "grad_norm": 1.1487404108047485, + "learning_rate": 9.809439291318764e-06, + "loss": 0.09, + "step": 533 + }, + { + "epoch": 0.23177083333333334, + "grad_norm": 0.7295951247215271, + "learning_rate": 9.808476981556706e-06, + "loss": 0.0361, + "step": 534 + }, + { + "epoch": 0.2322048611111111, + "grad_norm": 0.7190055251121521, + "learning_rate": 9.807512295575125e-06, + "loss": 0.0418, + "step": 535 + }, + { + "epoch": 0.2326388888888889, + "grad_norm": 1.2923922538757324, + "learning_rate": 9.806545233850746e-06, + "loss": 0.0802, + "step": 536 + }, + { + "epoch": 0.23307291666666666, + "grad_norm": 1.0366101264953613, + "learning_rate": 9.805575796861461e-06, + "loss": 0.0744, + "step": 537 + }, + { + "epoch": 0.23350694444444445, + "grad_norm": 0.6617826819419861, + "learning_rate": 9.80460398508634e-06, + "loss": 0.0483, + "step": 538 + }, + { + "epoch": 0.2339409722222222, + "grad_norm": 1.1316651105880737, + "learning_rate": 9.80362979900563e-06, + "loss": 0.0825, + "step": 539 + }, + { + "epoch": 0.234375, + "grad_norm": 0.773888349533081, + "learning_rate": 9.802653239100742e-06, + "loss": 0.0719, + "step": 540 + }, + { + "epoch": 0.2348090277777778, + "grad_norm": 0.8033248782157898, + "learning_rate": 9.801674305854268e-06, + "loss": 0.0775, + "step": 541 + }, + { + "epoch": 0.23524305555555555, + "grad_norm": 0.8055210113525391, + "learning_rate": 9.800692999749973e-06, + "loss": 0.0765, + "step": 542 + }, + { + "epoch": 0.23567708333333334, + "grad_norm": 0.6195559501647949, + "learning_rate": 9.799709321272787e-06, + "loss": 0.0603, + "step": 543 + }, + { + "epoch": 0.2361111111111111, + "grad_norm": 0.9171823263168335, + "learning_rate": 9.79872327090882e-06, + "loss": 0.0691, + "step": 544 + }, + { + "epoch": 0.2365451388888889, + "grad_norm": 0.5740596055984497, + "learning_rate": 9.797734849145352e-06, + "loss": 0.054, + "step": 545 + }, + { + "epoch": 0.23697916666666666, + "grad_norm": 0.632985532283783, + "learning_rate": 9.796744056470832e-06, + "loss": 0.0605, + "step": 546 + }, + { + "epoch": 0.23741319444444445, + "grad_norm": 0.6397607326507568, + "learning_rate": 9.795750893374887e-06, + "loss": 0.0451, + "step": 547 + }, + { + "epoch": 0.2378472222222222, + "grad_norm": 0.7479575276374817, + "learning_rate": 9.794755360348309e-06, + "loss": 0.0541, + "step": 548 + }, + { + "epoch": 0.23828125, + "grad_norm": 1.0319230556488037, + "learning_rate": 9.793757457883062e-06, + "loss": 0.0708, + "step": 549 + }, + { + "epoch": 0.2387152777777778, + "grad_norm": 1.3229392766952515, + "learning_rate": 9.792757186472285e-06, + "loss": 0.0931, + "step": 550 + }, + { + "epoch": 0.23914930555555555, + "grad_norm": 1.2721326351165771, + "learning_rate": 9.791754546610284e-06, + "loss": 0.0785, + "step": 551 + }, + { + "epoch": 0.23958333333333334, + "grad_norm": 0.6858370304107666, + "learning_rate": 9.790749538792537e-06, + "loss": 0.0555, + "step": 552 + }, + { + "epoch": 0.2400173611111111, + "grad_norm": 0.7393696904182434, + "learning_rate": 9.789742163515689e-06, + "loss": 0.0775, + "step": 553 + }, + { + "epoch": 0.2404513888888889, + "grad_norm": 0.5606722831726074, + "learning_rate": 9.788732421277563e-06, + "loss": 0.0438, + "step": 554 + }, + { + "epoch": 0.24088541666666666, + "grad_norm": 0.6640605330467224, + "learning_rate": 9.787720312577142e-06, + "loss": 0.0712, + "step": 555 + }, + { + "epoch": 0.24131944444444445, + "grad_norm": 0.8472843170166016, + "learning_rate": 9.786705837914584e-06, + "loss": 0.0681, + "step": 556 + }, + { + "epoch": 0.2417534722222222, + "grad_norm": 0.5820501446723938, + "learning_rate": 9.785688997791218e-06, + "loss": 0.045, + "step": 557 + }, + { + "epoch": 0.2421875, + "grad_norm": 0.7918657064437866, + "learning_rate": 9.784669792709534e-06, + "loss": 0.08, + "step": 558 + }, + { + "epoch": 0.2426215277777778, + "grad_norm": 0.6729964017868042, + "learning_rate": 9.783648223173197e-06, + "loss": 0.058, + "step": 559 + }, + { + "epoch": 0.24305555555555555, + "grad_norm": 0.6780051589012146, + "learning_rate": 9.78262428968704e-06, + "loss": 0.0592, + "step": 560 + }, + { + "epoch": 0.24348958333333334, + "grad_norm": 0.7747189402580261, + "learning_rate": 9.781597992757066e-06, + "loss": 0.0537, + "step": 561 + }, + { + "epoch": 0.2439236111111111, + "grad_norm": 1.1778326034545898, + "learning_rate": 9.78056933289044e-06, + "loss": 0.0714, + "step": 562 + }, + { + "epoch": 0.2443576388888889, + "grad_norm": 0.8010048270225525, + "learning_rate": 9.779538310595495e-06, + "loss": 0.0547, + "step": 563 + }, + { + "epoch": 0.24479166666666666, + "grad_norm": 1.0312997102737427, + "learning_rate": 9.77850492638174e-06, + "loss": 0.0834, + "step": 564 + }, + { + "epoch": 0.24522569444444445, + "grad_norm": 1.1206309795379639, + "learning_rate": 9.777469180759844e-06, + "loss": 0.0829, + "step": 565 + }, + { + "epoch": 0.2456597222222222, + "grad_norm": 0.7379752993583679, + "learning_rate": 9.776431074241643e-06, + "loss": 0.0557, + "step": 566 + }, + { + "epoch": 0.24609375, + "grad_norm": 0.6607487797737122, + "learning_rate": 9.775390607340142e-06, + "loss": 0.0535, + "step": 567 + }, + { + "epoch": 0.2465277777777778, + "grad_norm": 0.7935291528701782, + "learning_rate": 9.774347780569513e-06, + "loss": 0.0608, + "step": 568 + }, + { + "epoch": 0.24696180555555555, + "grad_norm": 0.7309617400169373, + "learning_rate": 9.77330259444509e-06, + "loss": 0.0704, + "step": 569 + }, + { + "epoch": 0.24739583333333334, + "grad_norm": 0.9455975294113159, + "learning_rate": 9.772255049483378e-06, + "loss": 0.0779, + "step": 570 + }, + { + "epoch": 0.2478298611111111, + "grad_norm": 0.716031551361084, + "learning_rate": 9.771205146202045e-06, + "loss": 0.0576, + "step": 571 + }, + { + "epoch": 0.2482638888888889, + "grad_norm": 0.8444411754608154, + "learning_rate": 9.770152885119925e-06, + "loss": 0.0709, + "step": 572 + }, + { + "epoch": 0.24869791666666666, + "grad_norm": 0.5682095289230347, + "learning_rate": 9.769098266757015e-06, + "loss": 0.0494, + "step": 573 + }, + { + "epoch": 0.24913194444444445, + "grad_norm": 0.7053081393241882, + "learning_rate": 9.768041291634483e-06, + "loss": 0.0473, + "step": 574 + }, + { + "epoch": 0.2495659722222222, + "grad_norm": 0.7969498038291931, + "learning_rate": 9.766981960274653e-06, + "loss": 0.061, + "step": 575 + }, + { + "epoch": 0.25, + "grad_norm": 0.7599273324012756, + "learning_rate": 9.765920273201021e-06, + "loss": 0.0515, + "step": 576 + }, + { + "epoch": 0.2504340277777778, + "grad_norm": 0.46713560819625854, + "learning_rate": 9.764856230938244e-06, + "loss": 0.0313, + "step": 577 + }, + { + "epoch": 0.2508680555555556, + "grad_norm": 1.0053895711898804, + "learning_rate": 9.763789834012141e-06, + "loss": 0.0377, + "step": 578 + }, + { + "epoch": 0.2513020833333333, + "grad_norm": 1.5747969150543213, + "learning_rate": 9.7627210829497e-06, + "loss": 0.0576, + "step": 579 + }, + { + "epoch": 0.2517361111111111, + "grad_norm": 1.6479192972183228, + "learning_rate": 9.761649978279065e-06, + "loss": 0.0642, + "step": 580 + }, + { + "epoch": 0.2521701388888889, + "grad_norm": 1.494860053062439, + "learning_rate": 9.760576520529548e-06, + "loss": 0.0698, + "step": 581 + }, + { + "epoch": 0.2526041666666667, + "grad_norm": 1.0137051343917847, + "learning_rate": 9.759500710231624e-06, + "loss": 0.0597, + "step": 582 + }, + { + "epoch": 0.2530381944444444, + "grad_norm": 1.1509759426116943, + "learning_rate": 9.758422547916931e-06, + "loss": 0.0637, + "step": 583 + }, + { + "epoch": 0.2534722222222222, + "grad_norm": 1.0559190511703491, + "learning_rate": 9.757342034118262e-06, + "loss": 0.0696, + "step": 584 + }, + { + "epoch": 0.25390625, + "grad_norm": 0.6210541129112244, + "learning_rate": 9.756259169369584e-06, + "loss": 0.0615, + "step": 585 + }, + { + "epoch": 0.2543402777777778, + "grad_norm": 0.6341214776039124, + "learning_rate": 9.755173954206016e-06, + "loss": 0.0633, + "step": 586 + }, + { + "epoch": 0.2547743055555556, + "grad_norm": 0.6376259326934814, + "learning_rate": 9.754086389163843e-06, + "loss": 0.0604, + "step": 587 + }, + { + "epoch": 0.2552083333333333, + "grad_norm": 1.0872777700424194, + "learning_rate": 9.752996474780512e-06, + "loss": 0.0542, + "step": 588 + }, + { + "epoch": 0.2556423611111111, + "grad_norm": 0.6137789487838745, + "learning_rate": 9.751904211594627e-06, + "loss": 0.0724, + "step": 589 + }, + { + "epoch": 0.2560763888888889, + "grad_norm": 0.6856177449226379, + "learning_rate": 9.750809600145955e-06, + "loss": 0.0747, + "step": 590 + }, + { + "epoch": 0.2565104166666667, + "grad_norm": 0.47774896025657654, + "learning_rate": 9.749712640975424e-06, + "loss": 0.0477, + "step": 591 + }, + { + "epoch": 0.2569444444444444, + "grad_norm": 1.0445431470870972, + "learning_rate": 9.748613334625122e-06, + "loss": 0.0501, + "step": 592 + }, + { + "epoch": 0.2573784722222222, + "grad_norm": 1.031905174255371, + "learning_rate": 9.747511681638295e-06, + "loss": 0.086, + "step": 593 + }, + { + "epoch": 0.2578125, + "grad_norm": 0.8072367906570435, + "learning_rate": 9.746407682559352e-06, + "loss": 0.069, + "step": 594 + }, + { + "epoch": 0.2582465277777778, + "grad_norm": 0.7235659956932068, + "learning_rate": 9.745301337933856e-06, + "loss": 0.0611, + "step": 595 + }, + { + "epoch": 0.2586805555555556, + "grad_norm": 2.203366994857788, + "learning_rate": 9.744192648308539e-06, + "loss": 0.0646, + "step": 596 + }, + { + "epoch": 0.2591145833333333, + "grad_norm": 0.9374741315841675, + "learning_rate": 9.743081614231283e-06, + "loss": 0.0763, + "step": 597 + }, + { + "epoch": 0.2595486111111111, + "grad_norm": 1.1199864149093628, + "learning_rate": 9.741968236251126e-06, + "loss": 0.0582, + "step": 598 + }, + { + "epoch": 0.2599826388888889, + "grad_norm": 0.6270304918289185, + "learning_rate": 9.740852514918275e-06, + "loss": 0.0439, + "step": 599 + }, + { + "epoch": 0.2604166666666667, + "grad_norm": 2.2203094959259033, + "learning_rate": 9.739734450784088e-06, + "loss": 0.1011, + "step": 600 + }, + { + "epoch": 0.2604166666666667, + "eval_layout_rm_rank_test_accuracy": 0.9295603101782307, + "eval_layout_rm_rank_test_chosen_scores": 0.7769265174865723, + "eval_layout_rm_rank_test_loss": 0.06331586837768555, + "eval_layout_rm_rank_test_rejected_scores": 0.17147307097911835, + "eval_layout_rm_rank_test_runtime": 1181.1502, + "eval_layout_rm_rank_test_samples_per_second": 20.853, + "eval_layout_rm_rank_test_steps_per_second": 0.652, + "step": 600 + }, + { + "epoch": 0.2608506944444444, + "grad_norm": 2.3830530643463135, + "learning_rate": 9.738614044401083e-06, + "loss": 0.0879, + "step": 601 + }, + { + "epoch": 0.2612847222222222, + "grad_norm": 0.488438218832016, + "learning_rate": 9.737491296322931e-06, + "loss": 0.0515, + "step": 602 + }, + { + "epoch": 0.26171875, + "grad_norm": 0.635438859462738, + "learning_rate": 9.73636620710447e-06, + "loss": 0.0663, + "step": 603 + }, + { + "epoch": 0.2621527777777778, + "grad_norm": 0.7975000739097595, + "learning_rate": 9.735238777301682e-06, + "loss": 0.0548, + "step": 604 + }, + { + "epoch": 0.2625868055555556, + "grad_norm": 0.7051642537117004, + "learning_rate": 9.734109007471715e-06, + "loss": 0.0533, + "step": 605 + }, + { + "epoch": 0.2630208333333333, + "grad_norm": 1.0443001985549927, + "learning_rate": 9.732976898172872e-06, + "loss": 0.0629, + "step": 606 + }, + { + "epoch": 0.2634548611111111, + "grad_norm": 1.143349051475525, + "learning_rate": 9.731842449964609e-06, + "loss": 0.0601, + "step": 607 + }, + { + "epoch": 0.2638888888888889, + "grad_norm": 1.0804708003997803, + "learning_rate": 9.730705663407541e-06, + "loss": 0.0595, + "step": 608 + }, + { + "epoch": 0.2643229166666667, + "grad_norm": 0.9708920121192932, + "learning_rate": 9.729566539063438e-06, + "loss": 0.0532, + "step": 609 + }, + { + "epoch": 0.2647569444444444, + "grad_norm": 0.7470878958702087, + "learning_rate": 9.728425077495222e-06, + "loss": 0.0556, + "step": 610 + }, + { + "epoch": 0.2651909722222222, + "grad_norm": 1.107346773147583, + "learning_rate": 9.727281279266973e-06, + "loss": 0.0589, + "step": 611 + }, + { + "epoch": 0.265625, + "grad_norm": 0.8929963707923889, + "learning_rate": 9.726135144943928e-06, + "loss": 0.0675, + "step": 612 + }, + { + "epoch": 0.2660590277777778, + "grad_norm": 0.6872580647468567, + "learning_rate": 9.724986675092473e-06, + "loss": 0.0551, + "step": 613 + }, + { + "epoch": 0.2664930555555556, + "grad_norm": 0.7875890135765076, + "learning_rate": 9.723835870280152e-06, + "loss": 0.0664, + "step": 614 + }, + { + "epoch": 0.2669270833333333, + "grad_norm": 0.7004210948944092, + "learning_rate": 9.722682731075658e-06, + "loss": 0.0559, + "step": 615 + }, + { + "epoch": 0.2673611111111111, + "grad_norm": 1.1991928815841675, + "learning_rate": 9.721527258048849e-06, + "loss": 0.0877, + "step": 616 + }, + { + "epoch": 0.2677951388888889, + "grad_norm": 0.785410463809967, + "learning_rate": 9.72036945177072e-06, + "loss": 0.0748, + "step": 617 + }, + { + "epoch": 0.2682291666666667, + "grad_norm": 1.1013269424438477, + "learning_rate": 9.719209312813434e-06, + "loss": 0.0641, + "step": 618 + }, + { + "epoch": 0.2686631944444444, + "grad_norm": 0.5591453909873962, + "learning_rate": 9.718046841750297e-06, + "loss": 0.0509, + "step": 619 + }, + { + "epoch": 0.2690972222222222, + "grad_norm": 0.45849889516830444, + "learning_rate": 9.716882039155773e-06, + "loss": 0.0555, + "step": 620 + }, + { + "epoch": 0.26953125, + "grad_norm": 0.895866870880127, + "learning_rate": 9.715714905605473e-06, + "loss": 0.0783, + "step": 621 + }, + { + "epoch": 0.2699652777777778, + "grad_norm": 0.8574803471565247, + "learning_rate": 9.714545441676167e-06, + "loss": 0.0629, + "step": 622 + }, + { + "epoch": 0.2703993055555556, + "grad_norm": 0.5782579779624939, + "learning_rate": 9.713373647945768e-06, + "loss": 0.0701, + "step": 623 + }, + { + "epoch": 0.2708333333333333, + "grad_norm": 0.740162193775177, + "learning_rate": 9.712199524993347e-06, + "loss": 0.0684, + "step": 624 + }, + { + "epoch": 0.2712673611111111, + "grad_norm": 0.6386349201202393, + "learning_rate": 9.711023073399124e-06, + "loss": 0.052, + "step": 625 + }, + { + "epoch": 0.2717013888888889, + "grad_norm": 0.6506643295288086, + "learning_rate": 9.70984429374447e-06, + "loss": 0.0659, + "step": 626 + }, + { + "epoch": 0.2721354166666667, + "grad_norm": 1.0170015096664429, + "learning_rate": 9.708663186611907e-06, + "loss": 0.0864, + "step": 627 + }, + { + "epoch": 0.2725694444444444, + "grad_norm": 0.7113478183746338, + "learning_rate": 9.707479752585103e-06, + "loss": 0.0552, + "step": 628 + }, + { + "epoch": 0.2730034722222222, + "grad_norm": 0.9804077744483948, + "learning_rate": 9.70629399224888e-06, + "loss": 0.0902, + "step": 629 + }, + { + "epoch": 0.2734375, + "grad_norm": 0.5103983283042908, + "learning_rate": 9.705105906189212e-06, + "loss": 0.0439, + "step": 630 + }, + { + "epoch": 0.2738715277777778, + "grad_norm": 0.7018703818321228, + "learning_rate": 9.703915494993215e-06, + "loss": 0.0494, + "step": 631 + }, + { + "epoch": 0.2743055555555556, + "grad_norm": 0.6729298233985901, + "learning_rate": 9.702722759249161e-06, + "loss": 0.0637, + "step": 632 + }, + { + "epoch": 0.2747395833333333, + "grad_norm": 0.5301738977432251, + "learning_rate": 9.701527699546465e-06, + "loss": 0.0431, + "step": 633 + }, + { + "epoch": 0.2751736111111111, + "grad_norm": 0.7354344725608826, + "learning_rate": 9.700330316475697e-06, + "loss": 0.0742, + "step": 634 + }, + { + "epoch": 0.2756076388888889, + "grad_norm": 0.8999168276786804, + "learning_rate": 9.69913061062857e-06, + "loss": 0.0602, + "step": 635 + }, + { + "epoch": 0.2760416666666667, + "grad_norm": 0.6171438694000244, + "learning_rate": 9.697928582597942e-06, + "loss": 0.0565, + "step": 636 + }, + { + "epoch": 0.2764756944444444, + "grad_norm": 0.9136935472488403, + "learning_rate": 9.696724232977829e-06, + "loss": 0.0715, + "step": 637 + }, + { + "epoch": 0.2769097222222222, + "grad_norm": 0.827450692653656, + "learning_rate": 9.695517562363385e-06, + "loss": 0.0705, + "step": 638 + }, + { + "epoch": 0.27734375, + "grad_norm": 0.5673283934593201, + "learning_rate": 9.694308571350914e-06, + "loss": 0.0611, + "step": 639 + }, + { + "epoch": 0.2777777777777778, + "grad_norm": 0.8250726461410522, + "learning_rate": 9.693097260537867e-06, + "loss": 0.0567, + "step": 640 + }, + { + "epoch": 0.2782118055555556, + "grad_norm": 0.677777886390686, + "learning_rate": 9.691883630522842e-06, + "loss": 0.0698, + "step": 641 + }, + { + "epoch": 0.2786458333333333, + "grad_norm": 0.7001973390579224, + "learning_rate": 9.690667681905582e-06, + "loss": 0.0364, + "step": 642 + }, + { + "epoch": 0.2790798611111111, + "grad_norm": 0.5818181037902832, + "learning_rate": 9.689449415286974e-06, + "loss": 0.0449, + "step": 643 + }, + { + "epoch": 0.2795138888888889, + "grad_norm": 0.7768517732620239, + "learning_rate": 9.688228831269056e-06, + "loss": 0.0649, + "step": 644 + }, + { + "epoch": 0.2799479166666667, + "grad_norm": 0.8156334161758423, + "learning_rate": 9.687005930455002e-06, + "loss": 0.0785, + "step": 645 + }, + { + "epoch": 0.2803819444444444, + "grad_norm": 0.9127948880195618, + "learning_rate": 9.685780713449143e-06, + "loss": 0.0817, + "step": 646 + }, + { + "epoch": 0.2808159722222222, + "grad_norm": 0.822266161441803, + "learning_rate": 9.684553180856945e-06, + "loss": 0.0688, + "step": 647 + }, + { + "epoch": 0.28125, + "grad_norm": 0.5691503882408142, + "learning_rate": 9.683323333285022e-06, + "loss": 0.0477, + "step": 648 + }, + { + "epoch": 0.2816840277777778, + "grad_norm": 0.7925100922584534, + "learning_rate": 9.68209117134113e-06, + "loss": 0.0538, + "step": 649 + }, + { + "epoch": 0.2821180555555556, + "grad_norm": 0.7010947465896606, + "learning_rate": 9.680856695634174e-06, + "loss": 0.0671, + "step": 650 + }, + { + "epoch": 0.2825520833333333, + "grad_norm": 0.9301623702049255, + "learning_rate": 9.679619906774193e-06, + "loss": 0.0634, + "step": 651 + }, + { + "epoch": 0.2829861111111111, + "grad_norm": 0.7186834216117859, + "learning_rate": 9.678380805372381e-06, + "loss": 0.0542, + "step": 652 + }, + { + "epoch": 0.2834201388888889, + "grad_norm": 0.8715153336524963, + "learning_rate": 9.677139392041062e-06, + "loss": 0.0525, + "step": 653 + }, + { + "epoch": 0.2838541666666667, + "grad_norm": 1.021162748336792, + "learning_rate": 9.675895667393713e-06, + "loss": 0.0723, + "step": 654 + }, + { + "epoch": 0.2842881944444444, + "grad_norm": 0.7757608890533447, + "learning_rate": 9.67464963204495e-06, + "loss": 0.044, + "step": 655 + }, + { + "epoch": 0.2847222222222222, + "grad_norm": 1.0429720878601074, + "learning_rate": 9.673401286610527e-06, + "loss": 0.0768, + "step": 656 + }, + { + "epoch": 0.28515625, + "grad_norm": 0.6734326481819153, + "learning_rate": 9.672150631707343e-06, + "loss": 0.0602, + "step": 657 + }, + { + "epoch": 0.2855902777777778, + "grad_norm": 0.598980724811554, + "learning_rate": 9.67089766795344e-06, + "loss": 0.0576, + "step": 658 + }, + { + "epoch": 0.2860243055555556, + "grad_norm": 0.7078015208244324, + "learning_rate": 9.669642395967997e-06, + "loss": 0.0677, + "step": 659 + }, + { + "epoch": 0.2864583333333333, + "grad_norm": 0.7563789486885071, + "learning_rate": 9.668384816371337e-06, + "loss": 0.0676, + "step": 660 + }, + { + "epoch": 0.2868923611111111, + "grad_norm": 0.6212117671966553, + "learning_rate": 9.667124929784922e-06, + "loss": 0.0415, + "step": 661 + }, + { + "epoch": 0.2873263888888889, + "grad_norm": 0.9101629257202148, + "learning_rate": 9.665862736831351e-06, + "loss": 0.0587, + "step": 662 + }, + { + "epoch": 0.2877604166666667, + "grad_norm": 0.7955373525619507, + "learning_rate": 9.664598238134369e-06, + "loss": 0.0563, + "step": 663 + }, + { + "epoch": 0.2881944444444444, + "grad_norm": 0.933698832988739, + "learning_rate": 9.663331434318855e-06, + "loss": 0.0659, + "step": 664 + }, + { + "epoch": 0.2886284722222222, + "grad_norm": 0.6098387241363525, + "learning_rate": 9.662062326010831e-06, + "loss": 0.0519, + "step": 665 + }, + { + "epoch": 0.2890625, + "grad_norm": 0.6711975336074829, + "learning_rate": 9.660790913837455e-06, + "loss": 0.0694, + "step": 666 + }, + { + "epoch": 0.2894965277777778, + "grad_norm": 0.680465817451477, + "learning_rate": 9.659517198427024e-06, + "loss": 0.0552, + "step": 667 + }, + { + "epoch": 0.2899305555555556, + "grad_norm": 0.6448178887367249, + "learning_rate": 9.658241180408974e-06, + "loss": 0.0498, + "step": 668 + }, + { + "epoch": 0.2903645833333333, + "grad_norm": 0.9048629403114319, + "learning_rate": 9.65696286041388e-06, + "loss": 0.0576, + "step": 669 + }, + { + "epoch": 0.2907986111111111, + "grad_norm": 0.7213271856307983, + "learning_rate": 9.655682239073452e-06, + "loss": 0.0548, + "step": 670 + }, + { + "epoch": 0.2912326388888889, + "grad_norm": 1.198144793510437, + "learning_rate": 9.65439931702054e-06, + "loss": 0.0648, + "step": 671 + }, + { + "epoch": 0.2916666666666667, + "grad_norm": 0.9591615796089172, + "learning_rate": 9.653114094889128e-06, + "loss": 0.0649, + "step": 672 + }, + { + "epoch": 0.2921006944444444, + "grad_norm": 0.8612697124481201, + "learning_rate": 9.651826573314338e-06, + "loss": 0.0481, + "step": 673 + }, + { + "epoch": 0.2925347222222222, + "grad_norm": 0.5536429286003113, + "learning_rate": 9.650536752932429e-06, + "loss": 0.0356, + "step": 674 + }, + { + "epoch": 0.29296875, + "grad_norm": 0.7021438479423523, + "learning_rate": 9.649244634380794e-06, + "loss": 0.0543, + "step": 675 + }, + { + "epoch": 0.2934027777777778, + "grad_norm": 0.7661638259887695, + "learning_rate": 9.647950218297966e-06, + "loss": 0.0589, + "step": 676 + }, + { + "epoch": 0.2938368055555556, + "grad_norm": 0.5391556620597839, + "learning_rate": 9.646653505323609e-06, + "loss": 0.041, + "step": 677 + }, + { + "epoch": 0.2942708333333333, + "grad_norm": 0.7588377594947815, + "learning_rate": 9.645354496098522e-06, + "loss": 0.0576, + "step": 678 + }, + { + "epoch": 0.2947048611111111, + "grad_norm": 1.0408495664596558, + "learning_rate": 9.644053191264643e-06, + "loss": 0.0923, + "step": 679 + }, + { + "epoch": 0.2951388888888889, + "grad_norm": 0.9011924266815186, + "learning_rate": 9.64274959146504e-06, + "loss": 0.0629, + "step": 680 + }, + { + "epoch": 0.2955729166666667, + "grad_norm": 1.4155269861221313, + "learning_rate": 9.641443697343918e-06, + "loss": 0.0723, + "step": 681 + }, + { + "epoch": 0.2960069444444444, + "grad_norm": 0.5065480470657349, + "learning_rate": 9.640135509546613e-06, + "loss": 0.0445, + "step": 682 + }, + { + "epoch": 0.2964409722222222, + "grad_norm": 0.5630937218666077, + "learning_rate": 9.638825028719596e-06, + "loss": 0.0435, + "step": 683 + }, + { + "epoch": 0.296875, + "grad_norm": 0.796134889125824, + "learning_rate": 9.637512255510475e-06, + "loss": 0.0415, + "step": 684 + }, + { + "epoch": 0.2973090277777778, + "grad_norm": 0.8620226383209229, + "learning_rate": 9.636197190567983e-06, + "loss": 0.087, + "step": 685 + }, + { + "epoch": 0.2977430555555556, + "grad_norm": 0.8691263198852539, + "learning_rate": 9.634879834541989e-06, + "loss": 0.0651, + "step": 686 + }, + { + "epoch": 0.2981770833333333, + "grad_norm": 0.4917859733104706, + "learning_rate": 9.633560188083498e-06, + "loss": 0.0367, + "step": 687 + }, + { + "epoch": 0.2986111111111111, + "grad_norm": 0.7439955472946167, + "learning_rate": 9.632238251844643e-06, + "loss": 0.0367, + "step": 688 + }, + { + "epoch": 0.2990451388888889, + "grad_norm": 1.3402535915374756, + "learning_rate": 9.630914026478688e-06, + "loss": 0.097, + "step": 689 + }, + { + "epoch": 0.2994791666666667, + "grad_norm": 1.0623676776885986, + "learning_rate": 9.62958751264003e-06, + "loss": 0.0483, + "step": 690 + }, + { + "epoch": 0.2999131944444444, + "grad_norm": 0.9007038474082947, + "learning_rate": 9.628258710984195e-06, + "loss": 0.0773, + "step": 691 + }, + { + "epoch": 0.3003472222222222, + "grad_norm": 0.7735984921455383, + "learning_rate": 9.626927622167844e-06, + "loss": 0.0538, + "step": 692 + }, + { + "epoch": 0.30078125, + "grad_norm": 0.6532759666442871, + "learning_rate": 9.625594246848762e-06, + "loss": 0.068, + "step": 693 + }, + { + "epoch": 0.3012152777777778, + "grad_norm": 1.0619275569915771, + "learning_rate": 9.62425858568587e-06, + "loss": 0.065, + "step": 694 + }, + { + "epoch": 0.3016493055555556, + "grad_norm": 0.8146920800209045, + "learning_rate": 9.622920639339212e-06, + "loss": 0.0624, + "step": 695 + }, + { + "epoch": 0.3020833333333333, + "grad_norm": 0.6527515053749084, + "learning_rate": 9.62158040846997e-06, + "loss": 0.0526, + "step": 696 + }, + { + "epoch": 0.3025173611111111, + "grad_norm": 0.6193270087242126, + "learning_rate": 9.620237893740446e-06, + "loss": 0.0566, + "step": 697 + }, + { + "epoch": 0.3029513888888889, + "grad_norm": 0.8723997473716736, + "learning_rate": 9.618893095814074e-06, + "loss": 0.058, + "step": 698 + }, + { + "epoch": 0.3033854166666667, + "grad_norm": 0.761968195438385, + "learning_rate": 9.61754601535542e-06, + "loss": 0.0479, + "step": 699 + }, + { + "epoch": 0.3038194444444444, + "grad_norm": 1.001143217086792, + "learning_rate": 9.616196653030173e-06, + "loss": 0.0608, + "step": 700 + }, + { + "epoch": 0.3042534722222222, + "grad_norm": 0.9322062730789185, + "learning_rate": 9.614845009505151e-06, + "loss": 0.0626, + "step": 701 + }, + { + "epoch": 0.3046875, + "grad_norm": 1.3316640853881836, + "learning_rate": 9.613491085448302e-06, + "loss": 0.0631, + "step": 702 + }, + { + "epoch": 0.3051215277777778, + "grad_norm": 0.8206167221069336, + "learning_rate": 9.612134881528695e-06, + "loss": 0.0455, + "step": 703 + }, + { + "epoch": 0.3055555555555556, + "grad_norm": 0.9950183629989624, + "learning_rate": 9.610776398416533e-06, + "loss": 0.0833, + "step": 704 + }, + { + "epoch": 0.3059895833333333, + "grad_norm": 0.6813333630561829, + "learning_rate": 9.609415636783138e-06, + "loss": 0.0583, + "step": 705 + }, + { + "epoch": 0.3064236111111111, + "grad_norm": 0.817216694355011, + "learning_rate": 9.608052597300965e-06, + "loss": 0.0722, + "step": 706 + }, + { + "epoch": 0.3068576388888889, + "grad_norm": 0.5392987132072449, + "learning_rate": 9.606687280643591e-06, + "loss": 0.0425, + "step": 707 + }, + { + "epoch": 0.3072916666666667, + "grad_norm": 0.7456854581832886, + "learning_rate": 9.605319687485718e-06, + "loss": 0.0659, + "step": 708 + }, + { + "epoch": 0.3077256944444444, + "grad_norm": 0.5358279347419739, + "learning_rate": 9.60394981850317e-06, + "loss": 0.0631, + "step": 709 + }, + { + "epoch": 0.3081597222222222, + "grad_norm": 0.5878068208694458, + "learning_rate": 9.602577674372907e-06, + "loss": 0.0358, + "step": 710 + }, + { + "epoch": 0.30859375, + "grad_norm": 0.7481245994567871, + "learning_rate": 9.601203255773e-06, + "loss": 0.0706, + "step": 711 + }, + { + "epoch": 0.3090277777777778, + "grad_norm": 0.7426470518112183, + "learning_rate": 9.599826563382649e-06, + "loss": 0.0496, + "step": 712 + }, + { + "epoch": 0.3094618055555556, + "grad_norm": 0.6087659001350403, + "learning_rate": 9.598447597882181e-06, + "loss": 0.0415, + "step": 713 + }, + { + "epoch": 0.3098958333333333, + "grad_norm": 0.8318248987197876, + "learning_rate": 9.597066359953043e-06, + "loss": 0.0335, + "step": 714 + }, + { + "epoch": 0.3103298611111111, + "grad_norm": 1.3557584285736084, + "learning_rate": 9.595682850277806e-06, + "loss": 0.0925, + "step": 715 + }, + { + "epoch": 0.3107638888888889, + "grad_norm": 0.9173680543899536, + "learning_rate": 9.59429706954016e-06, + "loss": 0.0587, + "step": 716 + }, + { + "epoch": 0.3111979166666667, + "grad_norm": 0.8296968340873718, + "learning_rate": 9.592909018424924e-06, + "loss": 0.0489, + "step": 717 + }, + { + "epoch": 0.3116319444444444, + "grad_norm": 0.6870502233505249, + "learning_rate": 9.591518697618033e-06, + "loss": 0.05, + "step": 718 + }, + { + "epoch": 0.3120659722222222, + "grad_norm": 1.0346730947494507, + "learning_rate": 9.59012610780655e-06, + "loss": 0.0654, + "step": 719 + }, + { + "epoch": 0.3125, + "grad_norm": 1.2370779514312744, + "learning_rate": 9.588731249678648e-06, + "loss": 0.0848, + "step": 720 + }, + { + "epoch": 0.3129340277777778, + "grad_norm": 0.6258195638656616, + "learning_rate": 9.587334123923634e-06, + "loss": 0.0511, + "step": 721 + }, + { + "epoch": 0.3133680555555556, + "grad_norm": 0.6205534338951111, + "learning_rate": 9.585934731231931e-06, + "loss": 0.0646, + "step": 722 + }, + { + "epoch": 0.3138020833333333, + "grad_norm": 0.8664010167121887, + "learning_rate": 9.584533072295077e-06, + "loss": 0.0706, + "step": 723 + }, + { + "epoch": 0.3142361111111111, + "grad_norm": 0.5968353152275085, + "learning_rate": 9.583129147805734e-06, + "loss": 0.0627, + "step": 724 + }, + { + "epoch": 0.3146701388888889, + "grad_norm": 0.464321106672287, + "learning_rate": 9.581722958457687e-06, + "loss": 0.0566, + "step": 725 + }, + { + "epoch": 0.3151041666666667, + "grad_norm": 0.5013965964317322, + "learning_rate": 9.580314504945834e-06, + "loss": 0.0379, + "step": 726 + }, + { + "epoch": 0.3155381944444444, + "grad_norm": 0.46232596039772034, + "learning_rate": 9.578903787966195e-06, + "loss": 0.0416, + "step": 727 + }, + { + "epoch": 0.3159722222222222, + "grad_norm": 0.6010047197341919, + "learning_rate": 9.57749080821591e-06, + "loss": 0.0531, + "step": 728 + }, + { + "epoch": 0.31640625, + "grad_norm": 0.5406182408332825, + "learning_rate": 9.576075566393232e-06, + "loss": 0.0513, + "step": 729 + }, + { + "epoch": 0.3168402777777778, + "grad_norm": 0.82621830701828, + "learning_rate": 9.57465806319754e-06, + "loss": 0.061, + "step": 730 + }, + { + "epoch": 0.3172743055555556, + "grad_norm": 1.1684813499450684, + "learning_rate": 9.573238299329323e-06, + "loss": 0.0912, + "step": 731 + }, + { + "epoch": 0.3177083333333333, + "grad_norm": 0.9138361811637878, + "learning_rate": 9.571816275490191e-06, + "loss": 0.0876, + "step": 732 + }, + { + "epoch": 0.3181423611111111, + "grad_norm": 0.9768379926681519, + "learning_rate": 9.570391992382869e-06, + "loss": 0.0627, + "step": 733 + }, + { + "epoch": 0.3185763888888889, + "grad_norm": 0.8773151636123657, + "learning_rate": 9.568965450711199e-06, + "loss": 0.0771, + "step": 734 + }, + { + "epoch": 0.3190104166666667, + "grad_norm": 0.6589311361312866, + "learning_rate": 9.567536651180139e-06, + "loss": 0.0638, + "step": 735 + }, + { + "epoch": 0.3194444444444444, + "grad_norm": 1.1911635398864746, + "learning_rate": 9.566105594495765e-06, + "loss": 0.0747, + "step": 736 + }, + { + "epoch": 0.3198784722222222, + "grad_norm": 0.592398464679718, + "learning_rate": 9.564672281365263e-06, + "loss": 0.0522, + "step": 737 + }, + { + "epoch": 0.3203125, + "grad_norm": 0.5514065623283386, + "learning_rate": 9.563236712496944e-06, + "loss": 0.0714, + "step": 738 + }, + { + "epoch": 0.3207465277777778, + "grad_norm": 0.7221823930740356, + "learning_rate": 9.561798888600222e-06, + "loss": 0.0463, + "step": 739 + }, + { + "epoch": 0.3211805555555556, + "grad_norm": 0.5506830811500549, + "learning_rate": 9.560358810385632e-06, + "loss": 0.068, + "step": 740 + }, + { + "epoch": 0.3216145833333333, + "grad_norm": 0.6145944595336914, + "learning_rate": 9.558916478564824e-06, + "loss": 0.0485, + "step": 741 + }, + { + "epoch": 0.3220486111111111, + "grad_norm": 0.5052325129508972, + "learning_rate": 9.557471893850556e-06, + "loss": 0.0425, + "step": 742 + }, + { + "epoch": 0.3224826388888889, + "grad_norm": 0.7461121678352356, + "learning_rate": 9.556025056956704e-06, + "loss": 0.0755, + "step": 743 + }, + { + "epoch": 0.3229166666666667, + "grad_norm": 0.7425948977470398, + "learning_rate": 9.554575968598257e-06, + "loss": 0.0411, + "step": 744 + }, + { + "epoch": 0.3233506944444444, + "grad_norm": 0.7867175340652466, + "learning_rate": 9.553124629491314e-06, + "loss": 0.0774, + "step": 745 + }, + { + "epoch": 0.3237847222222222, + "grad_norm": 0.5744606852531433, + "learning_rate": 9.551671040353088e-06, + "loss": 0.049, + "step": 746 + }, + { + "epoch": 0.32421875, + "grad_norm": 0.69539475440979, + "learning_rate": 9.550215201901904e-06, + "loss": 0.0381, + "step": 747 + }, + { + "epoch": 0.3246527777777778, + "grad_norm": 0.952163815498352, + "learning_rate": 9.548757114857198e-06, + "loss": 0.0714, + "step": 748 + }, + { + "epoch": 0.3250868055555556, + "grad_norm": 1.1044210195541382, + "learning_rate": 9.547296779939517e-06, + "loss": 0.0658, + "step": 749 + }, + { + "epoch": 0.3255208333333333, + "grad_norm": 1.186783790588379, + "learning_rate": 9.545834197870519e-06, + "loss": 0.0736, + "step": 750 + }, + { + "epoch": 0.3259548611111111, + "grad_norm": 0.6580379605293274, + "learning_rate": 9.54436936937297e-06, + "loss": 0.06, + "step": 751 + }, + { + "epoch": 0.3263888888888889, + "grad_norm": 0.7105837464332581, + "learning_rate": 9.542902295170755e-06, + "loss": 0.0486, + "step": 752 + }, + { + "epoch": 0.3268229166666667, + "grad_norm": 1.1437960863113403, + "learning_rate": 9.541432975988856e-06, + "loss": 0.0747, + "step": 753 + }, + { + "epoch": 0.3272569444444444, + "grad_norm": 0.8399361968040466, + "learning_rate": 9.539961412553375e-06, + "loss": 0.0799, + "step": 754 + }, + { + "epoch": 0.3276909722222222, + "grad_norm": 0.6523122191429138, + "learning_rate": 9.538487605591518e-06, + "loss": 0.0495, + "step": 755 + }, + { + "epoch": 0.328125, + "grad_norm": 0.6750720739364624, + "learning_rate": 9.5370115558316e-06, + "loss": 0.0621, + "step": 756 + }, + { + "epoch": 0.3285590277777778, + "grad_norm": 0.5642672181129456, + "learning_rate": 9.535533264003043e-06, + "loss": 0.0557, + "step": 757 + }, + { + "epoch": 0.3289930555555556, + "grad_norm": 0.47357630729675293, + "learning_rate": 9.534052730836383e-06, + "loss": 0.0506, + "step": 758 + }, + { + "epoch": 0.3294270833333333, + "grad_norm": 0.704122006893158, + "learning_rate": 9.532569957063257e-06, + "loss": 0.0553, + "step": 759 + }, + { + "epoch": 0.3298611111111111, + "grad_norm": 0.5729414820671082, + "learning_rate": 9.531084943416413e-06, + "loss": 0.0509, + "step": 760 + }, + { + "epoch": 0.3302951388888889, + "grad_norm": 0.6760425567626953, + "learning_rate": 9.529597690629702e-06, + "loss": 0.0471, + "step": 761 + }, + { + "epoch": 0.3307291666666667, + "grad_norm": 1.0773210525512695, + "learning_rate": 9.528108199438088e-06, + "loss": 0.0687, + "step": 762 + }, + { + "epoch": 0.3311631944444444, + "grad_norm": 0.6594577431678772, + "learning_rate": 9.526616470577636e-06, + "loss": 0.045, + "step": 763 + }, + { + "epoch": 0.3315972222222222, + "grad_norm": 1.134124517440796, + "learning_rate": 9.525122504785515e-06, + "loss": 0.0778, + "step": 764 + }, + { + "epoch": 0.33203125, + "grad_norm": 0.9284763932228088, + "learning_rate": 9.523626302800009e-06, + "loss": 0.0531, + "step": 765 + }, + { + "epoch": 0.3324652777777778, + "grad_norm": 0.855278491973877, + "learning_rate": 9.522127865360494e-06, + "loss": 0.0503, + "step": 766 + }, + { + "epoch": 0.3328993055555556, + "grad_norm": 1.1203644275665283, + "learning_rate": 9.520627193207462e-06, + "loss": 0.054, + "step": 767 + }, + { + "epoch": 0.3333333333333333, + "grad_norm": 0.8028052449226379, + "learning_rate": 9.5191242870825e-06, + "loss": 0.0824, + "step": 768 + }, + { + "epoch": 0.3337673611111111, + "grad_norm": 0.6946297883987427, + "learning_rate": 9.51761914772831e-06, + "loss": 0.0484, + "step": 769 + }, + { + "epoch": 0.3342013888888889, + "grad_norm": 0.531436562538147, + "learning_rate": 9.516111775888685e-06, + "loss": 0.04, + "step": 770 + }, + { + "epoch": 0.3346354166666667, + "grad_norm": 0.6533716917037964, + "learning_rate": 9.514602172308532e-06, + "loss": 0.0648, + "step": 771 + }, + { + "epoch": 0.3350694444444444, + "grad_norm": 1.1470931768417358, + "learning_rate": 9.513090337733852e-06, + "loss": 0.0774, + "step": 772 + }, + { + "epoch": 0.3355034722222222, + "grad_norm": 0.771140992641449, + "learning_rate": 9.511576272911757e-06, + "loss": 0.0708, + "step": 773 + }, + { + "epoch": 0.3359375, + "grad_norm": 0.6818678975105286, + "learning_rate": 9.510059978590456e-06, + "loss": 0.0732, + "step": 774 + }, + { + "epoch": 0.3363715277777778, + "grad_norm": 0.9621103405952454, + "learning_rate": 9.50854145551926e-06, + "loss": 0.0487, + "step": 775 + }, + { + "epoch": 0.3368055555555556, + "grad_norm": 0.7351205348968506, + "learning_rate": 9.507020704448581e-06, + "loss": 0.0747, + "step": 776 + }, + { + "epoch": 0.3372395833333333, + "grad_norm": 0.738548219203949, + "learning_rate": 9.505497726129935e-06, + "loss": 0.062, + "step": 777 + }, + { + "epoch": 0.3376736111111111, + "grad_norm": 0.5943174958229065, + "learning_rate": 9.503972521315935e-06, + "loss": 0.0591, + "step": 778 + }, + { + "epoch": 0.3381076388888889, + "grad_norm": 0.5805765986442566, + "learning_rate": 9.502445090760299e-06, + "loss": 0.0586, + "step": 779 + }, + { + "epoch": 0.3385416666666667, + "grad_norm": 0.6666689515113831, + "learning_rate": 9.50091543521784e-06, + "loss": 0.0748, + "step": 780 + }, + { + "epoch": 0.3389756944444444, + "grad_norm": 0.5110723972320557, + "learning_rate": 9.49938355544447e-06, + "loss": 0.0527, + "step": 781 + }, + { + "epoch": 0.3394097222222222, + "grad_norm": 1.0024490356445312, + "learning_rate": 9.497849452197207e-06, + "loss": 0.0641, + "step": 782 + }, + { + "epoch": 0.33984375, + "grad_norm": 0.682685911655426, + "learning_rate": 9.496313126234162e-06, + "loss": 0.0726, + "step": 783 + }, + { + "epoch": 0.3402777777777778, + "grad_norm": 0.6020975112915039, + "learning_rate": 9.494774578314545e-06, + "loss": 0.0635, + "step": 784 + }, + { + "epoch": 0.3407118055555556, + "grad_norm": 0.6309037804603577, + "learning_rate": 9.493233809198665e-06, + "loss": 0.0507, + "step": 785 + }, + { + "epoch": 0.3411458333333333, + "grad_norm": 0.7836750745773315, + "learning_rate": 9.491690819647927e-06, + "loss": 0.0488, + "step": 786 + }, + { + "epoch": 0.3415798611111111, + "grad_norm": 1.142842411994934, + "learning_rate": 9.49014561042484e-06, + "loss": 0.0704, + "step": 787 + }, + { + "epoch": 0.3420138888888889, + "grad_norm": 1.004252314567566, + "learning_rate": 9.488598182292998e-06, + "loss": 0.0467, + "step": 788 + }, + { + "epoch": 0.3424479166666667, + "grad_norm": 1.094494104385376, + "learning_rate": 9.487048536017103e-06, + "loss": 0.0831, + "step": 789 + }, + { + "epoch": 0.3428819444444444, + "grad_norm": 1.0647095441818237, + "learning_rate": 9.485496672362948e-06, + "loss": 0.0691, + "step": 790 + }, + { + "epoch": 0.3433159722222222, + "grad_norm": 0.8126431703567505, + "learning_rate": 9.48394259209742e-06, + "loss": 0.0514, + "step": 791 + }, + { + "epoch": 0.34375, + "grad_norm": 0.9838793873786926, + "learning_rate": 9.482386295988504e-06, + "loss": 0.0843, + "step": 792 + }, + { + "epoch": 0.3441840277777778, + "grad_norm": 0.7736327648162842, + "learning_rate": 9.480827784805278e-06, + "loss": 0.0423, + "step": 793 + }, + { + "epoch": 0.3446180555555556, + "grad_norm": 0.7374549508094788, + "learning_rate": 9.479267059317921e-06, + "loss": 0.0551, + "step": 794 + }, + { + "epoch": 0.3450520833333333, + "grad_norm": 0.9624958634376526, + "learning_rate": 9.477704120297698e-06, + "loss": 0.0599, + "step": 795 + }, + { + "epoch": 0.3454861111111111, + "grad_norm": 0.9699764847755432, + "learning_rate": 9.47613896851697e-06, + "loss": 0.0463, + "step": 796 + }, + { + "epoch": 0.3459201388888889, + "grad_norm": 0.7315773963928223, + "learning_rate": 9.474571604749196e-06, + "loss": 0.0585, + "step": 797 + }, + { + "epoch": 0.3463541666666667, + "grad_norm": 1.3439610004425049, + "learning_rate": 9.473002029768923e-06, + "loss": 0.0794, + "step": 798 + }, + { + "epoch": 0.3467881944444444, + "grad_norm": 0.6556320190429688, + "learning_rate": 9.471430244351793e-06, + "loss": 0.0647, + "step": 799 + }, + { + "epoch": 0.3472222222222222, + "grad_norm": 0.7467793822288513, + "learning_rate": 9.469856249274541e-06, + "loss": 0.0612, + "step": 800 + }, + { + "epoch": 0.34765625, + "grad_norm": 0.9917733073234558, + "learning_rate": 9.468280045314991e-06, + "loss": 0.0691, + "step": 801 + }, + { + "epoch": 0.3480902777777778, + "grad_norm": 0.577621579170227, + "learning_rate": 9.466701633252062e-06, + "loss": 0.0518, + "step": 802 + }, + { + "epoch": 0.3485243055555556, + "grad_norm": 0.8279784917831421, + "learning_rate": 9.465121013865761e-06, + "loss": 0.067, + "step": 803 + }, + { + "epoch": 0.3489583333333333, + "grad_norm": 0.5595852732658386, + "learning_rate": 9.46353818793719e-06, + "loss": 0.0478, + "step": 804 + }, + { + "epoch": 0.3493923611111111, + "grad_norm": 0.8331674933433533, + "learning_rate": 9.461953156248534e-06, + "loss": 0.0605, + "step": 805 + }, + { + "epoch": 0.3498263888888889, + "grad_norm": 0.7563780546188354, + "learning_rate": 9.460365919583077e-06, + "loss": 0.0374, + "step": 806 + }, + { + "epoch": 0.3502604166666667, + "grad_norm": 0.6543465852737427, + "learning_rate": 9.45877647872519e-06, + "loss": 0.0458, + "step": 807 + }, + { + "epoch": 0.3506944444444444, + "grad_norm": 0.9966139793395996, + "learning_rate": 9.457184834460327e-06, + "loss": 0.0615, + "step": 808 + }, + { + "epoch": 0.3511284722222222, + "grad_norm": 0.8658338189125061, + "learning_rate": 9.455590987575038e-06, + "loss": 0.0586, + "step": 809 + }, + { + "epoch": 0.3515625, + "grad_norm": 0.5670788884162903, + "learning_rate": 9.453994938856958e-06, + "loss": 0.0378, + "step": 810 + }, + { + "epoch": 0.3519965277777778, + "grad_norm": 0.9414218664169312, + "learning_rate": 9.452396689094815e-06, + "loss": 0.0694, + "step": 811 + }, + { + "epoch": 0.3524305555555556, + "grad_norm": 0.8863024115562439, + "learning_rate": 9.450796239078417e-06, + "loss": 0.0554, + "step": 812 + }, + { + "epoch": 0.3528645833333333, + "grad_norm": 0.7250595092773438, + "learning_rate": 9.449193589598664e-06, + "loss": 0.0565, + "step": 813 + }, + { + "epoch": 0.3532986111111111, + "grad_norm": 0.845879316329956, + "learning_rate": 9.447588741447542e-06, + "loss": 0.0733, + "step": 814 + }, + { + "epoch": 0.3537326388888889, + "grad_norm": 1.1272249221801758, + "learning_rate": 9.445981695418124e-06, + "loss": 0.0833, + "step": 815 + }, + { + "epoch": 0.3541666666666667, + "grad_norm": 0.516075611114502, + "learning_rate": 9.444372452304571e-06, + "loss": 0.0444, + "step": 816 + }, + { + "epoch": 0.3546006944444444, + "grad_norm": 0.5312237739562988, + "learning_rate": 9.442761012902125e-06, + "loss": 0.0417, + "step": 817 + }, + { + "epoch": 0.3550347222222222, + "grad_norm": 0.6717514991760254, + "learning_rate": 9.441147378007113e-06, + "loss": 0.0562, + "step": 818 + }, + { + "epoch": 0.35546875, + "grad_norm": 0.6433478593826294, + "learning_rate": 9.439531548416955e-06, + "loss": 0.0521, + "step": 819 + }, + { + "epoch": 0.3559027777777778, + "grad_norm": 0.7206553220748901, + "learning_rate": 9.437913524930148e-06, + "loss": 0.0727, + "step": 820 + }, + { + "epoch": 0.3563368055555556, + "grad_norm": 1.5485148429870605, + "learning_rate": 9.436293308346275e-06, + "loss": 0.0875, + "step": 821 + }, + { + "epoch": 0.3567708333333333, + "grad_norm": 1.1846742630004883, + "learning_rate": 9.434670899466007e-06, + "loss": 0.0647, + "step": 822 + }, + { + "epoch": 0.3572048611111111, + "grad_norm": 0.6657606959342957, + "learning_rate": 9.433046299091091e-06, + "loss": 0.0464, + "step": 823 + }, + { + "epoch": 0.3576388888888889, + "grad_norm": 0.6496115326881409, + "learning_rate": 9.431419508024359e-06, + "loss": 0.0716, + "step": 824 + }, + { + "epoch": 0.3580729166666667, + "grad_norm": 0.9403788447380066, + "learning_rate": 9.429790527069733e-06, + "loss": 0.0627, + "step": 825 + }, + { + "epoch": 0.3585069444444444, + "grad_norm": 0.5675524473190308, + "learning_rate": 9.428159357032207e-06, + "loss": 0.0527, + "step": 826 + }, + { + "epoch": 0.3589409722222222, + "grad_norm": 0.7261064648628235, + "learning_rate": 9.426525998717861e-06, + "loss": 0.0603, + "step": 827 + }, + { + "epoch": 0.359375, + "grad_norm": 0.8921183943748474, + "learning_rate": 9.42489045293386e-06, + "loss": 0.077, + "step": 828 + }, + { + "epoch": 0.3598090277777778, + "grad_norm": 0.4754103124141693, + "learning_rate": 9.423252720488445e-06, + "loss": 0.042, + "step": 829 + }, + { + "epoch": 0.3602430555555556, + "grad_norm": 0.5494545698165894, + "learning_rate": 9.421612802190936e-06, + "loss": 0.0545, + "step": 830 + }, + { + "epoch": 0.3606770833333333, + "grad_norm": 0.9630483984947205, + "learning_rate": 9.419970698851739e-06, + "loss": 0.0702, + "step": 831 + }, + { + "epoch": 0.3611111111111111, + "grad_norm": 0.7188891172409058, + "learning_rate": 9.418326411282338e-06, + "loss": 0.0555, + "step": 832 + }, + { + "epoch": 0.3615451388888889, + "grad_norm": 0.5888816714286804, + "learning_rate": 9.416679940295296e-06, + "loss": 0.0425, + "step": 833 + }, + { + "epoch": 0.3619791666666667, + "grad_norm": 0.7920419573783875, + "learning_rate": 9.415031286704253e-06, + "loss": 0.0584, + "step": 834 + }, + { + "epoch": 0.3624131944444444, + "grad_norm": 0.6306520700454712, + "learning_rate": 9.413380451323929e-06, + "loss": 0.0616, + "step": 835 + }, + { + "epoch": 0.3628472222222222, + "grad_norm": 0.8524011969566345, + "learning_rate": 9.411727434970121e-06, + "loss": 0.0443, + "step": 836 + }, + { + "epoch": 0.36328125, + "grad_norm": 0.9743005633354187, + "learning_rate": 9.410072238459709e-06, + "loss": 0.0522, + "step": 837 + }, + { + "epoch": 0.3637152777777778, + "grad_norm": 0.791600227355957, + "learning_rate": 9.408414862610645e-06, + "loss": 0.0479, + "step": 838 + }, + { + "epoch": 0.3641493055555556, + "grad_norm": 0.7155187726020813, + "learning_rate": 9.406755308241955e-06, + "loss": 0.0482, + "step": 839 + }, + { + "epoch": 0.3645833333333333, + "grad_norm": 0.7736154794692993, + "learning_rate": 9.405093576173752e-06, + "loss": 0.0328, + "step": 840 + }, + { + "epoch": 0.3650173611111111, + "grad_norm": 0.6123819351196289, + "learning_rate": 9.403429667227216e-06, + "loss": 0.0467, + "step": 841 + }, + { + "epoch": 0.3654513888888889, + "grad_norm": 0.43281838297843933, + "learning_rate": 9.401763582224606e-06, + "loss": 0.0335, + "step": 842 + }, + { + "epoch": 0.3658854166666667, + "grad_norm": 1.0776960849761963, + "learning_rate": 9.400095321989256e-06, + "loss": 0.0619, + "step": 843 + }, + { + "epoch": 0.3663194444444444, + "grad_norm": 1.0118237733840942, + "learning_rate": 9.398424887345577e-06, + "loss": 0.0668, + "step": 844 + }, + { + "epoch": 0.3667534722222222, + "grad_norm": 1.4794554710388184, + "learning_rate": 9.396752279119052e-06, + "loss": 0.0898, + "step": 845 + }, + { + "epoch": 0.3671875, + "grad_norm": 0.6297768354415894, + "learning_rate": 9.395077498136239e-06, + "loss": 0.0445, + "step": 846 + }, + { + "epoch": 0.3676215277777778, + "grad_norm": 0.8366924524307251, + "learning_rate": 9.393400545224766e-06, + "loss": 0.0539, + "step": 847 + }, + { + "epoch": 0.3680555555555556, + "grad_norm": 0.6405556797981262, + "learning_rate": 9.391721421213343e-06, + "loss": 0.0534, + "step": 848 + }, + { + "epoch": 0.3684895833333333, + "grad_norm": 0.711905300617218, + "learning_rate": 9.390040126931746e-06, + "loss": 0.0543, + "step": 849 + }, + { + "epoch": 0.3689236111111111, + "grad_norm": 0.6284782290458679, + "learning_rate": 9.388356663210824e-06, + "loss": 0.0685, + "step": 850 + }, + { + "epoch": 0.3693576388888889, + "grad_norm": 0.8485714793205261, + "learning_rate": 9.3866710308825e-06, + "loss": 0.0701, + "step": 851 + }, + { + "epoch": 0.3697916666666667, + "grad_norm": 0.847075879573822, + "learning_rate": 9.384983230779769e-06, + "loss": 0.0737, + "step": 852 + }, + { + "epoch": 0.3702256944444444, + "grad_norm": 0.5557371377944946, + "learning_rate": 9.383293263736694e-06, + "loss": 0.0513, + "step": 853 + }, + { + "epoch": 0.3706597222222222, + "grad_norm": 0.8466718196868896, + "learning_rate": 9.381601130588414e-06, + "loss": 0.0659, + "step": 854 + }, + { + "epoch": 0.37109375, + "grad_norm": 0.6541239023208618, + "learning_rate": 9.379906832171135e-06, + "loss": 0.0709, + "step": 855 + }, + { + "epoch": 0.3715277777777778, + "grad_norm": 0.604433536529541, + "learning_rate": 9.37821036932213e-06, + "loss": 0.0592, + "step": 856 + }, + { + "epoch": 0.3719618055555556, + "grad_norm": 0.701170027256012, + "learning_rate": 9.376511742879749e-06, + "loss": 0.074, + "step": 857 + }, + { + "epoch": 0.3723958333333333, + "grad_norm": 0.7357682585716248, + "learning_rate": 9.374810953683407e-06, + "loss": 0.0583, + "step": 858 + }, + { + "epoch": 0.3728298611111111, + "grad_norm": 0.7110288143157959, + "learning_rate": 9.373108002573586e-06, + "loss": 0.0659, + "step": 859 + }, + { + "epoch": 0.3732638888888889, + "grad_norm": 0.5729966759681702, + "learning_rate": 9.371402890391839e-06, + "loss": 0.0611, + "step": 860 + }, + { + "epoch": 0.3736979166666667, + "grad_norm": 0.46971800923347473, + "learning_rate": 9.369695617980789e-06, + "loss": 0.0402, + "step": 861 + }, + { + "epoch": 0.3741319444444444, + "grad_norm": 0.7789297103881836, + "learning_rate": 9.36798618618412e-06, + "loss": 0.0589, + "step": 862 + }, + { + "epoch": 0.3745659722222222, + "grad_norm": 0.6817132234573364, + "learning_rate": 9.366274595846592e-06, + "loss": 0.0618, + "step": 863 + }, + { + "epoch": 0.375, + "grad_norm": 0.9710595607757568, + "learning_rate": 9.36456084781402e-06, + "loss": 0.0818, + "step": 864 + }, + { + "epoch": 0.3754340277777778, + "grad_norm": 0.7942180037498474, + "learning_rate": 9.362844942933297e-06, + "loss": 0.0528, + "step": 865 + }, + { + "epoch": 0.3758680555555556, + "grad_norm": 0.5004309415817261, + "learning_rate": 9.361126882052378e-06, + "loss": 0.0429, + "step": 866 + }, + { + "epoch": 0.3763020833333333, + "grad_norm": 0.6871036291122437, + "learning_rate": 9.359406666020279e-06, + "loss": 0.0609, + "step": 867 + }, + { + "epoch": 0.3767361111111111, + "grad_norm": 0.5601425170898438, + "learning_rate": 9.357684295687084e-06, + "loss": 0.0543, + "step": 868 + }, + { + "epoch": 0.3771701388888889, + "grad_norm": 0.5933509469032288, + "learning_rate": 9.355959771903945e-06, + "loss": 0.0577, + "step": 869 + }, + { + "epoch": 0.3776041666666667, + "grad_norm": 0.7128217816352844, + "learning_rate": 9.354233095523071e-06, + "loss": 0.0635, + "step": 870 + }, + { + "epoch": 0.3780381944444444, + "grad_norm": 0.6243704557418823, + "learning_rate": 9.352504267397742e-06, + "loss": 0.0712, + "step": 871 + }, + { + "epoch": 0.3784722222222222, + "grad_norm": 0.5012828707695007, + "learning_rate": 9.350773288382297e-06, + "loss": 0.0405, + "step": 872 + }, + { + "epoch": 0.37890625, + "grad_norm": 0.5739850997924805, + "learning_rate": 9.349040159332141e-06, + "loss": 0.0549, + "step": 873 + }, + { + "epoch": 0.3793402777777778, + "grad_norm": 0.6758148074150085, + "learning_rate": 9.347304881103738e-06, + "loss": 0.0522, + "step": 874 + }, + { + "epoch": 0.3797743055555556, + "grad_norm": 0.5921206474304199, + "learning_rate": 9.345567454554614e-06, + "loss": 0.0523, + "step": 875 + }, + { + "epoch": 0.3802083333333333, + "grad_norm": 0.523183286190033, + "learning_rate": 9.343827880543362e-06, + "loss": 0.0388, + "step": 876 + }, + { + "epoch": 0.3806423611111111, + "grad_norm": 0.65218186378479, + "learning_rate": 9.342086159929629e-06, + "loss": 0.0501, + "step": 877 + }, + { + "epoch": 0.3810763888888889, + "grad_norm": 0.7650300860404968, + "learning_rate": 9.34034229357413e-06, + "loss": 0.0551, + "step": 878 + }, + { + "epoch": 0.3815104166666667, + "grad_norm": 0.9504824876785278, + "learning_rate": 9.338596282338632e-06, + "loss": 0.0323, + "step": 879 + }, + { + "epoch": 0.3819444444444444, + "grad_norm": 0.980655312538147, + "learning_rate": 9.336848127085972e-06, + "loss": 0.0538, + "step": 880 + }, + { + "epoch": 0.3823784722222222, + "grad_norm": 1.006734848022461, + "learning_rate": 9.335097828680035e-06, + "loss": 0.0497, + "step": 881 + }, + { + "epoch": 0.3828125, + "grad_norm": 1.2040218114852905, + "learning_rate": 9.333345387985777e-06, + "loss": 0.0711, + "step": 882 + }, + { + "epoch": 0.3832465277777778, + "grad_norm": 1.9897956848144531, + "learning_rate": 9.331590805869204e-06, + "loss": 0.0861, + "step": 883 + }, + { + "epoch": 0.3836805555555556, + "grad_norm": 1.4332646131515503, + "learning_rate": 9.329834083197382e-06, + "loss": 0.0663, + "step": 884 + }, + { + "epoch": 0.3841145833333333, + "grad_norm": 1.001236915588379, + "learning_rate": 9.32807522083844e-06, + "loss": 0.0723, + "step": 885 + }, + { + "epoch": 0.3845486111111111, + "grad_norm": 1.7063037157058716, + "learning_rate": 9.326314219661556e-06, + "loss": 0.0749, + "step": 886 + }, + { + "epoch": 0.3849826388888889, + "grad_norm": 0.6326516270637512, + "learning_rate": 9.324551080536972e-06, + "loss": 0.0539, + "step": 887 + }, + { + "epoch": 0.3854166666666667, + "grad_norm": 0.8508548736572266, + "learning_rate": 9.32278580433598e-06, + "loss": 0.046, + "step": 888 + }, + { + "epoch": 0.3858506944444444, + "grad_norm": 0.5909950137138367, + "learning_rate": 9.321018391930939e-06, + "loss": 0.0591, + "step": 889 + }, + { + "epoch": 0.3862847222222222, + "grad_norm": 0.4594472050666809, + "learning_rate": 9.319248844195248e-06, + "loss": 0.0428, + "step": 890 + }, + { + "epoch": 0.38671875, + "grad_norm": 0.42079484462738037, + "learning_rate": 9.317477162003376e-06, + "loss": 0.0352, + "step": 891 + }, + { + "epoch": 0.3871527777777778, + "grad_norm": 0.9136016964912415, + "learning_rate": 9.315703346230837e-06, + "loss": 0.0628, + "step": 892 + }, + { + "epoch": 0.3875868055555556, + "grad_norm": 0.7341428995132446, + "learning_rate": 9.313927397754205e-06, + "loss": 0.0535, + "step": 893 + }, + { + "epoch": 0.3880208333333333, + "grad_norm": 0.6203869581222534, + "learning_rate": 9.312149317451107e-06, + "loss": 0.0409, + "step": 894 + }, + { + "epoch": 0.3884548611111111, + "grad_norm": 0.4617060422897339, + "learning_rate": 9.310369106200218e-06, + "loss": 0.0214, + "step": 895 + }, + { + "epoch": 0.3888888888888889, + "grad_norm": 0.8191637992858887, + "learning_rate": 9.308586764881273e-06, + "loss": 0.0526, + "step": 896 + }, + { + "epoch": 0.3893229166666667, + "grad_norm": 1.3729549646377563, + "learning_rate": 9.306802294375056e-06, + "loss": 0.0779, + "step": 897 + }, + { + "epoch": 0.3897569444444444, + "grad_norm": 1.1696858406066895, + "learning_rate": 9.305015695563406e-06, + "loss": 0.0521, + "step": 898 + }, + { + "epoch": 0.3901909722222222, + "grad_norm": 1.3810319900512695, + "learning_rate": 9.30322696932921e-06, + "loss": 0.0716, + "step": 899 + }, + { + "epoch": 0.390625, + "grad_norm": 1.531559705734253, + "learning_rate": 9.30143611655641e-06, + "loss": 0.0768, + "step": 900 + }, + { + "epoch": 0.390625, + "eval_layout_rm_rank_test_accuracy": 0.9460030043441192, + "eval_layout_rm_rank_test_chosen_scores": 0.6126484870910645, + "eval_layout_rm_rank_test_loss": 0.06283911317586899, + "eval_layout_rm_rank_test_rejected_scores": 0.09528347104787827, + "eval_layout_rm_rank_test_runtime": 1176.1022, + "eval_layout_rm_rank_test_samples_per_second": 20.943, + "eval_layout_rm_rank_test_steps_per_second": 0.655, + "step": 900 + }, + { + "epoch": 0.3910590277777778, + "grad_norm": 1.5760308504104614, + "learning_rate": 9.299643138129995e-06, + "loss": 0.0661, + "step": 901 + }, + { + "epoch": 0.3914930555555556, + "grad_norm": 0.9271313548088074, + "learning_rate": 9.297848034936007e-06, + "loss": 0.0591, + "step": 902 + }, + { + "epoch": 0.3919270833333333, + "grad_norm": 1.0862689018249512, + "learning_rate": 9.296050807861538e-06, + "loss": 0.0826, + "step": 903 + }, + { + "epoch": 0.3923611111111111, + "grad_norm": 0.6337171196937561, + "learning_rate": 9.29425145779473e-06, + "loss": 0.0509, + "step": 904 + }, + { + "epoch": 0.3927951388888889, + "grad_norm": 0.7217728495597839, + "learning_rate": 9.292449985624771e-06, + "loss": 0.0407, + "step": 905 + }, + { + "epoch": 0.3932291666666667, + "grad_norm": 0.6203964948654175, + "learning_rate": 9.290646392241904e-06, + "loss": 0.0633, + "step": 906 + }, + { + "epoch": 0.3936631944444444, + "grad_norm": 0.4933103322982788, + "learning_rate": 9.288840678537412e-06, + "loss": 0.0455, + "step": 907 + }, + { + "epoch": 0.3940972222222222, + "grad_norm": 0.571537435054779, + "learning_rate": 9.28703284540363e-06, + "loss": 0.0584, + "step": 908 + }, + { + "epoch": 0.39453125, + "grad_norm": 0.4912552237510681, + "learning_rate": 9.285222893733943e-06, + "loss": 0.0415, + "step": 909 + }, + { + "epoch": 0.3949652777777778, + "grad_norm": 0.5689923167228699, + "learning_rate": 9.28341082442278e-06, + "loss": 0.0708, + "step": 910 + }, + { + "epoch": 0.3953993055555556, + "grad_norm": 0.49310195446014404, + "learning_rate": 9.281596638365615e-06, + "loss": 0.0424, + "step": 911 + }, + { + "epoch": 0.3958333333333333, + "grad_norm": 0.538809061050415, + "learning_rate": 9.27978033645897e-06, + "loss": 0.0513, + "step": 912 + }, + { + "epoch": 0.3962673611111111, + "grad_norm": 0.7360814213752747, + "learning_rate": 9.277961919600413e-06, + "loss": 0.0598, + "step": 913 + }, + { + "epoch": 0.3967013888888889, + "grad_norm": 0.6100806593894958, + "learning_rate": 9.276141388688557e-06, + "loss": 0.0347, + "step": 914 + }, + { + "epoch": 0.3971354166666667, + "grad_norm": 0.7792655229568481, + "learning_rate": 9.27431874462306e-06, + "loss": 0.0456, + "step": 915 + }, + { + "epoch": 0.3975694444444444, + "grad_norm": 1.1377533674240112, + "learning_rate": 9.272493988304621e-06, + "loss": 0.0721, + "step": 916 + }, + { + "epoch": 0.3980034722222222, + "grad_norm": 1.3492984771728516, + "learning_rate": 9.270667120634985e-06, + "loss": 0.0954, + "step": 917 + }, + { + "epoch": 0.3984375, + "grad_norm": 0.84878009557724, + "learning_rate": 9.268838142516943e-06, + "loss": 0.0449, + "step": 918 + }, + { + "epoch": 0.3988715277777778, + "grad_norm": 0.7643359899520874, + "learning_rate": 9.267007054854327e-06, + "loss": 0.0472, + "step": 919 + }, + { + "epoch": 0.3993055555555556, + "grad_norm": 0.8544363975524902, + "learning_rate": 9.26517385855201e-06, + "loss": 0.0617, + "step": 920 + }, + { + "epoch": 0.3997395833333333, + "grad_norm": 0.6072924733161926, + "learning_rate": 9.263338554515906e-06, + "loss": 0.0382, + "step": 921 + }, + { + "epoch": 0.4001736111111111, + "grad_norm": 0.9464804530143738, + "learning_rate": 9.261501143652975e-06, + "loss": 0.041, + "step": 922 + }, + { + "epoch": 0.4006076388888889, + "grad_norm": 0.6826421022415161, + "learning_rate": 9.259661626871217e-06, + "loss": 0.0559, + "step": 923 + }, + { + "epoch": 0.4010416666666667, + "grad_norm": 0.871494710445404, + "learning_rate": 9.25782000507967e-06, + "loss": 0.0757, + "step": 924 + }, + { + "epoch": 0.4014756944444444, + "grad_norm": 0.7773535251617432, + "learning_rate": 9.255976279188412e-06, + "loss": 0.0597, + "step": 925 + }, + { + "epoch": 0.4019097222222222, + "grad_norm": 0.6741392612457275, + "learning_rate": 9.254130450108567e-06, + "loss": 0.0409, + "step": 926 + }, + { + "epoch": 0.40234375, + "grad_norm": 0.9787711501121521, + "learning_rate": 9.25228251875229e-06, + "loss": 0.0507, + "step": 927 + }, + { + "epoch": 0.4027777777777778, + "grad_norm": 0.7004270553588867, + "learning_rate": 9.25043248603278e-06, + "loss": 0.0441, + "step": 928 + }, + { + "epoch": 0.4032118055555556, + "grad_norm": 1.1758992671966553, + "learning_rate": 9.248580352864273e-06, + "loss": 0.0761, + "step": 929 + }, + { + "epoch": 0.4036458333333333, + "grad_norm": 1.0420787334442139, + "learning_rate": 9.246726120162046e-06, + "loss": 0.0704, + "step": 930 + }, + { + "epoch": 0.4040798611111111, + "grad_norm": 0.7530882954597473, + "learning_rate": 9.244869788842406e-06, + "loss": 0.051, + "step": 931 + }, + { + "epoch": 0.4045138888888889, + "grad_norm": 0.8708016872406006, + "learning_rate": 9.243011359822704e-06, + "loss": 0.0427, + "step": 932 + }, + { + "epoch": 0.4049479166666667, + "grad_norm": 0.7921614646911621, + "learning_rate": 9.241150834021326e-06, + "loss": 0.0502, + "step": 933 + }, + { + "epoch": 0.4053819444444444, + "grad_norm": 0.6828484535217285, + "learning_rate": 9.239288212357694e-06, + "loss": 0.0558, + "step": 934 + }, + { + "epoch": 0.4058159722222222, + "grad_norm": 1.3188660144805908, + "learning_rate": 9.237423495752265e-06, + "loss": 0.0686, + "step": 935 + }, + { + "epoch": 0.40625, + "grad_norm": 1.3304487466812134, + "learning_rate": 9.23555668512653e-06, + "loss": 0.0532, + "step": 936 + }, + { + "epoch": 0.4066840277777778, + "grad_norm": 3.606907606124878, + "learning_rate": 9.233687781403017e-06, + "loss": 0.0708, + "step": 937 + }, + { + "epoch": 0.4071180555555556, + "grad_norm": 1.2716693878173828, + "learning_rate": 9.23181678550529e-06, + "loss": 0.101, + "step": 938 + }, + { + "epoch": 0.4075520833333333, + "grad_norm": 92.44752502441406, + "learning_rate": 9.229943698357942e-06, + "loss": 0.1493, + "step": 939 + }, + { + "epoch": 0.4079861111111111, + "grad_norm": 7.013112545013428, + "learning_rate": 9.228068520886604e-06, + "loss": 0.1524, + "step": 940 + }, + { + "epoch": 0.4084201388888889, + "grad_norm": 1.4616622924804688, + "learning_rate": 9.226191254017937e-06, + "loss": 0.1457, + "step": 941 + }, + { + "epoch": 0.4088541666666667, + "grad_norm": 78.1362533569336, + "learning_rate": 9.224311898679637e-06, + "loss": 0.1682, + "step": 942 + }, + { + "epoch": 0.4092881944444444, + "grad_norm": 34.54515075683594, + "learning_rate": 9.222430455800427e-06, + "loss": 0.1422, + "step": 943 + }, + { + "epoch": 0.4097222222222222, + "grad_norm": 39.45749282836914, + "learning_rate": 9.22054692631007e-06, + "loss": 0.1298, + "step": 944 + }, + { + "epoch": 0.41015625, + "grad_norm": 10.574701309204102, + "learning_rate": 9.218661311139352e-06, + "loss": 0.1894, + "step": 945 + }, + { + "epoch": 0.4105902777777778, + "grad_norm": 68.54234313964844, + "learning_rate": 9.216773611220094e-06, + "loss": 0.234, + "step": 946 + }, + { + "epoch": 0.4110243055555556, + "grad_norm": 48.71664810180664, + "learning_rate": 9.214883827485149e-06, + "loss": 0.256, + "step": 947 + }, + { + "epoch": 0.4114583333333333, + "grad_norm": 43.572425842285156, + "learning_rate": 9.212991960868392e-06, + "loss": 0.2555, + "step": 948 + }, + { + "epoch": 0.4118923611111111, + "grad_norm": 39.20548629760742, + "learning_rate": 9.211098012304736e-06, + "loss": 0.238, + "step": 949 + }, + { + "epoch": 0.4123263888888889, + "grad_norm": 67.0279769897461, + "learning_rate": 9.209201982730118e-06, + "loss": 0.3597, + "step": 950 + }, + { + "epoch": 0.4127604166666667, + "grad_norm": 43.72623062133789, + "learning_rate": 9.207303873081506e-06, + "loss": 0.3006, + "step": 951 + }, + { + "epoch": 0.4131944444444444, + "grad_norm": 11.67404556274414, + "learning_rate": 9.205403684296895e-06, + "loss": 0.2382, + "step": 952 + }, + { + "epoch": 0.4136284722222222, + "grad_norm": 11.699862480163574, + "learning_rate": 9.203501417315303e-06, + "loss": 0.2221, + "step": 953 + }, + { + "epoch": 0.4140625, + "grad_norm": 13.271985054016113, + "learning_rate": 9.201597073076784e-06, + "loss": 0.2537, + "step": 954 + }, + { + "epoch": 0.4144965277777778, + "grad_norm": 62.22166061401367, + "learning_rate": 9.199690652522409e-06, + "loss": 0.2715, + "step": 955 + }, + { + "epoch": 0.4149305555555556, + "grad_norm": 23.315994262695312, + "learning_rate": 9.197782156594284e-06, + "loss": 0.3128, + "step": 956 + }, + { + "epoch": 0.4153645833333333, + "grad_norm": 49.76742935180664, + "learning_rate": 9.195871586235531e-06, + "loss": 0.2271, + "step": 957 + }, + { + "epoch": 0.4157986111111111, + "grad_norm": 4.723781585693359, + "learning_rate": 9.19395894239031e-06, + "loss": 0.1952, + "step": 958 + }, + { + "epoch": 0.4162326388888889, + "grad_norm": 43.25361251831055, + "learning_rate": 9.19204422600379e-06, + "loss": 0.2309, + "step": 959 + }, + { + "epoch": 0.4166666666666667, + "grad_norm": 6.243020057678223, + "learning_rate": 9.190127438022175e-06, + "loss": 0.2332, + "step": 960 + }, + { + "epoch": 0.4171006944444444, + "grad_norm": 8.359853744506836, + "learning_rate": 9.18820857939269e-06, + "loss": 0.2122, + "step": 961 + }, + { + "epoch": 0.4175347222222222, + "grad_norm": 1.9266326427459717, + "learning_rate": 9.186287651063584e-06, + "loss": 0.1968, + "step": 962 + }, + { + "epoch": 0.41796875, + "grad_norm": 3.176651954650879, + "learning_rate": 9.184364653984126e-06, + "loss": 0.1657, + "step": 963 + }, + { + "epoch": 0.4184027777777778, + "grad_norm": 2.063450574874878, + "learning_rate": 9.182439589104612e-06, + "loss": 0.151, + "step": 964 + }, + { + "epoch": 0.4188368055555556, + "grad_norm": 27.922056198120117, + "learning_rate": 9.180512457376356e-06, + "loss": 0.1588, + "step": 965 + }, + { + "epoch": 0.4192708333333333, + "grad_norm": 4.232086181640625, + "learning_rate": 9.178583259751695e-06, + "loss": 0.1896, + "step": 966 + }, + { + "epoch": 0.4197048611111111, + "grad_norm": 4.373309135437012, + "learning_rate": 9.176651997183983e-06, + "loss": 0.1747, + "step": 967 + }, + { + "epoch": 0.4201388888888889, + "grad_norm": 3.191253900527954, + "learning_rate": 9.1747186706276e-06, + "loss": 0.1562, + "step": 968 + }, + { + "epoch": 0.4205729166666667, + "grad_norm": 8.11868667602539, + "learning_rate": 9.172783281037947e-06, + "loss": 0.1517, + "step": 969 + }, + { + "epoch": 0.4210069444444444, + "grad_norm": 28.279821395874023, + "learning_rate": 9.170845829371435e-06, + "loss": 0.1372, + "step": 970 + }, + { + "epoch": 0.4214409722222222, + "grad_norm": 23.073993682861328, + "learning_rate": 9.168906316585506e-06, + "loss": 0.1703, + "step": 971 + }, + { + "epoch": 0.421875, + "grad_norm": 6.043205261230469, + "learning_rate": 9.166964743638612e-06, + "loss": 0.15, + "step": 972 + }, + { + "epoch": 0.4223090277777778, + "grad_norm": 20.86329460144043, + "learning_rate": 9.165021111490226e-06, + "loss": 0.145, + "step": 973 + }, + { + "epoch": 0.4227430555555556, + "grad_norm": 7.663023471832275, + "learning_rate": 9.16307542110084e-06, + "loss": 0.123, + "step": 974 + }, + { + "epoch": 0.4231770833333333, + "grad_norm": 4.735728740692139, + "learning_rate": 9.16112767343196e-06, + "loss": 0.1135, + "step": 975 + }, + { + "epoch": 0.4236111111111111, + "grad_norm": 9.393878936767578, + "learning_rate": 9.159177869446112e-06, + "loss": 0.1371, + "step": 976 + }, + { + "epoch": 0.4240451388888889, + "grad_norm": 33.43635940551758, + "learning_rate": 9.157226010106835e-06, + "loss": 0.2043, + "step": 977 + }, + { + "epoch": 0.4244791666666667, + "grad_norm": 13.90460205078125, + "learning_rate": 9.155272096378685e-06, + "loss": 0.1522, + "step": 978 + }, + { + "epoch": 0.4249131944444444, + "grad_norm": 4.844347953796387, + "learning_rate": 9.153316129227234e-06, + "loss": 0.1117, + "step": 979 + }, + { + "epoch": 0.4253472222222222, + "grad_norm": 2.9909286499023438, + "learning_rate": 9.15135810961907e-06, + "loss": 0.0999, + "step": 980 + }, + { + "epoch": 0.42578125, + "grad_norm": 9.572826385498047, + "learning_rate": 9.14939803852179e-06, + "loss": 0.1549, + "step": 981 + }, + { + "epoch": 0.4262152777777778, + "grad_norm": 2.522773027420044, + "learning_rate": 9.147435916904011e-06, + "loss": 0.1092, + "step": 982 + }, + { + "epoch": 0.4266493055555556, + "grad_norm": 16.659290313720703, + "learning_rate": 9.145471745735359e-06, + "loss": 0.121, + "step": 983 + }, + { + "epoch": 0.4270833333333333, + "grad_norm": 4.65269136428833, + "learning_rate": 9.143505525986476e-06, + "loss": 0.1374, + "step": 984 + }, + { + "epoch": 0.4275173611111111, + "grad_norm": 1.8901381492614746, + "learning_rate": 9.141537258629012e-06, + "loss": 0.1186, + "step": 985 + }, + { + "epoch": 0.4279513888888889, + "grad_norm": 2.9920859336853027, + "learning_rate": 9.139566944635634e-06, + "loss": 0.1187, + "step": 986 + }, + { + "epoch": 0.4283854166666667, + "grad_norm": 1.7451281547546387, + "learning_rate": 9.137594584980017e-06, + "loss": 0.0959, + "step": 987 + }, + { + "epoch": 0.4288194444444444, + "grad_norm": 3.1266098022460938, + "learning_rate": 9.135620180636848e-06, + "loss": 0.1327, + "step": 988 + }, + { + "epoch": 0.4292534722222222, + "grad_norm": 1.3737506866455078, + "learning_rate": 9.133643732581824e-06, + "loss": 0.0942, + "step": 989 + }, + { + "epoch": 0.4296875, + "grad_norm": 1.38458251953125, + "learning_rate": 9.131665241791654e-06, + "loss": 0.0729, + "step": 990 + }, + { + "epoch": 0.4301215277777778, + "grad_norm": 1.934052586555481, + "learning_rate": 9.129684709244053e-06, + "loss": 0.0995, + "step": 991 + }, + { + "epoch": 0.4305555555555556, + "grad_norm": 21.256853103637695, + "learning_rate": 9.127702135917747e-06, + "loss": 0.0876, + "step": 992 + }, + { + "epoch": 0.4309895833333333, + "grad_norm": 7.968606472015381, + "learning_rate": 9.12571752279247e-06, + "loss": 0.0875, + "step": 993 + }, + { + "epoch": 0.4314236111111111, + "grad_norm": 6.5588250160217285, + "learning_rate": 9.123730870848963e-06, + "loss": 0.1262, + "step": 994 + }, + { + "epoch": 0.4318576388888889, + "grad_norm": 28.4091739654541, + "learning_rate": 9.12174218106898e-06, + "loss": 0.1219, + "step": 995 + }, + { + "epoch": 0.4322916666666667, + "grad_norm": 4.547117233276367, + "learning_rate": 9.119751454435275e-06, + "loss": 0.1213, + "step": 996 + }, + { + "epoch": 0.4327256944444444, + "grad_norm": 4.822508335113525, + "learning_rate": 9.11775869193161e-06, + "loss": 0.1223, + "step": 997 + }, + { + "epoch": 0.4331597222222222, + "grad_norm": 1.3423714637756348, + "learning_rate": 9.115763894542758e-06, + "loss": 0.0783, + "step": 998 + }, + { + "epoch": 0.43359375, + "grad_norm": 1.7178046703338623, + "learning_rate": 9.11376706325449e-06, + "loss": 0.0802, + "step": 999 + }, + { + "epoch": 0.4340277777777778, + "grad_norm": 5.980998992919922, + "learning_rate": 9.111768199053588e-06, + "loss": 0.0992, + "step": 1000 + }, + { + "epoch": 0.4344618055555556, + "grad_norm": 1.5980966091156006, + "learning_rate": 9.109767302927839e-06, + "loss": 0.0947, + "step": 1001 + }, + { + "epoch": 0.4348958333333333, + "grad_norm": 1.964292049407959, + "learning_rate": 9.107764375866028e-06, + "loss": 0.1008, + "step": 1002 + }, + { + "epoch": 0.4353298611111111, + "grad_norm": 1.4555721282958984, + "learning_rate": 9.10575941885795e-06, + "loss": 0.084, + "step": 1003 + }, + { + "epoch": 0.4357638888888889, + "grad_norm": 1.8440698385238647, + "learning_rate": 9.103752432894398e-06, + "loss": 0.1092, + "step": 1004 + }, + { + "epoch": 0.4361979166666667, + "grad_norm": 10.594161033630371, + "learning_rate": 9.101743418967174e-06, + "loss": 0.0622, + "step": 1005 + }, + { + "epoch": 0.4366319444444444, + "grad_norm": 1.5675795078277588, + "learning_rate": 9.099732378069076e-06, + "loss": 0.1005, + "step": 1006 + }, + { + "epoch": 0.4370659722222222, + "grad_norm": 1.1693589687347412, + "learning_rate": 9.097719311193907e-06, + "loss": 0.1029, + "step": 1007 + }, + { + "epoch": 0.4375, + "grad_norm": 0.9129657745361328, + "learning_rate": 9.09570421933647e-06, + "loss": 0.0799, + "step": 1008 + }, + { + "epoch": 0.4379340277777778, + "grad_norm": 0.9033944010734558, + "learning_rate": 9.093687103492568e-06, + "loss": 0.0671, + "step": 1009 + }, + { + "epoch": 0.4383680555555556, + "grad_norm": 1.1482619047164917, + "learning_rate": 9.091667964659006e-06, + "loss": 0.0788, + "step": 1010 + }, + { + "epoch": 0.4388020833333333, + "grad_norm": 1.9808341264724731, + "learning_rate": 9.089646803833589e-06, + "loss": 0.0724, + "step": 1011 + }, + { + "epoch": 0.4392361111111111, + "grad_norm": 1.0761967897415161, + "learning_rate": 9.08762362201512e-06, + "loss": 0.0662, + "step": 1012 + }, + { + "epoch": 0.4396701388888889, + "grad_norm": 1.2144149541854858, + "learning_rate": 9.085598420203399e-06, + "loss": 0.0727, + "step": 1013 + }, + { + "epoch": 0.4401041666666667, + "grad_norm": 0.7854107618331909, + "learning_rate": 9.083571199399226e-06, + "loss": 0.0791, + "step": 1014 + }, + { + "epoch": 0.4405381944444444, + "grad_norm": 1.2169189453125, + "learning_rate": 9.0815419606044e-06, + "loss": 0.0784, + "step": 1015 + }, + { + "epoch": 0.4409722222222222, + "grad_norm": 0.8333733081817627, + "learning_rate": 9.079510704821718e-06, + "loss": 0.0771, + "step": 1016 + }, + { + "epoch": 0.44140625, + "grad_norm": 2.004037857055664, + "learning_rate": 9.077477433054969e-06, + "loss": 0.0582, + "step": 1017 + }, + { + "epoch": 0.4418402777777778, + "grad_norm": 1.4720275402069092, + "learning_rate": 9.075442146308942e-06, + "loss": 0.073, + "step": 1018 + }, + { + "epoch": 0.4422743055555556, + "grad_norm": 0.634091854095459, + "learning_rate": 9.07340484558942e-06, + "loss": 0.0397, + "step": 1019 + }, + { + "epoch": 0.4427083333333333, + "grad_norm": 1.0841357707977295, + "learning_rate": 9.071365531903182e-06, + "loss": 0.0938, + "step": 1020 + }, + { + "epoch": 0.4431423611111111, + "grad_norm": 2.343041181564331, + "learning_rate": 9.069324206258003e-06, + "loss": 0.0891, + "step": 1021 + }, + { + "epoch": 0.4435763888888889, + "grad_norm": 1.1680972576141357, + "learning_rate": 9.06728086966265e-06, + "loss": 0.0546, + "step": 1022 + }, + { + "epoch": 0.4440104166666667, + "grad_norm": 0.9939153790473938, + "learning_rate": 9.065235523126886e-06, + "loss": 0.0732, + "step": 1023 + }, + { + "epoch": 0.4444444444444444, + "grad_norm": 1.5492585897445679, + "learning_rate": 9.063188167661463e-06, + "loss": 0.0707, + "step": 1024 + }, + { + "epoch": 0.4448784722222222, + "grad_norm": 1.1362593173980713, + "learning_rate": 9.061138804278134e-06, + "loss": 0.0978, + "step": 1025 + }, + { + "epoch": 0.4453125, + "grad_norm": 1.016692876815796, + "learning_rate": 9.059087433989633e-06, + "loss": 0.0663, + "step": 1026 + }, + { + "epoch": 0.4457465277777778, + "grad_norm": 1.037646770477295, + "learning_rate": 9.057034057809697e-06, + "loss": 0.0792, + "step": 1027 + }, + { + "epoch": 0.4461805555555556, + "grad_norm": 2.41023850440979, + "learning_rate": 9.054978676753045e-06, + "loss": 0.0951, + "step": 1028 + }, + { + "epoch": 0.4466145833333333, + "grad_norm": 1.0070242881774902, + "learning_rate": 9.052921291835395e-06, + "loss": 0.0646, + "step": 1029 + }, + { + "epoch": 0.4470486111111111, + "grad_norm": 2.8583602905273438, + "learning_rate": 9.050861904073447e-06, + "loss": 0.082, + "step": 1030 + }, + { + "epoch": 0.4474826388888889, + "grad_norm": 1.827767014503479, + "learning_rate": 9.048800514484897e-06, + "loss": 0.0818, + "step": 1031 + }, + { + "epoch": 0.4479166666666667, + "grad_norm": 2.483755350112915, + "learning_rate": 9.046737124088428e-06, + "loss": 0.0591, + "step": 1032 + }, + { + "epoch": 0.4483506944444444, + "grad_norm": 1.2786849737167358, + "learning_rate": 9.044671733903712e-06, + "loss": 0.0886, + "step": 1033 + }, + { + "epoch": 0.4487847222222222, + "grad_norm": 0.9632377624511719, + "learning_rate": 9.04260434495141e-06, + "loss": 0.0591, + "step": 1034 + }, + { + "epoch": 0.44921875, + "grad_norm": 1.3248785734176636, + "learning_rate": 9.040534958253166e-06, + "loss": 0.0667, + "step": 1035 + }, + { + "epoch": 0.4496527777777778, + "grad_norm": 0.7189981937408447, + "learning_rate": 9.038463574831617e-06, + "loss": 0.0641, + "step": 1036 + }, + { + "epoch": 0.4500868055555556, + "grad_norm": 0.7724950313568115, + "learning_rate": 9.036390195710388e-06, + "loss": 0.0513, + "step": 1037 + }, + { + "epoch": 0.4505208333333333, + "grad_norm": 1.7117736339569092, + "learning_rate": 9.034314821914083e-06, + "loss": 0.0645, + "step": 1038 + }, + { + "epoch": 0.4509548611111111, + "grad_norm": 32.858551025390625, + "learning_rate": 9.032237454468298e-06, + "loss": 0.0486, + "step": 1039 + }, + { + "epoch": 0.4513888888888889, + "grad_norm": 0.7045724391937256, + "learning_rate": 9.03015809439961e-06, + "loss": 0.0514, + "step": 1040 + }, + { + "epoch": 0.4518229166666667, + "grad_norm": 0.6073102355003357, + "learning_rate": 9.028076742735583e-06, + "loss": 0.0487, + "step": 1041 + }, + { + "epoch": 0.4522569444444444, + "grad_norm": 1.0917977094650269, + "learning_rate": 9.025993400504766e-06, + "loss": 0.0608, + "step": 1042 + }, + { + "epoch": 0.4526909722222222, + "grad_norm": 4.855196475982666, + "learning_rate": 9.023908068736688e-06, + "loss": 0.0763, + "step": 1043 + }, + { + "epoch": 0.453125, + "grad_norm": 2.3733937740325928, + "learning_rate": 9.021820748461868e-06, + "loss": 0.077, + "step": 1044 + }, + { + "epoch": 0.4535590277777778, + "grad_norm": 11.598405838012695, + "learning_rate": 9.0197314407118e-06, + "loss": 0.0901, + "step": 1045 + }, + { + "epoch": 0.4539930555555556, + "grad_norm": 0.9859849810600281, + "learning_rate": 9.017640146518964e-06, + "loss": 0.029, + "step": 1046 + }, + { + "epoch": 0.4544270833333333, + "grad_norm": 0.6398284435272217, + "learning_rate": 9.015546866916823e-06, + "loss": 0.0509, + "step": 1047 + }, + { + "epoch": 0.4548611111111111, + "grad_norm": 0.8597634434700012, + "learning_rate": 9.013451602939818e-06, + "loss": 0.0555, + "step": 1048 + }, + { + "epoch": 0.4552951388888889, + "grad_norm": 0.7557821869850159, + "learning_rate": 9.01135435562337e-06, + "loss": 0.0634, + "step": 1049 + }, + { + "epoch": 0.4557291666666667, + "grad_norm": 1.052837610244751, + "learning_rate": 9.009255126003883e-06, + "loss": 0.0639, + "step": 1050 + }, + { + "epoch": 0.4561631944444444, + "grad_norm": 0.8531177043914795, + "learning_rate": 9.00715391511874e-06, + "loss": 0.0651, + "step": 1051 + }, + { + "epoch": 0.4565972222222222, + "grad_norm": 0.8589138984680176, + "learning_rate": 9.005050724006305e-06, + "loss": 0.0677, + "step": 1052 + }, + { + "epoch": 0.45703125, + "grad_norm": 0.5279788970947266, + "learning_rate": 9.002945553705915e-06, + "loss": 0.0518, + "step": 1053 + }, + { + "epoch": 0.4574652777777778, + "grad_norm": 0.5493040680885315, + "learning_rate": 9.00083840525789e-06, + "loss": 0.0452, + "step": 1054 + }, + { + "epoch": 0.4578993055555556, + "grad_norm": 0.4292406737804413, + "learning_rate": 8.998729279703527e-06, + "loss": 0.0248, + "step": 1055 + }, + { + "epoch": 0.4583333333333333, + "grad_norm": 0.5679145455360413, + "learning_rate": 8.996618178085095e-06, + "loss": 0.0414, + "step": 1056 + }, + { + "epoch": 0.4587673611111111, + "grad_norm": 0.8989630937576294, + "learning_rate": 8.994505101445845e-06, + "loss": 0.0517, + "step": 1057 + }, + { + "epoch": 0.4592013888888889, + "grad_norm": 0.522352933883667, + "learning_rate": 8.992390050830005e-06, + "loss": 0.0354, + "step": 1058 + }, + { + "epoch": 0.4596354166666667, + "grad_norm": 0.7745173573493958, + "learning_rate": 8.990273027282774e-06, + "loss": 0.0589, + "step": 1059 + }, + { + "epoch": 0.4600694444444444, + "grad_norm": 1.2535181045532227, + "learning_rate": 8.988154031850328e-06, + "loss": 0.0539, + "step": 1060 + }, + { + "epoch": 0.4605034722222222, + "grad_norm": 0.5835824012756348, + "learning_rate": 8.986033065579819e-06, + "loss": 0.035, + "step": 1061 + }, + { + "epoch": 0.4609375, + "grad_norm": 1.5330524444580078, + "learning_rate": 8.983910129519367e-06, + "loss": 0.0911, + "step": 1062 + }, + { + "epoch": 0.4613715277777778, + "grad_norm": 0.8115177750587463, + "learning_rate": 8.981785224718073e-06, + "loss": 0.065, + "step": 1063 + }, + { + "epoch": 0.4618055555555556, + "grad_norm": 0.8075844049453735, + "learning_rate": 8.979658352226006e-06, + "loss": 0.0637, + "step": 1064 + }, + { + "epoch": 0.4622395833333333, + "grad_norm": 0.8678948879241943, + "learning_rate": 8.977529513094212e-06, + "loss": 0.0705, + "step": 1065 + }, + { + "epoch": 0.4626736111111111, + "grad_norm": 1.0560071468353271, + "learning_rate": 8.975398708374702e-06, + "loss": 0.072, + "step": 1066 + }, + { + "epoch": 0.4631076388888889, + "grad_norm": 0.5890426635742188, + "learning_rate": 8.973265939120465e-06, + "loss": 0.0577, + "step": 1067 + }, + { + "epoch": 0.4635416666666667, + "grad_norm": 0.9163851737976074, + "learning_rate": 8.971131206385457e-06, + "loss": 0.0633, + "step": 1068 + }, + { + "epoch": 0.4639756944444444, + "grad_norm": 0.6653317809104919, + "learning_rate": 8.968994511224606e-06, + "loss": 0.0658, + "step": 1069 + }, + { + "epoch": 0.4644097222222222, + "grad_norm": 0.8337730169296265, + "learning_rate": 8.966855854693805e-06, + "loss": 0.0492, + "step": 1070 + }, + { + "epoch": 0.46484375, + "grad_norm": 0.7300952672958374, + "learning_rate": 8.964715237849928e-06, + "loss": 0.0733, + "step": 1071 + }, + { + "epoch": 0.4652777777777778, + "grad_norm": 1.1915796995162964, + "learning_rate": 8.962572661750802e-06, + "loss": 0.0896, + "step": 1072 + }, + { + "epoch": 0.4657118055555556, + "grad_norm": 0.898739218711853, + "learning_rate": 8.960428127455236e-06, + "loss": 0.0527, + "step": 1073 + }, + { + "epoch": 0.4661458333333333, + "grad_norm": 0.6602903604507446, + "learning_rate": 8.958281636022998e-06, + "loss": 0.0511, + "step": 1074 + }, + { + "epoch": 0.4665798611111111, + "grad_norm": 0.735146164894104, + "learning_rate": 8.956133188514827e-06, + "loss": 0.0372, + "step": 1075 + }, + { + "epoch": 0.4670138888888889, + "grad_norm": 0.6647898554801941, + "learning_rate": 8.953982785992428e-06, + "loss": 0.031, + "step": 1076 + }, + { + "epoch": 0.4674479166666667, + "grad_norm": 1.3238353729248047, + "learning_rate": 8.951830429518471e-06, + "loss": 0.0795, + "step": 1077 + }, + { + "epoch": 0.4678819444444444, + "grad_norm": 0.9088767766952515, + "learning_rate": 8.949676120156593e-06, + "loss": 0.0487, + "step": 1078 + }, + { + "epoch": 0.4683159722222222, + "grad_norm": 0.882414698600769, + "learning_rate": 8.947519858971398e-06, + "loss": 0.0464, + "step": 1079 + }, + { + "epoch": 0.46875, + "grad_norm": 1.1095579862594604, + "learning_rate": 8.945361647028447e-06, + "loss": 0.0553, + "step": 1080 + }, + { + "epoch": 0.4691840277777778, + "grad_norm": 1.1162978410720825, + "learning_rate": 8.943201485394276e-06, + "loss": 0.0579, + "step": 1081 + }, + { + "epoch": 0.4696180555555556, + "grad_norm": 0.7105571627616882, + "learning_rate": 8.94103937513637e-06, + "loss": 0.0349, + "step": 1082 + }, + { + "epoch": 0.4700520833333333, + "grad_norm": 0.9360809326171875, + "learning_rate": 8.938875317323198e-06, + "loss": 0.058, + "step": 1083 + }, + { + "epoch": 0.4704861111111111, + "grad_norm": 0.7727855443954468, + "learning_rate": 8.936709313024165e-06, + "loss": 0.043, + "step": 1084 + }, + { + "epoch": 0.4709201388888889, + "grad_norm": 0.773892879486084, + "learning_rate": 8.934541363309664e-06, + "loss": 0.0568, + "step": 1085 + }, + { + "epoch": 0.4713541666666667, + "grad_norm": 1.0830212831497192, + "learning_rate": 8.932371469251028e-06, + "loss": 0.0757, + "step": 1086 + }, + { + "epoch": 0.4717881944444444, + "grad_norm": 0.6006203889846802, + "learning_rate": 8.930199631920567e-06, + "loss": 0.044, + "step": 1087 + }, + { + "epoch": 0.4722222222222222, + "grad_norm": 0.7148776650428772, + "learning_rate": 8.928025852391539e-06, + "loss": 0.0493, + "step": 1088 + }, + { + "epoch": 0.47265625, + "grad_norm": 0.6793000102043152, + "learning_rate": 8.925850131738169e-06, + "loss": 0.0616, + "step": 1089 + }, + { + "epoch": 0.4730902777777778, + "grad_norm": 0.6601008772850037, + "learning_rate": 8.923672471035643e-06, + "loss": 0.0675, + "step": 1090 + }, + { + "epoch": 0.4735243055555556, + "grad_norm": 0.603920578956604, + "learning_rate": 8.921492871360098e-06, + "loss": 0.0612, + "step": 1091 + }, + { + "epoch": 0.4739583333333333, + "grad_norm": 0.7035510540008545, + "learning_rate": 8.919311333788633e-06, + "loss": 0.0479, + "step": 1092 + }, + { + "epoch": 0.4743923611111111, + "grad_norm": 0.6087512373924255, + "learning_rate": 8.917127859399308e-06, + "loss": 0.0476, + "step": 1093 + }, + { + "epoch": 0.4748263888888889, + "grad_norm": 0.6310524940490723, + "learning_rate": 8.914942449271134e-06, + "loss": 0.0549, + "step": 1094 + }, + { + "epoch": 0.4752604166666667, + "grad_norm": 0.7847822904586792, + "learning_rate": 8.912755104484085e-06, + "loss": 0.0612, + "step": 1095 + }, + { + "epoch": 0.4756944444444444, + "grad_norm": 0.5045846700668335, + "learning_rate": 8.910565826119084e-06, + "loss": 0.0364, + "step": 1096 + }, + { + "epoch": 0.4761284722222222, + "grad_norm": 0.7675495743751526, + "learning_rate": 8.908374615258016e-06, + "loss": 0.062, + "step": 1097 + }, + { + "epoch": 0.4765625, + "grad_norm": 0.784545361995697, + "learning_rate": 8.906181472983717e-06, + "loss": 0.0458, + "step": 1098 + }, + { + "epoch": 0.4769965277777778, + "grad_norm": 1.0218698978424072, + "learning_rate": 8.90398640037998e-06, + "loss": 0.0656, + "step": 1099 + }, + { + "epoch": 0.4774305555555556, + "grad_norm": 0.5361638069152832, + "learning_rate": 8.901789398531551e-06, + "loss": 0.0456, + "step": 1100 + }, + { + "epoch": 0.4778645833333333, + "grad_norm": 0.6312048435211182, + "learning_rate": 8.899590468524125e-06, + "loss": 0.0485, + "step": 1101 + }, + { + "epoch": 0.4782986111111111, + "grad_norm": 0.8763497471809387, + "learning_rate": 8.897389611444359e-06, + "loss": 0.0427, + "step": 1102 + }, + { + "epoch": 0.4787326388888889, + "grad_norm": 0.9114484786987305, + "learning_rate": 8.895186828379855e-06, + "loss": 0.0651, + "step": 1103 + }, + { + "epoch": 0.4791666666666667, + "grad_norm": 1.0543360710144043, + "learning_rate": 8.892982120419166e-06, + "loss": 0.0765, + "step": 1104 + }, + { + "epoch": 0.4796006944444444, + "grad_norm": 1.0119502544403076, + "learning_rate": 8.890775488651804e-06, + "loss": 0.0747, + "step": 1105 + }, + { + "epoch": 0.4800347222222222, + "grad_norm": 0.5783020257949829, + "learning_rate": 8.888566934168223e-06, + "loss": 0.0371, + "step": 1106 + }, + { + "epoch": 0.48046875, + "grad_norm": 0.5611697435379028, + "learning_rate": 8.886356458059836e-06, + "loss": 0.0469, + "step": 1107 + }, + { + "epoch": 0.4809027777777778, + "grad_norm": 1.545137882232666, + "learning_rate": 8.884144061418993e-06, + "loss": 0.0574, + "step": 1108 + }, + { + "epoch": 0.4813368055555556, + "grad_norm": 0.4729805290699005, + "learning_rate": 8.881929745339008e-06, + "loss": 0.041, + "step": 1109 + }, + { + "epoch": 0.4817708333333333, + "grad_norm": 0.6251962780952454, + "learning_rate": 8.879713510914132e-06, + "loss": 0.0535, + "step": 1110 + }, + { + "epoch": 0.4822048611111111, + "grad_norm": 0.5359340906143188, + "learning_rate": 8.87749535923957e-06, + "loss": 0.0436, + "step": 1111 + }, + { + "epoch": 0.4826388888888889, + "grad_norm": 0.6215574741363525, + "learning_rate": 8.875275291411473e-06, + "loss": 0.0552, + "step": 1112 + }, + { + "epoch": 0.4830729166666667, + "grad_norm": 0.5203942656517029, + "learning_rate": 8.873053308526938e-06, + "loss": 0.0617, + "step": 1113 + }, + { + "epoch": 0.4835069444444444, + "grad_norm": 0.4644133150577545, + "learning_rate": 8.870829411684007e-06, + "loss": 0.031, + "step": 1114 + }, + { + "epoch": 0.4839409722222222, + "grad_norm": 0.5880697965621948, + "learning_rate": 8.868603601981672e-06, + "loss": 0.0415, + "step": 1115 + }, + { + "epoch": 0.484375, + "grad_norm": 1.0086796283721924, + "learning_rate": 8.866375880519868e-06, + "loss": 0.0526, + "step": 1116 + }, + { + "epoch": 0.4848090277777778, + "grad_norm": 1.1692084074020386, + "learning_rate": 8.864146248399476e-06, + "loss": 0.0713, + "step": 1117 + }, + { + "epoch": 0.4852430555555556, + "grad_norm": 0.7209148406982422, + "learning_rate": 8.861914706722316e-06, + "loss": 0.0501, + "step": 1118 + }, + { + "epoch": 0.4856770833333333, + "grad_norm": 0.5479106903076172, + "learning_rate": 8.85968125659116e-06, + "loss": 0.0426, + "step": 1119 + }, + { + "epoch": 0.4861111111111111, + "grad_norm": 0.8221974968910217, + "learning_rate": 8.857445899109716e-06, + "loss": 0.0591, + "step": 1120 + }, + { + "epoch": 0.4865451388888889, + "grad_norm": 0.5368654131889343, + "learning_rate": 8.855208635382637e-06, + "loss": 0.0327, + "step": 1121 + }, + { + "epoch": 0.4869791666666667, + "grad_norm": 0.794711172580719, + "learning_rate": 8.852969466515522e-06, + "loss": 0.0695, + "step": 1122 + }, + { + "epoch": 0.4874131944444444, + "grad_norm": 0.8020004034042358, + "learning_rate": 8.850728393614903e-06, + "loss": 0.0524, + "step": 1123 + }, + { + "epoch": 0.4878472222222222, + "grad_norm": 0.6102003455162048, + "learning_rate": 8.848485417788262e-06, + "loss": 0.0399, + "step": 1124 + }, + { + "epoch": 0.48828125, + "grad_norm": 0.7068143486976624, + "learning_rate": 8.846240540144013e-06, + "loss": 0.0561, + "step": 1125 + }, + { + "epoch": 0.4887152777777778, + "grad_norm": 0.9423394203186035, + "learning_rate": 8.843993761791518e-06, + "loss": 0.0814, + "step": 1126 + }, + { + "epoch": 0.4891493055555556, + "grad_norm": 0.5660754442214966, + "learning_rate": 8.841745083841071e-06, + "loss": 0.047, + "step": 1127 + }, + { + "epoch": 0.4895833333333333, + "grad_norm": 0.5480889678001404, + "learning_rate": 8.839494507403909e-06, + "loss": 0.0481, + "step": 1128 + }, + { + "epoch": 0.4900173611111111, + "grad_norm": 0.6959866881370544, + "learning_rate": 8.837242033592208e-06, + "loss": 0.0436, + "step": 1129 + }, + { + "epoch": 0.4904513888888889, + "grad_norm": 0.867567777633667, + "learning_rate": 8.834987663519079e-06, + "loss": 0.0702, + "step": 1130 + }, + { + "epoch": 0.4908854166666667, + "grad_norm": 0.6856876015663147, + "learning_rate": 8.832731398298567e-06, + "loss": 0.0743, + "step": 1131 + }, + { + "epoch": 0.4913194444444444, + "grad_norm": 0.5049101114273071, + "learning_rate": 8.830473239045663e-06, + "loss": 0.0554, + "step": 1132 + }, + { + "epoch": 0.4917534722222222, + "grad_norm": 0.5671724081039429, + "learning_rate": 8.828213186876283e-06, + "loss": 0.0438, + "step": 1133 + }, + { + "epoch": 0.4921875, + "grad_norm": 0.6206570863723755, + "learning_rate": 8.825951242907288e-06, + "loss": 0.0399, + "step": 1134 + }, + { + "epoch": 0.4926215277777778, + "grad_norm": 0.5550037622451782, + "learning_rate": 8.823687408256468e-06, + "loss": 0.0433, + "step": 1135 + }, + { + "epoch": 0.4930555555555556, + "grad_norm": 0.8700177669525146, + "learning_rate": 8.82142168404255e-06, + "loss": 0.0727, + "step": 1136 + }, + { + "epoch": 0.4934895833333333, + "grad_norm": 0.9669756293296814, + "learning_rate": 8.81915407138519e-06, + "loss": 0.0738, + "step": 1137 + }, + { + "epoch": 0.4939236111111111, + "grad_norm": 0.7427315711975098, + "learning_rate": 8.816884571404987e-06, + "loss": 0.048, + "step": 1138 + }, + { + "epoch": 0.4943576388888889, + "grad_norm": 1.1001490354537964, + "learning_rate": 8.814613185223462e-06, + "loss": 0.0752, + "step": 1139 + }, + { + "epoch": 0.4947916666666667, + "grad_norm": 0.7257412075996399, + "learning_rate": 8.812339913963074e-06, + "loss": 0.0563, + "step": 1140 + }, + { + "epoch": 0.4952256944444444, + "grad_norm": 1.2158013582229614, + "learning_rate": 8.810064758747212e-06, + "loss": 0.0592, + "step": 1141 + }, + { + "epoch": 0.4956597222222222, + "grad_norm": 0.4755469262599945, + "learning_rate": 8.807787720700199e-06, + "loss": 0.0584, + "step": 1142 + }, + { + "epoch": 0.49609375, + "grad_norm": 0.4486525058746338, + "learning_rate": 8.80550880094728e-06, + "loss": 0.0314, + "step": 1143 + }, + { + "epoch": 0.4965277777777778, + "grad_norm": 0.6543397307395935, + "learning_rate": 8.803228000614642e-06, + "loss": 0.0534, + "step": 1144 + }, + { + "epoch": 0.4969618055555556, + "grad_norm": 0.8404059410095215, + "learning_rate": 8.800945320829391e-06, + "loss": 0.0746, + "step": 1145 + }, + { + "epoch": 0.4973958333333333, + "grad_norm": 0.7537004947662354, + "learning_rate": 8.798660762719567e-06, + "loss": 0.0408, + "step": 1146 + }, + { + "epoch": 0.4978298611111111, + "grad_norm": 0.8327997922897339, + "learning_rate": 8.796374327414135e-06, + "loss": 0.0513, + "step": 1147 + }, + { + "epoch": 0.4982638888888889, + "grad_norm": 0.8455732464790344, + "learning_rate": 8.794086016042993e-06, + "loss": 0.0599, + "step": 1148 + }, + { + "epoch": 0.4986979166666667, + "grad_norm": 0.5204771757125854, + "learning_rate": 8.79179582973696e-06, + "loss": 0.0413, + "step": 1149 + }, + { + "epoch": 0.4991319444444444, + "grad_norm": 0.8038812875747681, + "learning_rate": 8.789503769627786e-06, + "loss": 0.0552, + "step": 1150 + }, + { + "epoch": 0.4995659722222222, + "grad_norm": 0.781603217124939, + "learning_rate": 8.787209836848144e-06, + "loss": 0.0594, + "step": 1151 + }, + { + "epoch": 0.5, + "grad_norm": 0.508056640625, + "learning_rate": 8.784914032531636e-06, + "loss": 0.0434, + "step": 1152 + }, + { + "epoch": 0.5004340277777778, + "grad_norm": 0.5347216725349426, + "learning_rate": 8.782616357812784e-06, + "loss": 0.0614, + "step": 1153 + }, + { + "epoch": 0.5008680555555556, + "grad_norm": 0.7684710621833801, + "learning_rate": 8.780316813827037e-06, + "loss": 0.0601, + "step": 1154 + }, + { + "epoch": 0.5013020833333334, + "grad_norm": 0.625655472278595, + "learning_rate": 8.778015401710766e-06, + "loss": 0.0524, + "step": 1155 + }, + { + "epoch": 0.5017361111111112, + "grad_norm": 1.0762808322906494, + "learning_rate": 8.775712122601273e-06, + "loss": 0.0763, + "step": 1156 + }, + { + "epoch": 0.5021701388888888, + "grad_norm": 0.592757523059845, + "learning_rate": 8.773406977636769e-06, + "loss": 0.0445, + "step": 1157 + }, + { + "epoch": 0.5026041666666666, + "grad_norm": 0.8303428292274475, + "learning_rate": 8.771099967956398e-06, + "loss": 0.0608, + "step": 1158 + }, + { + "epoch": 0.5030381944444444, + "grad_norm": 1.07639479637146, + "learning_rate": 8.768791094700222e-06, + "loss": 0.0608, + "step": 1159 + }, + { + "epoch": 0.5034722222222222, + "grad_norm": 4.158738613128662, + "learning_rate": 8.766480359009224e-06, + "loss": 0.065, + "step": 1160 + }, + { + "epoch": 0.50390625, + "grad_norm": 0.5527358055114746, + "learning_rate": 8.764167762025306e-06, + "loss": 0.0453, + "step": 1161 + }, + { + "epoch": 0.5043402777777778, + "grad_norm": 0.7811183929443359, + "learning_rate": 8.761853304891293e-06, + "loss": 0.0733, + "step": 1162 + }, + { + "epoch": 0.5047743055555556, + "grad_norm": 0.8103577494621277, + "learning_rate": 8.759536988750922e-06, + "loss": 0.0657, + "step": 1163 + }, + { + "epoch": 0.5052083333333334, + "grad_norm": 0.6777807474136353, + "learning_rate": 8.75721881474886e-06, + "loss": 0.0614, + "step": 1164 + }, + { + "epoch": 0.5056423611111112, + "grad_norm": 0.7982746362686157, + "learning_rate": 8.754898784030685e-06, + "loss": 0.0604, + "step": 1165 + }, + { + "epoch": 0.5060763888888888, + "grad_norm": 0.5058833360671997, + "learning_rate": 8.75257689774289e-06, + "loss": 0.0461, + "step": 1166 + }, + { + "epoch": 0.5065104166666666, + "grad_norm": 0.5670070648193359, + "learning_rate": 8.750253157032893e-06, + "loss": 0.0578, + "step": 1167 + }, + { + "epoch": 0.5069444444444444, + "grad_norm": 0.5332913398742676, + "learning_rate": 8.74792756304902e-06, + "loss": 0.0598, + "step": 1168 + }, + { + "epoch": 0.5073784722222222, + "grad_norm": 0.6165770888328552, + "learning_rate": 8.745600116940518e-06, + "loss": 0.0678, + "step": 1169 + }, + { + "epoch": 0.5078125, + "grad_norm": 0.5424283146858215, + "learning_rate": 8.74327081985755e-06, + "loss": 0.0496, + "step": 1170 + }, + { + "epoch": 0.5082465277777778, + "grad_norm": 0.7060388326644897, + "learning_rate": 8.74093967295119e-06, + "loss": 0.0462, + "step": 1171 + }, + { + "epoch": 0.5086805555555556, + "grad_norm": 0.7423508167266846, + "learning_rate": 8.738606677373426e-06, + "loss": 0.0543, + "step": 1172 + }, + { + "epoch": 0.5091145833333334, + "grad_norm": 0.5764522552490234, + "learning_rate": 8.736271834277167e-06, + "loss": 0.0493, + "step": 1173 + }, + { + "epoch": 0.5095486111111112, + "grad_norm": 0.8961094617843628, + "learning_rate": 8.733935144816224e-06, + "loss": 0.0708, + "step": 1174 + }, + { + "epoch": 0.5099826388888888, + "grad_norm": 1.347813367843628, + "learning_rate": 8.731596610145327e-06, + "loss": 0.082, + "step": 1175 + }, + { + "epoch": 0.5104166666666666, + "grad_norm": 0.664887011051178, + "learning_rate": 8.72925623142012e-06, + "loss": 0.0464, + "step": 1176 + }, + { + "epoch": 0.5108506944444444, + "grad_norm": 0.4864520728588104, + "learning_rate": 8.726914009797152e-06, + "loss": 0.0359, + "step": 1177 + }, + { + "epoch": 0.5112847222222222, + "grad_norm": 0.5455866456031799, + "learning_rate": 8.724569946433884e-06, + "loss": 0.0514, + "step": 1178 + }, + { + "epoch": 0.51171875, + "grad_norm": 0.5329997539520264, + "learning_rate": 8.722224042488694e-06, + "loss": 0.0435, + "step": 1179 + }, + { + "epoch": 0.5121527777777778, + "grad_norm": 0.6375029683113098, + "learning_rate": 8.71987629912086e-06, + "loss": 0.0666, + "step": 1180 + }, + { + "epoch": 0.5125868055555556, + "grad_norm": 0.6707751750946045, + "learning_rate": 8.717526717490575e-06, + "loss": 0.0605, + "step": 1181 + }, + { + "epoch": 0.5130208333333334, + "grad_norm": 0.7720732688903809, + "learning_rate": 8.715175298758938e-06, + "loss": 0.0455, + "step": 1182 + }, + { + "epoch": 0.5134548611111112, + "grad_norm": 0.6852914094924927, + "learning_rate": 8.712822044087959e-06, + "loss": 0.0661, + "step": 1183 + }, + { + "epoch": 0.5138888888888888, + "grad_norm": 0.6983592510223389, + "learning_rate": 8.710466954640551e-06, + "loss": 0.0494, + "step": 1184 + }, + { + "epoch": 0.5143229166666666, + "grad_norm": 1.2967052459716797, + "learning_rate": 8.708110031580534e-06, + "loss": 0.0837, + "step": 1185 + }, + { + "epoch": 0.5147569444444444, + "grad_norm": 0.42146065831184387, + "learning_rate": 8.705751276072639e-06, + "loss": 0.0308, + "step": 1186 + }, + { + "epoch": 0.5151909722222222, + "grad_norm": 1.0599000453948975, + "learning_rate": 8.703390689282498e-06, + "loss": 0.0834, + "step": 1187 + }, + { + "epoch": 0.515625, + "grad_norm": 0.8723039031028748, + "learning_rate": 8.70102827237665e-06, + "loss": 0.0641, + "step": 1188 + }, + { + "epoch": 0.5160590277777778, + "grad_norm": 0.8208059668540955, + "learning_rate": 8.698664026522536e-06, + "loss": 0.07, + "step": 1189 + }, + { + "epoch": 0.5164930555555556, + "grad_norm": 0.5073002576828003, + "learning_rate": 8.696297952888503e-06, + "loss": 0.0469, + "step": 1190 + }, + { + "epoch": 0.5169270833333334, + "grad_norm": 0.7411544322967529, + "learning_rate": 8.693930052643803e-06, + "loss": 0.0574, + "step": 1191 + }, + { + "epoch": 0.5173611111111112, + "grad_norm": 0.5245022773742676, + "learning_rate": 8.691560326958584e-06, + "loss": 0.0458, + "step": 1192 + }, + { + "epoch": 0.5177951388888888, + "grad_norm": 0.6204422116279602, + "learning_rate": 8.689188777003906e-06, + "loss": 0.0533, + "step": 1193 + }, + { + "epoch": 0.5182291666666666, + "grad_norm": 0.9399925470352173, + "learning_rate": 8.686815403951719e-06, + "loss": 0.0675, + "step": 1194 + }, + { + "epoch": 0.5186631944444444, + "grad_norm": 0.6003794074058533, + "learning_rate": 8.684440208974884e-06, + "loss": 0.0674, + "step": 1195 + }, + { + "epoch": 0.5190972222222222, + "grad_norm": 0.580943763256073, + "learning_rate": 8.682063193247157e-06, + "loss": 0.045, + "step": 1196 + }, + { + "epoch": 0.51953125, + "grad_norm": 0.5490971803665161, + "learning_rate": 8.679684357943195e-06, + "loss": 0.0505, + "step": 1197 + }, + { + "epoch": 0.5199652777777778, + "grad_norm": 0.6057333946228027, + "learning_rate": 8.677303704238553e-06, + "loss": 0.0453, + "step": 1198 + }, + { + "epoch": 0.5203993055555556, + "grad_norm": 0.5196250081062317, + "learning_rate": 8.674921233309689e-06, + "loss": 0.0402, + "step": 1199 + }, + { + "epoch": 0.5208333333333334, + "grad_norm": 0.6459466814994812, + "learning_rate": 8.672536946333954e-06, + "loss": 0.0427, + "step": 1200 + }, + { + "epoch": 0.5208333333333334, + "eval_layout_rm_rank_test_accuracy": 0.9691039746660712, + "eval_layout_rm_rank_test_chosen_scores": 0.5440658926963806, + "eval_layout_rm_rank_test_loss": 0.052996955811977386, + "eval_layout_rm_rank_test_rejected_scores": 0.07446130365133286, + "eval_layout_rm_rank_test_runtime": 1175.8343, + "eval_layout_rm_rank_test_samples_per_second": 20.948, + "eval_layout_rm_rank_test_steps_per_second": 0.655, + "step": 1200 + }, + { + "epoch": 0.5212673611111112, + "grad_norm": 1.0727156400680542, + "learning_rate": 8.670150844489597e-06, + "loss": 0.0927, + "step": 1201 + }, + { + "epoch": 0.5217013888888888, + "grad_norm": 0.6895102262496948, + "learning_rate": 8.667762928955766e-06, + "loss": 0.0402, + "step": 1202 + }, + { + "epoch": 0.5221354166666666, + "grad_norm": 1.0803653001785278, + "learning_rate": 8.665373200912506e-06, + "loss": 0.0585, + "step": 1203 + }, + { + "epoch": 0.5225694444444444, + "grad_norm": 1.0241613388061523, + "learning_rate": 8.662981661540753e-06, + "loss": 0.0748, + "step": 1204 + }, + { + "epoch": 0.5230034722222222, + "grad_norm": 1.1917078495025635, + "learning_rate": 8.660588312022345e-06, + "loss": 0.0787, + "step": 1205 + }, + { + "epoch": 0.5234375, + "grad_norm": 0.8959947824478149, + "learning_rate": 8.658193153540006e-06, + "loss": 0.0604, + "step": 1206 + }, + { + "epoch": 0.5238715277777778, + "grad_norm": 0.5525012612342834, + "learning_rate": 8.655796187277364e-06, + "loss": 0.0398, + "step": 1207 + }, + { + "epoch": 0.5243055555555556, + "grad_norm": 0.6186550855636597, + "learning_rate": 8.65339741441893e-06, + "loss": 0.0442, + "step": 1208 + }, + { + "epoch": 0.5247395833333334, + "grad_norm": 0.5443890690803528, + "learning_rate": 8.650996836150117e-06, + "loss": 0.0361, + "step": 1209 + }, + { + "epoch": 0.5251736111111112, + "grad_norm": 0.6587841510772705, + "learning_rate": 8.648594453657224e-06, + "loss": 0.0485, + "step": 1210 + }, + { + "epoch": 0.5256076388888888, + "grad_norm": 0.6534464359283447, + "learning_rate": 8.646190268127444e-06, + "loss": 0.0595, + "step": 1211 + }, + { + "epoch": 0.5260416666666666, + "grad_norm": 0.5813679695129395, + "learning_rate": 8.64378428074886e-06, + "loss": 0.0537, + "step": 1212 + }, + { + "epoch": 0.5264756944444444, + "grad_norm": 0.6579888463020325, + "learning_rate": 8.641376492710445e-06, + "loss": 0.0393, + "step": 1213 + }, + { + "epoch": 0.5269097222222222, + "grad_norm": 0.6142035126686096, + "learning_rate": 8.638966905202064e-06, + "loss": 0.0438, + "step": 1214 + }, + { + "epoch": 0.52734375, + "grad_norm": 0.6465262770652771, + "learning_rate": 8.63655551941447e-06, + "loss": 0.0417, + "step": 1215 + }, + { + "epoch": 0.5277777777777778, + "grad_norm": 0.8512789607048035, + "learning_rate": 8.634142336539308e-06, + "loss": 0.0578, + "step": 1216 + }, + { + "epoch": 0.5282118055555556, + "grad_norm": 0.868228554725647, + "learning_rate": 8.631727357769103e-06, + "loss": 0.0677, + "step": 1217 + }, + { + "epoch": 0.5286458333333334, + "grad_norm": 1.382534384727478, + "learning_rate": 8.629310584297274e-06, + "loss": 0.072, + "step": 1218 + }, + { + "epoch": 0.5290798611111112, + "grad_norm": 0.895889937877655, + "learning_rate": 8.626892017318127e-06, + "loss": 0.0639, + "step": 1219 + }, + { + "epoch": 0.5295138888888888, + "grad_norm": 1.282185435295105, + "learning_rate": 8.624471658026848e-06, + "loss": 0.0581, + "step": 1220 + }, + { + "epoch": 0.5299479166666666, + "grad_norm": 0.7685604095458984, + "learning_rate": 8.622049507619519e-06, + "loss": 0.0589, + "step": 1221 + }, + { + "epoch": 0.5303819444444444, + "grad_norm": 0.9641804099082947, + "learning_rate": 8.619625567293096e-06, + "loss": 0.0525, + "step": 1222 + }, + { + "epoch": 0.5308159722222222, + "grad_norm": 0.6586122512817383, + "learning_rate": 8.61719983824543e-06, + "loss": 0.0557, + "step": 1223 + }, + { + "epoch": 0.53125, + "grad_norm": 0.625971257686615, + "learning_rate": 8.614772321675245e-06, + "loss": 0.0574, + "step": 1224 + }, + { + "epoch": 0.5316840277777778, + "grad_norm": 0.6470603346824646, + "learning_rate": 8.61234301878216e-06, + "loss": 0.0594, + "step": 1225 + }, + { + "epoch": 0.5321180555555556, + "grad_norm": 0.600440502166748, + "learning_rate": 8.609911930766668e-06, + "loss": 0.0591, + "step": 1226 + }, + { + "epoch": 0.5325520833333334, + "grad_norm": 0.5906641483306885, + "learning_rate": 8.607479058830144e-06, + "loss": 0.0453, + "step": 1227 + }, + { + "epoch": 0.5329861111111112, + "grad_norm": 0.800161600112915, + "learning_rate": 8.605044404174854e-06, + "loss": 0.0521, + "step": 1228 + }, + { + "epoch": 0.5334201388888888, + "grad_norm": 0.7596476078033447, + "learning_rate": 8.602607968003935e-06, + "loss": 0.0514, + "step": 1229 + }, + { + "epoch": 0.5338541666666666, + "grad_norm": 0.729378342628479, + "learning_rate": 8.60016975152141e-06, + "loss": 0.0489, + "step": 1230 + }, + { + "epoch": 0.5342881944444444, + "grad_norm": 0.6288821697235107, + "learning_rate": 8.597729755932175e-06, + "loss": 0.0494, + "step": 1231 + }, + { + "epoch": 0.5347222222222222, + "grad_norm": 0.841924250125885, + "learning_rate": 8.595287982442016e-06, + "loss": 0.0512, + "step": 1232 + }, + { + "epoch": 0.53515625, + "grad_norm": 0.7113053798675537, + "learning_rate": 8.592844432257588e-06, + "loss": 0.0411, + "step": 1233 + }, + { + "epoch": 0.5355902777777778, + "grad_norm": 0.6260560154914856, + "learning_rate": 8.590399106586429e-06, + "loss": 0.0517, + "step": 1234 + }, + { + "epoch": 0.5360243055555556, + "grad_norm": 0.754421055316925, + "learning_rate": 8.58795200663695e-06, + "loss": 0.0551, + "step": 1235 + }, + { + "epoch": 0.5364583333333334, + "grad_norm": 0.7027459144592285, + "learning_rate": 8.585503133618447e-06, + "loss": 0.058, + "step": 1236 + }, + { + "epoch": 0.5368923611111112, + "grad_norm": 0.7734227180480957, + "learning_rate": 8.583052488741081e-06, + "loss": 0.0474, + "step": 1237 + }, + { + "epoch": 0.5373263888888888, + "grad_norm": 0.6809510588645935, + "learning_rate": 8.5806000732159e-06, + "loss": 0.0562, + "step": 1238 + }, + { + "epoch": 0.5377604166666666, + "grad_norm": 0.7577083110809326, + "learning_rate": 8.578145888254819e-06, + "loss": 0.0789, + "step": 1239 + }, + { + "epoch": 0.5381944444444444, + "grad_norm": 0.45938336849212646, + "learning_rate": 8.575689935070628e-06, + "loss": 0.0381, + "step": 1240 + }, + { + "epoch": 0.5386284722222222, + "grad_norm": 0.43321073055267334, + "learning_rate": 8.573232214876996e-06, + "loss": 0.0473, + "step": 1241 + }, + { + "epoch": 0.5390625, + "grad_norm": 0.5972332954406738, + "learning_rate": 8.57077272888846e-06, + "loss": 0.0629, + "step": 1242 + }, + { + "epoch": 0.5394965277777778, + "grad_norm": 0.5967984199523926, + "learning_rate": 8.568311478320434e-06, + "loss": 0.0446, + "step": 1243 + }, + { + "epoch": 0.5399305555555556, + "grad_norm": 0.5215777158737183, + "learning_rate": 8.565848464389198e-06, + "loss": 0.0404, + "step": 1244 + }, + { + "epoch": 0.5403645833333334, + "grad_norm": 0.5675219297409058, + "learning_rate": 8.56338368831191e-06, + "loss": 0.0415, + "step": 1245 + }, + { + "epoch": 0.5407986111111112, + "grad_norm": 0.8522801399230957, + "learning_rate": 8.560917151306594e-06, + "loss": 0.0583, + "step": 1246 + }, + { + "epoch": 0.5412326388888888, + "grad_norm": 0.4838396906852722, + "learning_rate": 8.558448854592148e-06, + "loss": 0.0371, + "step": 1247 + }, + { + "epoch": 0.5416666666666666, + "grad_norm": 0.5317425727844238, + "learning_rate": 8.555978799388336e-06, + "loss": 0.0357, + "step": 1248 + }, + { + "epoch": 0.5421006944444444, + "grad_norm": 0.6244840025901794, + "learning_rate": 8.553506986915792e-06, + "loss": 0.0507, + "step": 1249 + }, + { + "epoch": 0.5425347222222222, + "grad_norm": 1.0391993522644043, + "learning_rate": 8.551033418396019e-06, + "loss": 0.0585, + "step": 1250 + }, + { + "epoch": 0.54296875, + "grad_norm": 0.9313353300094604, + "learning_rate": 8.548558095051391e-06, + "loss": 0.0814, + "step": 1251 + }, + { + "epoch": 0.5434027777777778, + "grad_norm": 0.835850715637207, + "learning_rate": 8.546081018105142e-06, + "loss": 0.0507, + "step": 1252 + }, + { + "epoch": 0.5438368055555556, + "grad_norm": 0.7506844401359558, + "learning_rate": 8.54360218878138e-06, + "loss": 0.0575, + "step": 1253 + }, + { + "epoch": 0.5442708333333334, + "grad_norm": 1.0784194469451904, + "learning_rate": 8.541121608305072e-06, + "loss": 0.0724, + "step": 1254 + }, + { + "epoch": 0.5447048611111112, + "grad_norm": 0.6933436393737793, + "learning_rate": 8.538639277902057e-06, + "loss": 0.0648, + "step": 1255 + }, + { + "epoch": 0.5451388888888888, + "grad_norm": 0.6776074171066284, + "learning_rate": 8.536155198799034e-06, + "loss": 0.0606, + "step": 1256 + }, + { + "epoch": 0.5455729166666666, + "grad_norm": 0.5607663989067078, + "learning_rate": 8.533669372223568e-06, + "loss": 0.0589, + "step": 1257 + }, + { + "epoch": 0.5460069444444444, + "grad_norm": 0.4452672600746155, + "learning_rate": 8.531181799404088e-06, + "loss": 0.0436, + "step": 1258 + }, + { + "epoch": 0.5464409722222222, + "grad_norm": 0.43998244404792786, + "learning_rate": 8.528692481569884e-06, + "loss": 0.0317, + "step": 1259 + }, + { + "epoch": 0.546875, + "grad_norm": 0.5600937604904175, + "learning_rate": 8.526201419951113e-06, + "loss": 0.0451, + "step": 1260 + }, + { + "epoch": 0.5473090277777778, + "grad_norm": 0.5619065165519714, + "learning_rate": 8.523708615778785e-06, + "loss": 0.0556, + "step": 1261 + }, + { + "epoch": 0.5477430555555556, + "grad_norm": 0.48861682415008545, + "learning_rate": 8.521214070284782e-06, + "loss": 0.0364, + "step": 1262 + }, + { + "epoch": 0.5481770833333334, + "grad_norm": 0.4935634732246399, + "learning_rate": 8.518717784701838e-06, + "loss": 0.0421, + "step": 1263 + }, + { + "epoch": 0.5486111111111112, + "grad_norm": 1.013365387916565, + "learning_rate": 8.51621976026355e-06, + "loss": 0.0546, + "step": 1264 + }, + { + "epoch": 0.5490451388888888, + "grad_norm": 0.9382243156433105, + "learning_rate": 8.513719998204374e-06, + "loss": 0.0687, + "step": 1265 + }, + { + "epoch": 0.5494791666666666, + "grad_norm": 0.8279402852058411, + "learning_rate": 8.511218499759625e-06, + "loss": 0.0526, + "step": 1266 + }, + { + "epoch": 0.5499131944444444, + "grad_norm": 0.8852784633636475, + "learning_rate": 8.508715266165479e-06, + "loss": 0.0669, + "step": 1267 + }, + { + "epoch": 0.5503472222222222, + "grad_norm": 0.8998196125030518, + "learning_rate": 8.506210298658964e-06, + "loss": 0.0599, + "step": 1268 + }, + { + "epoch": 0.55078125, + "grad_norm": 1.1187841892242432, + "learning_rate": 8.503703598477967e-06, + "loss": 0.0596, + "step": 1269 + }, + { + "epoch": 0.5512152777777778, + "grad_norm": 0.6818974018096924, + "learning_rate": 8.501195166861234e-06, + "loss": 0.0552, + "step": 1270 + }, + { + "epoch": 0.5516493055555556, + "grad_norm": 0.8382002711296082, + "learning_rate": 8.49868500504836e-06, + "loss": 0.0564, + "step": 1271 + }, + { + "epoch": 0.5520833333333334, + "grad_norm": 0.8586729764938354, + "learning_rate": 8.496173114279803e-06, + "loss": 0.0607, + "step": 1272 + }, + { + "epoch": 0.5525173611111112, + "grad_norm": 0.6577945947647095, + "learning_rate": 8.493659495796867e-06, + "loss": 0.0464, + "step": 1273 + }, + { + "epoch": 0.5529513888888888, + "grad_norm": 0.869432270526886, + "learning_rate": 8.49114415084172e-06, + "loss": 0.0598, + "step": 1274 + }, + { + "epoch": 0.5533854166666666, + "grad_norm": 0.4027788043022156, + "learning_rate": 8.488627080657375e-06, + "loss": 0.0307, + "step": 1275 + }, + { + "epoch": 0.5538194444444444, + "grad_norm": 0.7500314116477966, + "learning_rate": 8.486108286487698e-06, + "loss": 0.0472, + "step": 1276 + }, + { + "epoch": 0.5542534722222222, + "grad_norm": 0.5049936175346375, + "learning_rate": 8.483587769577414e-06, + "loss": 0.0402, + "step": 1277 + }, + { + "epoch": 0.5546875, + "grad_norm": 0.7624374628067017, + "learning_rate": 8.481065531172091e-06, + "loss": 0.0467, + "step": 1278 + }, + { + "epoch": 0.5551215277777778, + "grad_norm": 0.7847470045089722, + "learning_rate": 8.478541572518152e-06, + "loss": 0.0489, + "step": 1279 + }, + { + "epoch": 0.5555555555555556, + "grad_norm": 0.5217390656471252, + "learning_rate": 8.476015894862867e-06, + "loss": 0.0383, + "step": 1280 + }, + { + "epoch": 0.5559895833333334, + "grad_norm": 0.8662663102149963, + "learning_rate": 8.473488499454363e-06, + "loss": 0.0702, + "step": 1281 + }, + { + "epoch": 0.5564236111111112, + "grad_norm": 0.8891013860702515, + "learning_rate": 8.470959387541606e-06, + "loss": 0.0722, + "step": 1282 + }, + { + "epoch": 0.5568576388888888, + "grad_norm": 0.6959691047668457, + "learning_rate": 8.468428560374416e-06, + "loss": 0.0471, + "step": 1283 + }, + { + "epoch": 0.5572916666666666, + "grad_norm": 0.9065499901771545, + "learning_rate": 8.46589601920346e-06, + "loss": 0.0418, + "step": 1284 + }, + { + "epoch": 0.5577256944444444, + "grad_norm": 0.694658637046814, + "learning_rate": 8.463361765280252e-06, + "loss": 0.0469, + "step": 1285 + }, + { + "epoch": 0.5581597222222222, + "grad_norm": 0.7287715673446655, + "learning_rate": 8.460825799857148e-06, + "loss": 0.0484, + "step": 1286 + }, + { + "epoch": 0.55859375, + "grad_norm": 0.4626723825931549, + "learning_rate": 8.45828812418736e-06, + "loss": 0.0405, + "step": 1287 + }, + { + "epoch": 0.5590277777777778, + "grad_norm": 0.7418968677520752, + "learning_rate": 8.455748739524934e-06, + "loss": 0.0672, + "step": 1288 + }, + { + "epoch": 0.5594618055555556, + "grad_norm": 0.8423126935958862, + "learning_rate": 8.453207647124769e-06, + "loss": 0.0591, + "step": 1289 + }, + { + "epoch": 0.5598958333333334, + "grad_norm": 0.7815362215042114, + "learning_rate": 8.450664848242599e-06, + "loss": 0.0431, + "step": 1290 + }, + { + "epoch": 0.5603298611111112, + "grad_norm": 0.6513561010360718, + "learning_rate": 8.448120344135012e-06, + "loss": 0.0601, + "step": 1291 + }, + { + "epoch": 0.5607638888888888, + "grad_norm": 0.5690886378288269, + "learning_rate": 8.445574136059429e-06, + "loss": 0.0555, + "step": 1292 + }, + { + "epoch": 0.5611979166666666, + "grad_norm": 0.7079150676727295, + "learning_rate": 8.44302622527412e-06, + "loss": 0.0528, + "step": 1293 + }, + { + "epoch": 0.5616319444444444, + "grad_norm": 0.5692477822303772, + "learning_rate": 8.440476613038192e-06, + "loss": 0.0412, + "step": 1294 + }, + { + "epoch": 0.5620659722222222, + "grad_norm": 0.8501294851303101, + "learning_rate": 8.437925300611598e-06, + "loss": 0.052, + "step": 1295 + }, + { + "epoch": 0.5625, + "grad_norm": 0.8660826086997986, + "learning_rate": 8.435372289255123e-06, + "loss": 0.0473, + "step": 1296 + }, + { + "epoch": 0.5629340277777778, + "grad_norm": 0.6774151921272278, + "learning_rate": 8.4328175802304e-06, + "loss": 0.0582, + "step": 1297 + }, + { + "epoch": 0.5633680555555556, + "grad_norm": 0.6546648740768433, + "learning_rate": 8.430261174799897e-06, + "loss": 0.0432, + "step": 1298 + }, + { + "epoch": 0.5638020833333334, + "grad_norm": 0.7087470889091492, + "learning_rate": 8.427703074226916e-06, + "loss": 0.0367, + "step": 1299 + }, + { + "epoch": 0.5642361111111112, + "grad_norm": 0.6567984223365784, + "learning_rate": 8.425143279775607e-06, + "loss": 0.0455, + "step": 1300 + }, + { + "epoch": 0.5646701388888888, + "grad_norm": 1.0748047828674316, + "learning_rate": 8.422581792710948e-06, + "loss": 0.0557, + "step": 1301 + }, + { + "epoch": 0.5651041666666666, + "grad_norm": 0.6882471442222595, + "learning_rate": 8.420018614298758e-06, + "loss": 0.0524, + "step": 1302 + }, + { + "epoch": 0.5655381944444444, + "grad_norm": 0.7942001819610596, + "learning_rate": 8.41745374580569e-06, + "loss": 0.0609, + "step": 1303 + }, + { + "epoch": 0.5659722222222222, + "grad_norm": 1.2504606246948242, + "learning_rate": 8.414887188499233e-06, + "loss": 0.0386, + "step": 1304 + }, + { + "epoch": 0.56640625, + "grad_norm": 0.587224006652832, + "learning_rate": 8.41231894364771e-06, + "loss": 0.0586, + "step": 1305 + }, + { + "epoch": 0.5668402777777778, + "grad_norm": 0.8295255899429321, + "learning_rate": 8.409749012520278e-06, + "loss": 0.048, + "step": 1306 + }, + { + "epoch": 0.5672743055555556, + "grad_norm": 0.7238820791244507, + "learning_rate": 8.40717739638693e-06, + "loss": 0.0789, + "step": 1307 + }, + { + "epoch": 0.5677083333333334, + "grad_norm": 0.5116128325462341, + "learning_rate": 8.404604096518486e-06, + "loss": 0.0454, + "step": 1308 + }, + { + "epoch": 0.5681423611111112, + "grad_norm": 0.4796935021877289, + "learning_rate": 8.402029114186604e-06, + "loss": 0.0498, + "step": 1309 + }, + { + "epoch": 0.5685763888888888, + "grad_norm": 0.638282299041748, + "learning_rate": 8.39945245066377e-06, + "loss": 0.0459, + "step": 1310 + }, + { + "epoch": 0.5690104166666666, + "grad_norm": 0.4643474817276001, + "learning_rate": 8.396874107223298e-06, + "loss": 0.0386, + "step": 1311 + }, + { + "epoch": 0.5694444444444444, + "grad_norm": 0.6372327208518982, + "learning_rate": 8.394294085139343e-06, + "loss": 0.0641, + "step": 1312 + }, + { + "epoch": 0.5698784722222222, + "grad_norm": 0.7297875881195068, + "learning_rate": 8.391712385686877e-06, + "loss": 0.0651, + "step": 1313 + }, + { + "epoch": 0.5703125, + "grad_norm": 0.5666899681091309, + "learning_rate": 8.389129010141705e-06, + "loss": 0.0525, + "step": 1314 + }, + { + "epoch": 0.5707465277777778, + "grad_norm": 0.749977171421051, + "learning_rate": 8.386543959780465e-06, + "loss": 0.0591, + "step": 1315 + }, + { + "epoch": 0.5711805555555556, + "grad_norm": 0.6804627776145935, + "learning_rate": 8.383957235880617e-06, + "loss": 0.0495, + "step": 1316 + }, + { + "epoch": 0.5716145833333334, + "grad_norm": 0.5683985948562622, + "learning_rate": 8.381368839720454e-06, + "loss": 0.047, + "step": 1317 + }, + { + "epoch": 0.5720486111111112, + "grad_norm": 0.7459953427314758, + "learning_rate": 8.378778772579087e-06, + "loss": 0.0377, + "step": 1318 + }, + { + "epoch": 0.5724826388888888, + "grad_norm": 1.290941834449768, + "learning_rate": 8.376187035736458e-06, + "loss": 0.0675, + "step": 1319 + }, + { + "epoch": 0.5729166666666666, + "grad_norm": 0.6719200015068054, + "learning_rate": 8.373593630473334e-06, + "loss": 0.0664, + "step": 1320 + }, + { + "epoch": 0.5733506944444444, + "grad_norm": 0.9134562015533447, + "learning_rate": 8.370998558071308e-06, + "loss": 0.0481, + "step": 1321 + }, + { + "epoch": 0.5737847222222222, + "grad_norm": 0.6994345188140869, + "learning_rate": 8.36840181981279e-06, + "loss": 0.047, + "step": 1322 + }, + { + "epoch": 0.57421875, + "grad_norm": 0.7651628255844116, + "learning_rate": 8.365803416981022e-06, + "loss": 0.0486, + "step": 1323 + }, + { + "epoch": 0.5746527777777778, + "grad_norm": 0.8562982678413391, + "learning_rate": 8.363203350860064e-06, + "loss": 0.0383, + "step": 1324 + }, + { + "epoch": 0.5750868055555556, + "grad_norm": 0.8496001958847046, + "learning_rate": 8.360601622734796e-06, + "loss": 0.0565, + "step": 1325 + }, + { + "epoch": 0.5755208333333334, + "grad_norm": 0.7761862277984619, + "learning_rate": 8.357998233890925e-06, + "loss": 0.075, + "step": 1326 + }, + { + "epoch": 0.5759548611111112, + "grad_norm": 0.5182710289955139, + "learning_rate": 8.355393185614974e-06, + "loss": 0.0417, + "step": 1327 + }, + { + "epoch": 0.5763888888888888, + "grad_norm": 0.45891663432121277, + "learning_rate": 8.352786479194288e-06, + "loss": 0.028, + "step": 1328 + }, + { + "epoch": 0.5768229166666666, + "grad_norm": 0.6198312044143677, + "learning_rate": 8.35017811591703e-06, + "loss": 0.053, + "step": 1329 + }, + { + "epoch": 0.5772569444444444, + "grad_norm": 0.620874285697937, + "learning_rate": 8.347568097072183e-06, + "loss": 0.0599, + "step": 1330 + }, + { + "epoch": 0.5776909722222222, + "grad_norm": 0.6952489018440247, + "learning_rate": 8.344956423949547e-06, + "loss": 0.0645, + "step": 1331 + }, + { + "epoch": 0.578125, + "grad_norm": 0.9362801313400269, + "learning_rate": 8.342343097839743e-06, + "loss": 0.0738, + "step": 1332 + }, + { + "epoch": 0.5785590277777778, + "grad_norm": 0.7076898813247681, + "learning_rate": 8.339728120034204e-06, + "loss": 0.0539, + "step": 1333 + }, + { + "epoch": 0.5789930555555556, + "grad_norm": 0.4984985888004303, + "learning_rate": 8.337111491825183e-06, + "loss": 0.0365, + "step": 1334 + }, + { + "epoch": 0.5794270833333334, + "grad_norm": 0.4663236439228058, + "learning_rate": 8.334493214505746e-06, + "loss": 0.0405, + "step": 1335 + }, + { + "epoch": 0.5798611111111112, + "grad_norm": 0.7300651669502258, + "learning_rate": 8.331873289369774e-06, + "loss": 0.0512, + "step": 1336 + }, + { + "epoch": 0.5802951388888888, + "grad_norm": 0.8619731664657593, + "learning_rate": 8.329251717711964e-06, + "loss": 0.0483, + "step": 1337 + }, + { + "epoch": 0.5807291666666666, + "grad_norm": 1.2407528162002563, + "learning_rate": 8.326628500827826e-06, + "loss": 0.0408, + "step": 1338 + }, + { + "epoch": 0.5811631944444444, + "grad_norm": 0.7284979820251465, + "learning_rate": 8.324003640013685e-06, + "loss": 0.0612, + "step": 1339 + }, + { + "epoch": 0.5815972222222222, + "grad_norm": 0.565274715423584, + "learning_rate": 8.321377136566675e-06, + "loss": 0.0563, + "step": 1340 + }, + { + "epoch": 0.58203125, + "grad_norm": 0.6407968401908875, + "learning_rate": 8.318748991784742e-06, + "loss": 0.048, + "step": 1341 + }, + { + "epoch": 0.5824652777777778, + "grad_norm": 1.3209383487701416, + "learning_rate": 8.316119206966644e-06, + "loss": 0.0977, + "step": 1342 + }, + { + "epoch": 0.5828993055555556, + "grad_norm": 1.2708077430725098, + "learning_rate": 8.313487783411954e-06, + "loss": 0.082, + "step": 1343 + }, + { + "epoch": 0.5833333333333334, + "grad_norm": 0.7060262560844421, + "learning_rate": 8.310854722421044e-06, + "loss": 0.0539, + "step": 1344 + }, + { + "epoch": 0.5837673611111112, + "grad_norm": 0.7067934274673462, + "learning_rate": 8.308220025295107e-06, + "loss": 0.0442, + "step": 1345 + }, + { + "epoch": 0.5842013888888888, + "grad_norm": 0.5657163858413696, + "learning_rate": 8.305583693336135e-06, + "loss": 0.0424, + "step": 1346 + }, + { + "epoch": 0.5846354166666666, + "grad_norm": 0.8705339431762695, + "learning_rate": 8.302945727846934e-06, + "loss": 0.0774, + "step": 1347 + }, + { + "epoch": 0.5850694444444444, + "grad_norm": 0.651258647441864, + "learning_rate": 8.300306130131116e-06, + "loss": 0.048, + "step": 1348 + }, + { + "epoch": 0.5855034722222222, + "grad_norm": 0.7655849456787109, + "learning_rate": 8.297664901493098e-06, + "loss": 0.0577, + "step": 1349 + }, + { + "epoch": 0.5859375, + "grad_norm": 0.5116808414459229, + "learning_rate": 8.295022043238104e-06, + "loss": 0.0353, + "step": 1350 + }, + { + "epoch": 0.5863715277777778, + "grad_norm": 0.5585806965827942, + "learning_rate": 8.292377556672165e-06, + "loss": 0.0502, + "step": 1351 + }, + { + "epoch": 0.5868055555555556, + "grad_norm": 0.6864737272262573, + "learning_rate": 8.289731443102111e-06, + "loss": 0.0511, + "step": 1352 + }, + { + "epoch": 0.5872395833333334, + "grad_norm": 0.9937721490859985, + "learning_rate": 8.287083703835584e-06, + "loss": 0.0469, + "step": 1353 + }, + { + "epoch": 0.5876736111111112, + "grad_norm": 1.1562503576278687, + "learning_rate": 8.284434340181025e-06, + "loss": 0.0751, + "step": 1354 + }, + { + "epoch": 0.5881076388888888, + "grad_norm": 0.8913924098014832, + "learning_rate": 8.281783353447675e-06, + "loss": 0.0599, + "step": 1355 + }, + { + "epoch": 0.5885416666666666, + "grad_norm": 0.8412242531776428, + "learning_rate": 8.27913074494558e-06, + "loss": 0.04, + "step": 1356 + }, + { + "epoch": 0.5889756944444444, + "grad_norm": 0.9167409539222717, + "learning_rate": 8.27647651598559e-06, + "loss": 0.0371, + "step": 1357 + }, + { + "epoch": 0.5894097222222222, + "grad_norm": 1.0268688201904297, + "learning_rate": 8.273820667879351e-06, + "loss": 0.0747, + "step": 1358 + }, + { + "epoch": 0.58984375, + "grad_norm": 0.6086006164550781, + "learning_rate": 8.271163201939316e-06, + "loss": 0.0311, + "step": 1359 + }, + { + "epoch": 0.5902777777777778, + "grad_norm": 0.5621817708015442, + "learning_rate": 8.268504119478729e-06, + "loss": 0.047, + "step": 1360 + }, + { + "epoch": 0.5907118055555556, + "grad_norm": 0.8023841381072998, + "learning_rate": 8.265843421811636e-06, + "loss": 0.0553, + "step": 1361 + }, + { + "epoch": 0.5911458333333334, + "grad_norm": 0.9353199601173401, + "learning_rate": 8.263181110252885e-06, + "loss": 0.0641, + "step": 1362 + }, + { + "epoch": 0.5915798611111112, + "grad_norm": 0.6181360483169556, + "learning_rate": 8.260517186118116e-06, + "loss": 0.0639, + "step": 1363 + }, + { + "epoch": 0.5920138888888888, + "grad_norm": 0.5545939803123474, + "learning_rate": 8.25785165072377e-06, + "loss": 0.0585, + "step": 1364 + }, + { + "epoch": 0.5924479166666666, + "grad_norm": 0.6505069136619568, + "learning_rate": 8.255184505387082e-06, + "loss": 0.0544, + "step": 1365 + }, + { + "epoch": 0.5928819444444444, + "grad_norm": 0.4361746907234192, + "learning_rate": 8.252515751426082e-06, + "loss": 0.03, + "step": 1366 + }, + { + "epoch": 0.5933159722222222, + "grad_norm": 0.5293030142784119, + "learning_rate": 8.249845390159598e-06, + "loss": 0.0559, + "step": 1367 + }, + { + "epoch": 0.59375, + "grad_norm": 0.5465980172157288, + "learning_rate": 8.24717342290725e-06, + "loss": 0.0444, + "step": 1368 + }, + { + "epoch": 0.5941840277777778, + "grad_norm": 0.6178024411201477, + "learning_rate": 8.244499850989453e-06, + "loss": 0.0571, + "step": 1369 + }, + { + "epoch": 0.5946180555555556, + "grad_norm": 0.5191216468811035, + "learning_rate": 8.241824675727412e-06, + "loss": 0.0347, + "step": 1370 + }, + { + "epoch": 0.5950520833333334, + "grad_norm": 0.5588525533676147, + "learning_rate": 8.239147898443126e-06, + "loss": 0.0491, + "step": 1371 + }, + { + "epoch": 0.5954861111111112, + "grad_norm": 0.9492883682250977, + "learning_rate": 8.236469520459391e-06, + "loss": 0.0513, + "step": 1372 + }, + { + "epoch": 0.5959201388888888, + "grad_norm": 1.0945905447006226, + "learning_rate": 8.233789543099784e-06, + "loss": 0.0533, + "step": 1373 + }, + { + "epoch": 0.5963541666666666, + "grad_norm": 0.7427625060081482, + "learning_rate": 8.231107967688678e-06, + "loss": 0.0579, + "step": 1374 + }, + { + "epoch": 0.5967881944444444, + "grad_norm": 0.7160750031471252, + "learning_rate": 8.228424795551238e-06, + "loss": 0.045, + "step": 1375 + }, + { + "epoch": 0.5972222222222222, + "grad_norm": 0.7438015937805176, + "learning_rate": 8.225740028013413e-06, + "loss": 0.0457, + "step": 1376 + }, + { + "epoch": 0.59765625, + "grad_norm": 1.2287431955337524, + "learning_rate": 8.223053666401942e-06, + "loss": 0.0553, + "step": 1377 + }, + { + "epoch": 0.5980902777777778, + "grad_norm": 0.8054426908493042, + "learning_rate": 8.220365712044353e-06, + "loss": 0.0453, + "step": 1378 + }, + { + "epoch": 0.5985243055555556, + "grad_norm": 0.6114641427993774, + "learning_rate": 8.217676166268965e-06, + "loss": 0.0504, + "step": 1379 + }, + { + "epoch": 0.5989583333333334, + "grad_norm": 0.6696768403053284, + "learning_rate": 8.214985030404872e-06, + "loss": 0.0514, + "step": 1380 + }, + { + "epoch": 0.5993923611111112, + "grad_norm": 0.7519408464431763, + "learning_rate": 8.212292305781964e-06, + "loss": 0.0686, + "step": 1381 + }, + { + "epoch": 0.5998263888888888, + "grad_norm": 0.48269274830818176, + "learning_rate": 8.20959799373091e-06, + "loss": 0.0331, + "step": 1382 + }, + { + "epoch": 0.6002604166666666, + "grad_norm": 0.7272772192955017, + "learning_rate": 8.206902095583174e-06, + "loss": 0.0518, + "step": 1383 + }, + { + "epoch": 0.6006944444444444, + "grad_norm": 0.5394783020019531, + "learning_rate": 8.204204612670987e-06, + "loss": 0.0481, + "step": 1384 + }, + { + "epoch": 0.6011284722222222, + "grad_norm": 0.5102927088737488, + "learning_rate": 8.201505546327377e-06, + "loss": 0.0378, + "step": 1385 + }, + { + "epoch": 0.6015625, + "grad_norm": 0.6936540007591248, + "learning_rate": 8.19880489788615e-06, + "loss": 0.0534, + "step": 1386 + }, + { + "epoch": 0.6019965277777778, + "grad_norm": 0.6128141283988953, + "learning_rate": 8.196102668681891e-06, + "loss": 0.0477, + "step": 1387 + }, + { + "epoch": 0.6024305555555556, + "grad_norm": 0.7972360849380493, + "learning_rate": 8.19339886004997e-06, + "loss": 0.0593, + "step": 1388 + }, + { + "epoch": 0.6028645833333334, + "grad_norm": 0.8966538906097412, + "learning_rate": 8.190693473326536e-06, + "loss": 0.0715, + "step": 1389 + }, + { + "epoch": 0.6032986111111112, + "grad_norm": 0.6679428815841675, + "learning_rate": 8.18798650984852e-06, + "loss": 0.0416, + "step": 1390 + }, + { + "epoch": 0.6037326388888888, + "grad_norm": 0.45375049114227295, + "learning_rate": 8.185277970953627e-06, + "loss": 0.0371, + "step": 1391 + }, + { + "epoch": 0.6041666666666666, + "grad_norm": 0.8478909134864807, + "learning_rate": 8.182567857980347e-06, + "loss": 0.0746, + "step": 1392 + }, + { + "epoch": 0.6046006944444444, + "grad_norm": 1.3403058052062988, + "learning_rate": 8.179856172267942e-06, + "loss": 0.064, + "step": 1393 + }, + { + "epoch": 0.6050347222222222, + "grad_norm": 0.5338839292526245, + "learning_rate": 8.177142915156455e-06, + "loss": 0.0311, + "step": 1394 + }, + { + "epoch": 0.60546875, + "grad_norm": 0.8129806518554688, + "learning_rate": 8.174428087986705e-06, + "loss": 0.0701, + "step": 1395 + }, + { + "epoch": 0.6059027777777778, + "grad_norm": 0.76915442943573, + "learning_rate": 8.171711692100286e-06, + "loss": 0.0345, + "step": 1396 + }, + { + "epoch": 0.6063368055555556, + "grad_norm": 0.6448995471000671, + "learning_rate": 8.168993728839567e-06, + "loss": 0.0431, + "step": 1397 + }, + { + "epoch": 0.6067708333333334, + "grad_norm": 0.5470688343048096, + "learning_rate": 8.166274199547692e-06, + "loss": 0.034, + "step": 1398 + }, + { + "epoch": 0.6072048611111112, + "grad_norm": 1.3783670663833618, + "learning_rate": 8.16355310556858e-06, + "loss": 0.0559, + "step": 1399 + }, + { + "epoch": 0.6076388888888888, + "grad_norm": 0.6607531905174255, + "learning_rate": 8.160830448246922e-06, + "loss": 0.0434, + "step": 1400 + }, + { + "epoch": 0.6080729166666666, + "grad_norm": 1.23941969871521, + "learning_rate": 8.15810622892818e-06, + "loss": 0.0739, + "step": 1401 + }, + { + "epoch": 0.6085069444444444, + "grad_norm": 0.6775326728820801, + "learning_rate": 8.155380448958589e-06, + "loss": 0.0522, + "step": 1402 + }, + { + "epoch": 0.6089409722222222, + "grad_norm": 0.5448042750358582, + "learning_rate": 8.152653109685163e-06, + "loss": 0.0347, + "step": 1403 + }, + { + "epoch": 0.609375, + "grad_norm": 0.6932007670402527, + "learning_rate": 8.14992421245567e-06, + "loss": 0.0492, + "step": 1404 + }, + { + "epoch": 0.6098090277777778, + "grad_norm": 0.7181015014648438, + "learning_rate": 8.147193758618664e-06, + "loss": 0.0485, + "step": 1405 + }, + { + "epoch": 0.6102430555555556, + "grad_norm": 0.5694456696510315, + "learning_rate": 8.144461749523458e-06, + "loss": 0.0484, + "step": 1406 + }, + { + "epoch": 0.6106770833333334, + "grad_norm": 0.649516224861145, + "learning_rate": 8.141728186520138e-06, + "loss": 0.048, + "step": 1407 + }, + { + "epoch": 0.6111111111111112, + "grad_norm": 0.6933144330978394, + "learning_rate": 8.138993070959557e-06, + "loss": 0.0624, + "step": 1408 + }, + { + "epoch": 0.6115451388888888, + "grad_norm": 1.0121896266937256, + "learning_rate": 8.136256404193336e-06, + "loss": 0.0527, + "step": 1409 + }, + { + "epoch": 0.6119791666666666, + "grad_norm": 0.766247034072876, + "learning_rate": 8.133518187573864e-06, + "loss": 0.0443, + "step": 1410 + }, + { + "epoch": 0.6124131944444444, + "grad_norm": 0.8101610541343689, + "learning_rate": 8.13077842245429e-06, + "loss": 0.0502, + "step": 1411 + }, + { + "epoch": 0.6128472222222222, + "grad_norm": 0.47420060634613037, + "learning_rate": 8.128037110188531e-06, + "loss": 0.0233, + "step": 1412 + }, + { + "epoch": 0.61328125, + "grad_norm": 1.5402026176452637, + "learning_rate": 8.125294252131273e-06, + "loss": 0.0785, + "step": 1413 + }, + { + "epoch": 0.6137152777777778, + "grad_norm": 0.6392644047737122, + "learning_rate": 8.122549849637961e-06, + "loss": 0.0357, + "step": 1414 + }, + { + "epoch": 0.6141493055555556, + "grad_norm": 0.7949815392494202, + "learning_rate": 8.119803904064805e-06, + "loss": 0.0415, + "step": 1415 + }, + { + "epoch": 0.6145833333333334, + "grad_norm": 1.336789608001709, + "learning_rate": 8.117056416768776e-06, + "loss": 0.0581, + "step": 1416 + }, + { + "epoch": 0.6150173611111112, + "grad_norm": 0.9395679831504822, + "learning_rate": 8.11430738910761e-06, + "loss": 0.0379, + "step": 1417 + }, + { + "epoch": 0.6154513888888888, + "grad_norm": 1.2471702098846436, + "learning_rate": 8.111556822439801e-06, + "loss": 0.0507, + "step": 1418 + }, + { + "epoch": 0.6158854166666666, + "grad_norm": 0.8362313508987427, + "learning_rate": 8.108804718124606e-06, + "loss": 0.0548, + "step": 1419 + }, + { + "epoch": 0.6163194444444444, + "grad_norm": 0.6505581736564636, + "learning_rate": 8.106051077522037e-06, + "loss": 0.0469, + "step": 1420 + }, + { + "epoch": 0.6167534722222222, + "grad_norm": 0.6922495365142822, + "learning_rate": 8.103295901992874e-06, + "loss": 0.0505, + "step": 1421 + }, + { + "epoch": 0.6171875, + "grad_norm": 0.8996455669403076, + "learning_rate": 8.100539192898648e-06, + "loss": 0.05, + "step": 1422 + }, + { + "epoch": 0.6176215277777778, + "grad_norm": 0.46958956122398376, + "learning_rate": 8.097780951601647e-06, + "loss": 0.0282, + "step": 1423 + }, + { + "epoch": 0.6180555555555556, + "grad_norm": 0.9579825401306152, + "learning_rate": 8.095021179464926e-06, + "loss": 0.0497, + "step": 1424 + }, + { + "epoch": 0.6184895833333334, + "grad_norm": 0.49152427911758423, + "learning_rate": 8.092259877852287e-06, + "loss": 0.0315, + "step": 1425 + }, + { + "epoch": 0.6189236111111112, + "grad_norm": 1.2000850439071655, + "learning_rate": 8.089497048128288e-06, + "loss": 0.0611, + "step": 1426 + }, + { + "epoch": 0.6193576388888888, + "grad_norm": 1.0212475061416626, + "learning_rate": 8.086732691658247e-06, + "loss": 0.0651, + "step": 1427 + }, + { + "epoch": 0.6197916666666666, + "grad_norm": 0.9792412519454956, + "learning_rate": 8.083966809808234e-06, + "loss": 0.0486, + "step": 1428 + }, + { + "epoch": 0.6202256944444444, + "grad_norm": 0.9700614213943481, + "learning_rate": 8.081199403945075e-06, + "loss": 0.0617, + "step": 1429 + }, + { + "epoch": 0.6206597222222222, + "grad_norm": 0.6700954437255859, + "learning_rate": 8.078430475436343e-06, + "loss": 0.0431, + "step": 1430 + }, + { + "epoch": 0.62109375, + "grad_norm": 0.9820812344551086, + "learning_rate": 8.075660025650374e-06, + "loss": 0.0478, + "step": 1431 + }, + { + "epoch": 0.6215277777777778, + "grad_norm": 0.6720173954963684, + "learning_rate": 8.072888055956246e-06, + "loss": 0.0436, + "step": 1432 + }, + { + "epoch": 0.6219618055555556, + "grad_norm": 0.5771016478538513, + "learning_rate": 8.070114567723788e-06, + "loss": 0.0377, + "step": 1433 + }, + { + "epoch": 0.6223958333333334, + "grad_norm": 0.7792287468910217, + "learning_rate": 8.06733956232359e-06, + "loss": 0.0638, + "step": 1434 + }, + { + "epoch": 0.6228298611111112, + "grad_norm": 0.9756418466567993, + "learning_rate": 8.064563041126981e-06, + "loss": 0.0557, + "step": 1435 + }, + { + "epoch": 0.6232638888888888, + "grad_norm": 0.5423485040664673, + "learning_rate": 8.06178500550604e-06, + "loss": 0.0351, + "step": 1436 + }, + { + "epoch": 0.6236979166666666, + "grad_norm": 0.6620336771011353, + "learning_rate": 8.059005456833604e-06, + "loss": 0.0478, + "step": 1437 + }, + { + "epoch": 0.6241319444444444, + "grad_norm": 0.6537438631057739, + "learning_rate": 8.056224396483246e-06, + "loss": 0.0382, + "step": 1438 + }, + { + "epoch": 0.6245659722222222, + "grad_norm": 0.6597551107406616, + "learning_rate": 8.05344182582929e-06, + "loss": 0.0465, + "step": 1439 + }, + { + "epoch": 0.625, + "grad_norm": 1.3213461637496948, + "learning_rate": 8.050657746246812e-06, + "loss": 0.0539, + "step": 1440 + }, + { + "epoch": 0.6254340277777778, + "grad_norm": 0.6487838625907898, + "learning_rate": 8.047872159111627e-06, + "loss": 0.0393, + "step": 1441 + }, + { + "epoch": 0.6258680555555556, + "grad_norm": 1.4973907470703125, + "learning_rate": 8.045085065800297e-06, + "loss": 0.0764, + "step": 1442 + }, + { + "epoch": 0.6263020833333334, + "grad_norm": 0.8121576905250549, + "learning_rate": 8.042296467690126e-06, + "loss": 0.0571, + "step": 1443 + }, + { + "epoch": 0.6267361111111112, + "grad_norm": 0.5702161192893982, + "learning_rate": 8.039506366159166e-06, + "loss": 0.0279, + "step": 1444 + }, + { + "epoch": 0.6271701388888888, + "grad_norm": 0.5276521444320679, + "learning_rate": 8.03671476258621e-06, + "loss": 0.0328, + "step": 1445 + }, + { + "epoch": 0.6276041666666666, + "grad_norm": 0.8054888844490051, + "learning_rate": 8.033921658350793e-06, + "loss": 0.0596, + "step": 1446 + }, + { + "epoch": 0.6280381944444444, + "grad_norm": 0.9613228440284729, + "learning_rate": 8.031127054833192e-06, + "loss": 0.0617, + "step": 1447 + }, + { + "epoch": 0.6284722222222222, + "grad_norm": 1.0140703916549683, + "learning_rate": 8.028330953414423e-06, + "loss": 0.0744, + "step": 1448 + }, + { + "epoch": 0.62890625, + "grad_norm": 0.652394711971283, + "learning_rate": 8.025533355476243e-06, + "loss": 0.037, + "step": 1449 + }, + { + "epoch": 0.6293402777777778, + "grad_norm": 0.636411726474762, + "learning_rate": 8.022734262401154e-06, + "loss": 0.0488, + "step": 1450 + }, + { + "epoch": 0.6297743055555556, + "grad_norm": 0.6581993103027344, + "learning_rate": 8.019933675572389e-06, + "loss": 0.0488, + "step": 1451 + }, + { + "epoch": 0.6302083333333334, + "grad_norm": 0.6974286437034607, + "learning_rate": 8.017131596373922e-06, + "loss": 0.0555, + "step": 1452 + }, + { + "epoch": 0.6306423611111112, + "grad_norm": 0.5334892868995667, + "learning_rate": 8.014328026190466e-06, + "loss": 0.0583, + "step": 1453 + }, + { + "epoch": 0.6310763888888888, + "grad_norm": 0.6498401761054993, + "learning_rate": 8.01152296640747e-06, + "loss": 0.0625, + "step": 1454 + }, + { + "epoch": 0.6315104166666666, + "grad_norm": 0.6348752379417419, + "learning_rate": 8.008716418411117e-06, + "loss": 0.0551, + "step": 1455 + }, + { + "epoch": 0.6319444444444444, + "grad_norm": 0.5549450516700745, + "learning_rate": 8.005908383588332e-06, + "loss": 0.048, + "step": 1456 + }, + { + "epoch": 0.6323784722222222, + "grad_norm": 0.5023066997528076, + "learning_rate": 8.003098863326764e-06, + "loss": 0.0485, + "step": 1457 + }, + { + "epoch": 0.6328125, + "grad_norm": 0.5993810892105103, + "learning_rate": 8.000287859014804e-06, + "loss": 0.0727, + "step": 1458 + }, + { + "epoch": 0.6332465277777778, + "grad_norm": 0.6572069525718689, + "learning_rate": 7.997475372041576e-06, + "loss": 0.0471, + "step": 1459 + }, + { + "epoch": 0.6336805555555556, + "grad_norm": 0.5218420028686523, + "learning_rate": 7.994661403796935e-06, + "loss": 0.0416, + "step": 1460 + }, + { + "epoch": 0.6341145833333334, + "grad_norm": 0.7828699946403503, + "learning_rate": 7.991845955671467e-06, + "loss": 0.0725, + "step": 1461 + }, + { + "epoch": 0.6345486111111112, + "grad_norm": 0.5772749781608582, + "learning_rate": 7.98902902905649e-06, + "loss": 0.0349, + "step": 1462 + }, + { + "epoch": 0.6349826388888888, + "grad_norm": 0.480615496635437, + "learning_rate": 7.986210625344054e-06, + "loss": 0.0505, + "step": 1463 + }, + { + "epoch": 0.6354166666666666, + "grad_norm": 0.6188299059867859, + "learning_rate": 7.98339074592694e-06, + "loss": 0.0374, + "step": 1464 + }, + { + "epoch": 0.6358506944444444, + "grad_norm": 0.5962648391723633, + "learning_rate": 7.980569392198651e-06, + "loss": 0.0526, + "step": 1465 + }, + { + "epoch": 0.6362847222222222, + "grad_norm": 0.9205628037452698, + "learning_rate": 7.97774656555343e-06, + "loss": 0.078, + "step": 1466 + }, + { + "epoch": 0.63671875, + "grad_norm": 0.6367231011390686, + "learning_rate": 7.974922267386239e-06, + "loss": 0.0384, + "step": 1467 + }, + { + "epoch": 0.6371527777777778, + "grad_norm": 0.9891589879989624, + "learning_rate": 7.972096499092767e-06, + "loss": 0.0694, + "step": 1468 + }, + { + "epoch": 0.6375868055555556, + "grad_norm": 0.7711750268936157, + "learning_rate": 7.969269262069438e-06, + "loss": 0.0552, + "step": 1469 + }, + { + "epoch": 0.6380208333333334, + "grad_norm": 0.623945951461792, + "learning_rate": 7.96644055771339e-06, + "loss": 0.0542, + "step": 1470 + }, + { + "epoch": 0.6384548611111112, + "grad_norm": 0.6759617328643799, + "learning_rate": 7.963610387422499e-06, + "loss": 0.0433, + "step": 1471 + }, + { + "epoch": 0.6388888888888888, + "grad_norm": 0.6105266809463501, + "learning_rate": 7.960778752595355e-06, + "loss": 0.0407, + "step": 1472 + }, + { + "epoch": 0.6393229166666666, + "grad_norm": 0.7171297669410706, + "learning_rate": 7.957945654631275e-06, + "loss": 0.0447, + "step": 1473 + }, + { + "epoch": 0.6397569444444444, + "grad_norm": 0.821792483329773, + "learning_rate": 7.9551110949303e-06, + "loss": 0.0581, + "step": 1474 + }, + { + "epoch": 0.6401909722222222, + "grad_norm": 0.6385071277618408, + "learning_rate": 7.952275074893195e-06, + "loss": 0.0569, + "step": 1475 + }, + { + "epoch": 0.640625, + "grad_norm": 0.8872828483581543, + "learning_rate": 7.949437595921443e-06, + "loss": 0.0658, + "step": 1476 + }, + { + "epoch": 0.6410590277777778, + "grad_norm": 0.6728609204292297, + "learning_rate": 7.946598659417249e-06, + "loss": 0.0396, + "step": 1477 + }, + { + "epoch": 0.6414930555555556, + "grad_norm": 0.6835218071937561, + "learning_rate": 7.943758266783539e-06, + "loss": 0.0656, + "step": 1478 + }, + { + "epoch": 0.6419270833333334, + "grad_norm": 0.6220277547836304, + "learning_rate": 7.940916419423959e-06, + "loss": 0.048, + "step": 1479 + }, + { + "epoch": 0.6423611111111112, + "grad_norm": 0.46455779671669006, + "learning_rate": 7.938073118742871e-06, + "loss": 0.0382, + "step": 1480 + }, + { + "epoch": 0.6427951388888888, + "grad_norm": 0.6037314534187317, + "learning_rate": 7.93522836614536e-06, + "loss": 0.0402, + "step": 1481 + }, + { + "epoch": 0.6432291666666666, + "grad_norm": 0.8157238364219666, + "learning_rate": 7.932382163037227e-06, + "loss": 0.0733, + "step": 1482 + }, + { + "epoch": 0.6436631944444444, + "grad_norm": 1.062686800956726, + "learning_rate": 7.929534510824985e-06, + "loss": 0.067, + "step": 1483 + }, + { + "epoch": 0.6440972222222222, + "grad_norm": 0.7600874900817871, + "learning_rate": 7.926685410915871e-06, + "loss": 0.0513, + "step": 1484 + }, + { + "epoch": 0.64453125, + "grad_norm": 0.9816238880157471, + "learning_rate": 7.923834864717828e-06, + "loss": 0.0582, + "step": 1485 + }, + { + "epoch": 0.6449652777777778, + "grad_norm": 0.4806270897388458, + "learning_rate": 7.920982873639524e-06, + "loss": 0.0378, + "step": 1486 + }, + { + "epoch": 0.6453993055555556, + "grad_norm": 0.7819468975067139, + "learning_rate": 7.918129439090335e-06, + "loss": 0.0562, + "step": 1487 + }, + { + "epoch": 0.6458333333333334, + "grad_norm": 0.6223804354667664, + "learning_rate": 7.915274562480349e-06, + "loss": 0.0492, + "step": 1488 + }, + { + "epoch": 0.6462673611111112, + "grad_norm": 0.6152021884918213, + "learning_rate": 7.912418245220372e-06, + "loss": 0.0537, + "step": 1489 + }, + { + "epoch": 0.6467013888888888, + "grad_norm": 0.695647656917572, + "learning_rate": 7.909560488721915e-06, + "loss": 0.0484, + "step": 1490 + }, + { + "epoch": 0.6471354166666666, + "grad_norm": 0.6031627655029297, + "learning_rate": 7.906701294397208e-06, + "loss": 0.0543, + "step": 1491 + }, + { + "epoch": 0.6475694444444444, + "grad_norm": 0.5799106955528259, + "learning_rate": 7.903840663659186e-06, + "loss": 0.0505, + "step": 1492 + }, + { + "epoch": 0.6480034722222222, + "grad_norm": 1.1803741455078125, + "learning_rate": 7.900978597921494e-06, + "loss": 0.0398, + "step": 1493 + }, + { + "epoch": 0.6484375, + "grad_norm": 0.9330025315284729, + "learning_rate": 7.898115098598492e-06, + "loss": 0.0815, + "step": 1494 + }, + { + "epoch": 0.6488715277777778, + "grad_norm": 0.6871426701545715, + "learning_rate": 7.895250167105239e-06, + "loss": 0.0501, + "step": 1495 + }, + { + "epoch": 0.6493055555555556, + "grad_norm": 0.7065111994743347, + "learning_rate": 7.89238380485751e-06, + "loss": 0.0519, + "step": 1496 + }, + { + "epoch": 0.6497395833333334, + "grad_norm": 0.7787265777587891, + "learning_rate": 7.889516013271782e-06, + "loss": 0.0652, + "step": 1497 + }, + { + "epoch": 0.6501736111111112, + "grad_norm": 0.5526673197746277, + "learning_rate": 7.88664679376524e-06, + "loss": 0.0435, + "step": 1498 + }, + { + "epoch": 0.6506076388888888, + "grad_norm": 0.7988463044166565, + "learning_rate": 7.883776147755777e-06, + "loss": 0.0602, + "step": 1499 + }, + { + "epoch": 0.6510416666666666, + "grad_norm": 0.6189616918563843, + "learning_rate": 7.880904076661988e-06, + "loss": 0.048, + "step": 1500 + }, + { + "epoch": 0.6510416666666666, + "eval_layout_rm_rank_test_accuracy": 0.9717835248264383, + "eval_layout_rm_rank_test_chosen_scores": 0.5814792513847351, + "eval_layout_rm_rank_test_loss": 0.05069770663976669, + "eval_layout_rm_rank_test_rejected_scores": 0.08338902145624161, + "eval_layout_rm_rank_test_runtime": 1178.5822, + "eval_layout_rm_rank_test_samples_per_second": 20.899, + "eval_layout_rm_rank_test_steps_per_second": 0.653, + "step": 1500 + }, + { + "epoch": 0.6514756944444444, + "grad_norm": 0.562228798866272, + "learning_rate": 7.878030581903171e-06, + "loss": 0.0367, + "step": 1501 + }, + { + "epoch": 0.6519097222222222, + "grad_norm": 0.6598196625709534, + "learning_rate": 7.875155664899334e-06, + "loss": 0.0547, + "step": 1502 + }, + { + "epoch": 0.65234375, + "grad_norm": 0.7484297752380371, + "learning_rate": 7.87227932707118e-06, + "loss": 0.0445, + "step": 1503 + }, + { + "epoch": 0.6527777777777778, + "grad_norm": 0.820593535900116, + "learning_rate": 7.869401569840117e-06, + "loss": 0.0469, + "step": 1504 + }, + { + "epoch": 0.6532118055555556, + "grad_norm": 1.0721542835235596, + "learning_rate": 7.866522394628256e-06, + "loss": 0.0588, + "step": 1505 + }, + { + "epoch": 0.6536458333333334, + "grad_norm": 0.735040009021759, + "learning_rate": 7.863641802858408e-06, + "loss": 0.0462, + "step": 1506 + }, + { + "epoch": 0.6540798611111112, + "grad_norm": 0.8309510946273804, + "learning_rate": 7.860759795954085e-06, + "loss": 0.0512, + "step": 1507 + }, + { + "epoch": 0.6545138888888888, + "grad_norm": 0.6727327704429626, + "learning_rate": 7.857876375339493e-06, + "loss": 0.051, + "step": 1508 + }, + { + "epoch": 0.6549479166666666, + "grad_norm": 0.9708459377288818, + "learning_rate": 7.854991542439544e-06, + "loss": 0.0706, + "step": 1509 + }, + { + "epoch": 0.6553819444444444, + "grad_norm": 0.7068577408790588, + "learning_rate": 7.852105298679844e-06, + "loss": 0.0523, + "step": 1510 + }, + { + "epoch": 0.6558159722222222, + "grad_norm": 0.5113874077796936, + "learning_rate": 7.849217645486695e-06, + "loss": 0.0292, + "step": 1511 + }, + { + "epoch": 0.65625, + "grad_norm": 0.5951700806617737, + "learning_rate": 7.8463285842871e-06, + "loss": 0.0374, + "step": 1512 + }, + { + "epoch": 0.6566840277777778, + "grad_norm": 1.22222101688385, + "learning_rate": 7.843438116508752e-06, + "loss": 0.0734, + "step": 1513 + }, + { + "epoch": 0.6571180555555556, + "grad_norm": 0.6699300408363342, + "learning_rate": 7.840546243580045e-06, + "loss": 0.0442, + "step": 1514 + }, + { + "epoch": 0.6575520833333334, + "grad_norm": 0.5567951202392578, + "learning_rate": 7.837652966930064e-06, + "loss": 0.0396, + "step": 1515 + }, + { + "epoch": 0.6579861111111112, + "grad_norm": 0.7641263008117676, + "learning_rate": 7.834758287988586e-06, + "loss": 0.0441, + "step": 1516 + }, + { + "epoch": 0.6584201388888888, + "grad_norm": 0.7480666041374207, + "learning_rate": 7.831862208186085e-06, + "loss": 0.0463, + "step": 1517 + }, + { + "epoch": 0.6588541666666666, + "grad_norm": 0.8478124141693115, + "learning_rate": 7.828964728953728e-06, + "loss": 0.051, + "step": 1518 + }, + { + "epoch": 0.6592881944444444, + "grad_norm": 0.7010301947593689, + "learning_rate": 7.826065851723366e-06, + "loss": 0.0378, + "step": 1519 + }, + { + "epoch": 0.6597222222222222, + "grad_norm": 0.7114124894142151, + "learning_rate": 7.823165577927547e-06, + "loss": 0.0587, + "step": 1520 + }, + { + "epoch": 0.66015625, + "grad_norm": 0.6724095940589905, + "learning_rate": 7.820263908999508e-06, + "loss": 0.0541, + "step": 1521 + }, + { + "epoch": 0.6605902777777778, + "grad_norm": 0.6992164254188538, + "learning_rate": 7.81736084637318e-06, + "loss": 0.0546, + "step": 1522 + }, + { + "epoch": 0.6610243055555556, + "grad_norm": 0.5857810974121094, + "learning_rate": 7.814456391483174e-06, + "loss": 0.0522, + "step": 1523 + }, + { + "epoch": 0.6614583333333334, + "grad_norm": 0.6716225147247314, + "learning_rate": 7.811550545764792e-06, + "loss": 0.0477, + "step": 1524 + }, + { + "epoch": 0.6618923611111112, + "grad_norm": 0.6821210980415344, + "learning_rate": 7.808643310654029e-06, + "loss": 0.0423, + "step": 1525 + }, + { + "epoch": 0.6623263888888888, + "grad_norm": 0.597215473651886, + "learning_rate": 7.805734687587559e-06, + "loss": 0.0376, + "step": 1526 + }, + { + "epoch": 0.6627604166666666, + "grad_norm": 0.8344597816467285, + "learning_rate": 7.802824678002744e-06, + "loss": 0.0553, + "step": 1527 + }, + { + "epoch": 0.6631944444444444, + "grad_norm": 0.7426319718360901, + "learning_rate": 7.799913283337634e-06, + "loss": 0.0537, + "step": 1528 + }, + { + "epoch": 0.6636284722222222, + "grad_norm": 0.7442464828491211, + "learning_rate": 7.797000505030962e-06, + "loss": 0.0667, + "step": 1529 + }, + { + "epoch": 0.6640625, + "grad_norm": 0.6925808191299438, + "learning_rate": 7.794086344522144e-06, + "loss": 0.039, + "step": 1530 + }, + { + "epoch": 0.6644965277777778, + "grad_norm": 0.8789548277854919, + "learning_rate": 7.79117080325128e-06, + "loss": 0.0631, + "step": 1531 + }, + { + "epoch": 0.6649305555555556, + "grad_norm": 0.6922028660774231, + "learning_rate": 7.78825388265915e-06, + "loss": 0.0556, + "step": 1532 + }, + { + "epoch": 0.6653645833333334, + "grad_norm": 0.6471952795982361, + "learning_rate": 7.78533558418722e-06, + "loss": 0.0449, + "step": 1533 + }, + { + "epoch": 0.6657986111111112, + "grad_norm": 0.6730069518089294, + "learning_rate": 7.782415909277632e-06, + "loss": 0.0557, + "step": 1534 + }, + { + "epoch": 0.6662326388888888, + "grad_norm": 0.6181557774543762, + "learning_rate": 7.77949485937321e-06, + "loss": 0.0572, + "step": 1535 + }, + { + "epoch": 0.6666666666666666, + "grad_norm": 0.48969897627830505, + "learning_rate": 7.77657243591746e-06, + "loss": 0.0443, + "step": 1536 + }, + { + "epoch": 0.6671006944444444, + "grad_norm": 0.6379843354225159, + "learning_rate": 7.773648640354564e-06, + "loss": 0.0525, + "step": 1537 + }, + { + "epoch": 0.6675347222222222, + "grad_norm": 0.6808762550354004, + "learning_rate": 7.770723474129385e-06, + "loss": 0.0578, + "step": 1538 + }, + { + "epoch": 0.66796875, + "grad_norm": 0.7974750995635986, + "learning_rate": 7.767796938687458e-06, + "loss": 0.0574, + "step": 1539 + }, + { + "epoch": 0.6684027777777778, + "grad_norm": 0.7461864352226257, + "learning_rate": 7.764869035474998e-06, + "loss": 0.0487, + "step": 1540 + }, + { + "epoch": 0.6688368055555556, + "grad_norm": 0.7737073302268982, + "learning_rate": 7.761939765938897e-06, + "loss": 0.0547, + "step": 1541 + }, + { + "epoch": 0.6692708333333334, + "grad_norm": 0.9133140444755554, + "learning_rate": 7.75900913152672e-06, + "loss": 0.0438, + "step": 1542 + }, + { + "epoch": 0.6697048611111112, + "grad_norm": 0.5385271906852722, + "learning_rate": 7.756077133686711e-06, + "loss": 0.0438, + "step": 1543 + }, + { + "epoch": 0.6701388888888888, + "grad_norm": 0.7025004029273987, + "learning_rate": 7.753143773867779e-06, + "loss": 0.0495, + "step": 1544 + }, + { + "epoch": 0.6705729166666666, + "grad_norm": 0.6356387138366699, + "learning_rate": 7.750209053519513e-06, + "loss": 0.0504, + "step": 1545 + }, + { + "epoch": 0.6710069444444444, + "grad_norm": 0.6328694224357605, + "learning_rate": 7.747272974092173e-06, + "loss": 0.0452, + "step": 1546 + }, + { + "epoch": 0.6714409722222222, + "grad_norm": 0.8696878552436829, + "learning_rate": 7.744335537036692e-06, + "loss": 0.0492, + "step": 1547 + }, + { + "epoch": 0.671875, + "grad_norm": 0.5818842649459839, + "learning_rate": 7.74139674380467e-06, + "loss": 0.0412, + "step": 1548 + }, + { + "epoch": 0.6723090277777778, + "grad_norm": 1.057603359222412, + "learning_rate": 7.738456595848378e-06, + "loss": 0.042, + "step": 1549 + }, + { + "epoch": 0.6727430555555556, + "grad_norm": 0.4587070643901825, + "learning_rate": 7.735515094620761e-06, + "loss": 0.0432, + "step": 1550 + }, + { + "epoch": 0.6731770833333334, + "grad_norm": 1.1908316612243652, + "learning_rate": 7.732572241575427e-06, + "loss": 0.059, + "step": 1551 + }, + { + "epoch": 0.6736111111111112, + "grad_norm": 0.5407928824424744, + "learning_rate": 7.729628038166655e-06, + "loss": 0.0471, + "step": 1552 + }, + { + "epoch": 0.6740451388888888, + "grad_norm": 1.1224613189697266, + "learning_rate": 7.726682485849392e-06, + "loss": 0.062, + "step": 1553 + }, + { + "epoch": 0.6744791666666666, + "grad_norm": 1.0806435346603394, + "learning_rate": 7.723735586079251e-06, + "loss": 0.0655, + "step": 1554 + }, + { + "epoch": 0.6749131944444444, + "grad_norm": 0.5161687731742859, + "learning_rate": 7.720787340312507e-06, + "loss": 0.0483, + "step": 1555 + }, + { + "epoch": 0.6753472222222222, + "grad_norm": 0.6101528406143188, + "learning_rate": 7.717837750006106e-06, + "loss": 0.0304, + "step": 1556 + }, + { + "epoch": 0.67578125, + "grad_norm": 0.6650863885879517, + "learning_rate": 7.714886816617656e-06, + "loss": 0.0492, + "step": 1557 + }, + { + "epoch": 0.6762152777777778, + "grad_norm": 0.9439712762832642, + "learning_rate": 7.711934541605427e-06, + "loss": 0.0711, + "step": 1558 + }, + { + "epoch": 0.6766493055555556, + "grad_norm": 1.069583773612976, + "learning_rate": 7.708980926428356e-06, + "loss": 0.0608, + "step": 1559 + }, + { + "epoch": 0.6770833333333334, + "grad_norm": 0.8776530623435974, + "learning_rate": 7.706025972546039e-06, + "loss": 0.0635, + "step": 1560 + }, + { + "epoch": 0.6775173611111112, + "grad_norm": 0.6253235936164856, + "learning_rate": 7.703069681418732e-06, + "loss": 0.0499, + "step": 1561 + }, + { + "epoch": 0.6779513888888888, + "grad_norm": 0.7380187511444092, + "learning_rate": 7.700112054507357e-06, + "loss": 0.0504, + "step": 1562 + }, + { + "epoch": 0.6783854166666666, + "grad_norm": 0.5171223282814026, + "learning_rate": 7.697153093273492e-06, + "loss": 0.0515, + "step": 1563 + }, + { + "epoch": 0.6788194444444444, + "grad_norm": 0.6803078055381775, + "learning_rate": 7.694192799179377e-06, + "loss": 0.0477, + "step": 1564 + }, + { + "epoch": 0.6792534722222222, + "grad_norm": 0.5703323483467102, + "learning_rate": 7.691231173687909e-06, + "loss": 0.0378, + "step": 1565 + }, + { + "epoch": 0.6796875, + "grad_norm": 0.6021401286125183, + "learning_rate": 7.688268218262645e-06, + "loss": 0.059, + "step": 1566 + }, + { + "epoch": 0.6801215277777778, + "grad_norm": 0.5687470436096191, + "learning_rate": 7.685303934367793e-06, + "loss": 0.046, + "step": 1567 + }, + { + "epoch": 0.6805555555555556, + "grad_norm": 0.9414650201797485, + "learning_rate": 7.682338323468227e-06, + "loss": 0.0621, + "step": 1568 + }, + { + "epoch": 0.6809895833333334, + "grad_norm": 0.8877261281013489, + "learning_rate": 7.679371387029468e-06, + "loss": 0.0636, + "step": 1569 + }, + { + "epoch": 0.6814236111111112, + "grad_norm": 0.7038198709487915, + "learning_rate": 7.676403126517698e-06, + "loss": 0.0404, + "step": 1570 + }, + { + "epoch": 0.6818576388888888, + "grad_norm": 1.1424853801727295, + "learning_rate": 7.67343354339975e-06, + "loss": 0.0591, + "step": 1571 + }, + { + "epoch": 0.6822916666666666, + "grad_norm": 1.11978280544281, + "learning_rate": 7.670462639143115e-06, + "loss": 0.0397, + "step": 1572 + }, + { + "epoch": 0.6827256944444444, + "grad_norm": 0.6603884696960449, + "learning_rate": 7.667490415215929e-06, + "loss": 0.0518, + "step": 1573 + }, + { + "epoch": 0.6831597222222222, + "grad_norm": 1.0959213972091675, + "learning_rate": 7.664516873086987e-06, + "loss": 0.0497, + "step": 1574 + }, + { + "epoch": 0.68359375, + "grad_norm": 0.6479515433311462, + "learning_rate": 7.661542014225735e-06, + "loss": 0.0527, + "step": 1575 + }, + { + "epoch": 0.6840277777777778, + "grad_norm": 0.9234191179275513, + "learning_rate": 7.658565840102265e-06, + "loss": 0.0648, + "step": 1576 + }, + { + "epoch": 0.6844618055555556, + "grad_norm": 0.9269822835922241, + "learning_rate": 7.655588352187325e-06, + "loss": 0.0615, + "step": 1577 + }, + { + "epoch": 0.6848958333333334, + "grad_norm": 0.6565536260604858, + "learning_rate": 7.652609551952306e-06, + "loss": 0.0344, + "step": 1578 + }, + { + "epoch": 0.6853298611111112, + "grad_norm": 0.5279817581176758, + "learning_rate": 7.649629440869254e-06, + "loss": 0.0355, + "step": 1579 + }, + { + "epoch": 0.6857638888888888, + "grad_norm": 0.5780903697013855, + "learning_rate": 7.646648020410857e-06, + "loss": 0.0485, + "step": 1580 + }, + { + "epoch": 0.6861979166666666, + "grad_norm": 0.5379993319511414, + "learning_rate": 7.64366529205045e-06, + "loss": 0.0319, + "step": 1581 + }, + { + "epoch": 0.6866319444444444, + "grad_norm": 0.559840738773346, + "learning_rate": 7.640681257262023e-06, + "loss": 0.0424, + "step": 1582 + }, + { + "epoch": 0.6870659722222222, + "grad_norm": 0.9723001718521118, + "learning_rate": 7.637695917520201e-06, + "loss": 0.0453, + "step": 1583 + }, + { + "epoch": 0.6875, + "grad_norm": 0.8327869772911072, + "learning_rate": 7.63470927430026e-06, + "loss": 0.0497, + "step": 1584 + }, + { + "epoch": 0.6879340277777778, + "grad_norm": 0.8263087868690491, + "learning_rate": 7.631721329078118e-06, + "loss": 0.0587, + "step": 1585 + }, + { + "epoch": 0.6883680555555556, + "grad_norm": 1.0317976474761963, + "learning_rate": 7.628732083330338e-06, + "loss": 0.0699, + "step": 1586 + }, + { + "epoch": 0.6888020833333334, + "grad_norm": 0.6567073464393616, + "learning_rate": 7.625741538534123e-06, + "loss": 0.052, + "step": 1587 + }, + { + "epoch": 0.6892361111111112, + "grad_norm": 0.8320260643959045, + "learning_rate": 7.622749696167318e-06, + "loss": 0.0462, + "step": 1588 + }, + { + "epoch": 0.6896701388888888, + "grad_norm": 1.1415224075317383, + "learning_rate": 7.619756557708415e-06, + "loss": 0.0721, + "step": 1589 + }, + { + "epoch": 0.6901041666666666, + "grad_norm": 0.49437031149864197, + "learning_rate": 7.61676212463654e-06, + "loss": 0.0414, + "step": 1590 + }, + { + "epoch": 0.6905381944444444, + "grad_norm": 0.6207118630409241, + "learning_rate": 7.613766398431462e-06, + "loss": 0.0658, + "step": 1591 + }, + { + "epoch": 0.6909722222222222, + "grad_norm": 0.9149729013442993, + "learning_rate": 7.610769380573584e-06, + "loss": 0.0702, + "step": 1592 + }, + { + "epoch": 0.69140625, + "grad_norm": 0.5002617239952087, + "learning_rate": 7.607771072543955e-06, + "loss": 0.041, + "step": 1593 + }, + { + "epoch": 0.6918402777777778, + "grad_norm": 0.49440690875053406, + "learning_rate": 7.604771475824258e-06, + "loss": 0.0492, + "step": 1594 + }, + { + "epoch": 0.6922743055555556, + "grad_norm": 0.47741520404815674, + "learning_rate": 7.6017705918968134e-06, + "loss": 0.0399, + "step": 1595 + }, + { + "epoch": 0.6927083333333334, + "grad_norm": 0.4327121376991272, + "learning_rate": 7.598768422244573e-06, + "loss": 0.0406, + "step": 1596 + }, + { + "epoch": 0.6931423611111112, + "grad_norm": 0.633120059967041, + "learning_rate": 7.595764968351132e-06, + "loss": 0.0511, + "step": 1597 + }, + { + "epoch": 0.6935763888888888, + "grad_norm": 0.8574790954589844, + "learning_rate": 7.592760231700715e-06, + "loss": 0.0784, + "step": 1598 + }, + { + "epoch": 0.6940104166666666, + "grad_norm": 0.6364694237709045, + "learning_rate": 7.589754213778181e-06, + "loss": 0.0475, + "step": 1599 + }, + { + "epoch": 0.6944444444444444, + "grad_norm": 1.3502572774887085, + "learning_rate": 7.586746916069023e-06, + "loss": 0.0619, + "step": 1600 + }, + { + "epoch": 0.6948784722222222, + "grad_norm": 0.5626755952835083, + "learning_rate": 7.583738340059368e-06, + "loss": 0.048, + "step": 1601 + }, + { + "epoch": 0.6953125, + "grad_norm": 0.6648493409156799, + "learning_rate": 7.580728487235971e-06, + "loss": 0.0498, + "step": 1602 + }, + { + "epoch": 0.6957465277777778, + "grad_norm": 0.8648276925086975, + "learning_rate": 7.577717359086221e-06, + "loss": 0.0558, + "step": 1603 + }, + { + "epoch": 0.6961805555555556, + "grad_norm": 0.9568964242935181, + "learning_rate": 7.574704957098135e-06, + "loss": 0.0615, + "step": 1604 + }, + { + "epoch": 0.6966145833333334, + "grad_norm": 0.7265316247940063, + "learning_rate": 7.571691282760364e-06, + "loss": 0.0465, + "step": 1605 + }, + { + "epoch": 0.6970486111111112, + "grad_norm": 0.5609695315361023, + "learning_rate": 7.568676337562182e-06, + "loss": 0.0579, + "step": 1606 + }, + { + "epoch": 0.6974826388888888, + "grad_norm": 0.4945490062236786, + "learning_rate": 7.565660122993494e-06, + "loss": 0.0338, + "step": 1607 + }, + { + "epoch": 0.6979166666666666, + "grad_norm": 0.691581130027771, + "learning_rate": 7.56264264054483e-06, + "loss": 0.0491, + "step": 1608 + }, + { + "epoch": 0.6983506944444444, + "grad_norm": 0.4934881329536438, + "learning_rate": 7.55962389170735e-06, + "loss": 0.0305, + "step": 1609 + }, + { + "epoch": 0.6987847222222222, + "grad_norm": 1.052710771560669, + "learning_rate": 7.556603877972841e-06, + "loss": 0.0774, + "step": 1610 + }, + { + "epoch": 0.69921875, + "grad_norm": 0.5597624778747559, + "learning_rate": 7.553582600833709e-06, + "loss": 0.0337, + "step": 1611 + }, + { + "epoch": 0.6996527777777778, + "grad_norm": 0.85475754737854, + "learning_rate": 7.550560061782986e-06, + "loss": 0.0528, + "step": 1612 + }, + { + "epoch": 0.7000868055555556, + "grad_norm": 0.5703578591346741, + "learning_rate": 7.5475362623143345e-06, + "loss": 0.027, + "step": 1613 + }, + { + "epoch": 0.7005208333333334, + "grad_norm": 0.6355430483818054, + "learning_rate": 7.544511203922028e-06, + "loss": 0.046, + "step": 1614 + }, + { + "epoch": 0.7009548611111112, + "grad_norm": 1.1655287742614746, + "learning_rate": 7.541484888100974e-06, + "loss": 0.0724, + "step": 1615 + }, + { + "epoch": 0.7013888888888888, + "grad_norm": 1.0697033405303955, + "learning_rate": 7.5384573163466945e-06, + "loss": 0.0626, + "step": 1616 + }, + { + "epoch": 0.7018229166666666, + "grad_norm": 0.6042693853378296, + "learning_rate": 7.535428490155332e-06, + "loss": 0.0394, + "step": 1617 + }, + { + "epoch": 0.7022569444444444, + "grad_norm": 1.1471283435821533, + "learning_rate": 7.532398411023654e-06, + "loss": 0.0676, + "step": 1618 + }, + { + "epoch": 0.7026909722222222, + "grad_norm": 0.6342034339904785, + "learning_rate": 7.5293670804490396e-06, + "loss": 0.0401, + "step": 1619 + }, + { + "epoch": 0.703125, + "grad_norm": 0.630740761756897, + "learning_rate": 7.526334499929495e-06, + "loss": 0.052, + "step": 1620 + }, + { + "epoch": 0.7035590277777778, + "grad_norm": 1.228358507156372, + "learning_rate": 7.523300670963636e-06, + "loss": 0.066, + "step": 1621 + }, + { + "epoch": 0.7039930555555556, + "grad_norm": 0.7143034934997559, + "learning_rate": 7.520265595050702e-06, + "loss": 0.0544, + "step": 1622 + }, + { + "epoch": 0.7044270833333334, + "grad_norm": 0.8387890458106995, + "learning_rate": 7.517229273690543e-06, + "loss": 0.0758, + "step": 1623 + }, + { + "epoch": 0.7048611111111112, + "grad_norm": 0.5054572820663452, + "learning_rate": 7.514191708383627e-06, + "loss": 0.0383, + "step": 1624 + }, + { + "epoch": 0.7052951388888888, + "grad_norm": 0.522240161895752, + "learning_rate": 7.511152900631038e-06, + "loss": 0.0473, + "step": 1625 + }, + { + "epoch": 0.7057291666666666, + "grad_norm": 0.5247807502746582, + "learning_rate": 7.508112851934475e-06, + "loss": 0.0567, + "step": 1626 + }, + { + "epoch": 0.7061631944444444, + "grad_norm": 0.579017698764801, + "learning_rate": 7.505071563796245e-06, + "loss": 0.0429, + "step": 1627 + }, + { + "epoch": 0.7065972222222222, + "grad_norm": 0.5940823554992676, + "learning_rate": 7.5020290377192696e-06, + "loss": 0.0493, + "step": 1628 + }, + { + "epoch": 0.70703125, + "grad_norm": 0.6563714742660522, + "learning_rate": 7.498985275207086e-06, + "loss": 0.0531, + "step": 1629 + }, + { + "epoch": 0.7074652777777778, + "grad_norm": 0.5494057536125183, + "learning_rate": 7.4959402777638365e-06, + "loss": 0.0298, + "step": 1630 + }, + { + "epoch": 0.7078993055555556, + "grad_norm": 0.4213770031929016, + "learning_rate": 7.492894046894279e-06, + "loss": 0.0349, + "step": 1631 + }, + { + "epoch": 0.7083333333333334, + "grad_norm": 0.6510403752326965, + "learning_rate": 7.489846584103778e-06, + "loss": 0.0414, + "step": 1632 + }, + { + "epoch": 0.7087673611111112, + "grad_norm": 0.7641744613647461, + "learning_rate": 7.486797890898307e-06, + "loss": 0.0448, + "step": 1633 + }, + { + "epoch": 0.7092013888888888, + "grad_norm": 0.8901313543319702, + "learning_rate": 7.483747968784449e-06, + "loss": 0.0624, + "step": 1634 + }, + { + "epoch": 0.7096354166666666, + "grad_norm": 0.6470595002174377, + "learning_rate": 7.480696819269391e-06, + "loss": 0.0396, + "step": 1635 + }, + { + "epoch": 0.7100694444444444, + "grad_norm": 1.072182059288025, + "learning_rate": 7.47764444386093e-06, + "loss": 0.0712, + "step": 1636 + }, + { + "epoch": 0.7105034722222222, + "grad_norm": 0.8364940881729126, + "learning_rate": 7.474590844067468e-06, + "loss": 0.0454, + "step": 1637 + }, + { + "epoch": 0.7109375, + "grad_norm": 0.7541149258613586, + "learning_rate": 7.471536021398009e-06, + "loss": 0.0518, + "step": 1638 + }, + { + "epoch": 0.7113715277777778, + "grad_norm": 0.7979973554611206, + "learning_rate": 7.468479977362169e-06, + "loss": 0.0439, + "step": 1639 + }, + { + "epoch": 0.7118055555555556, + "grad_norm": 0.6536715030670166, + "learning_rate": 7.465422713470156e-06, + "loss": 0.0482, + "step": 1640 + }, + { + "epoch": 0.7122395833333334, + "grad_norm": 1.1344897747039795, + "learning_rate": 7.462364231232794e-06, + "loss": 0.0427, + "step": 1641 + }, + { + "epoch": 0.7126736111111112, + "grad_norm": 0.6313517093658447, + "learning_rate": 7.459304532161498e-06, + "loss": 0.039, + "step": 1642 + }, + { + "epoch": 0.7131076388888888, + "grad_norm": 0.8508172631263733, + "learning_rate": 7.4562436177682885e-06, + "loss": 0.0594, + "step": 1643 + }, + { + "epoch": 0.7135416666666666, + "grad_norm": 0.7220401167869568, + "learning_rate": 7.4531814895657885e-06, + "loss": 0.0448, + "step": 1644 + }, + { + "epoch": 0.7139756944444444, + "grad_norm": 0.7740882635116577, + "learning_rate": 7.450118149067218e-06, + "loss": 0.0564, + "step": 1645 + }, + { + "epoch": 0.7144097222222222, + "grad_norm": 0.41988405585289, + "learning_rate": 7.447053597786398e-06, + "loss": 0.0383, + "step": 1646 + }, + { + "epoch": 0.71484375, + "grad_norm": 0.9427817463874817, + "learning_rate": 7.443987837237749e-06, + "loss": 0.0673, + "step": 1647 + }, + { + "epoch": 0.7152777777777778, + "grad_norm": 0.9876265525817871, + "learning_rate": 7.440920868936282e-06, + "loss": 0.0574, + "step": 1648 + }, + { + "epoch": 0.7157118055555556, + "grad_norm": 0.6729372143745422, + "learning_rate": 7.437852694397615e-06, + "loss": 0.0492, + "step": 1649 + }, + { + "epoch": 0.7161458333333334, + "grad_norm": 0.677540123462677, + "learning_rate": 7.434783315137953e-06, + "loss": 0.046, + "step": 1650 + }, + { + "epoch": 0.7165798611111112, + "grad_norm": 0.7205467820167542, + "learning_rate": 7.431712732674103e-06, + "loss": 0.0416, + "step": 1651 + }, + { + "epoch": 0.7170138888888888, + "grad_norm": 0.7260408401489258, + "learning_rate": 7.428640948523463e-06, + "loss": 0.0257, + "step": 1652 + }, + { + "epoch": 0.7174479166666666, + "grad_norm": 0.9158640503883362, + "learning_rate": 7.4255679642040265e-06, + "loss": 0.0706, + "step": 1653 + }, + { + "epoch": 0.7178819444444444, + "grad_norm": 1.2473180294036865, + "learning_rate": 7.422493781234379e-06, + "loss": 0.0516, + "step": 1654 + }, + { + "epoch": 0.7183159722222222, + "grad_norm": 0.8920367956161499, + "learning_rate": 7.419418401133698e-06, + "loss": 0.0554, + "step": 1655 + }, + { + "epoch": 0.71875, + "grad_norm": 0.6255639791488647, + "learning_rate": 7.416341825421755e-06, + "loss": 0.0463, + "step": 1656 + }, + { + "epoch": 0.7191840277777778, + "grad_norm": 0.8922591209411621, + "learning_rate": 7.413264055618909e-06, + "loss": 0.0701, + "step": 1657 + }, + { + "epoch": 0.7196180555555556, + "grad_norm": 0.8104597330093384, + "learning_rate": 7.4101850932461095e-06, + "loss": 0.0695, + "step": 1658 + }, + { + "epoch": 0.7200520833333334, + "grad_norm": 0.5503954887390137, + "learning_rate": 7.407104939824901e-06, + "loss": 0.0359, + "step": 1659 + }, + { + "epoch": 0.7204861111111112, + "grad_norm": 0.5614795684814453, + "learning_rate": 7.404023596877409e-06, + "loss": 0.0535, + "step": 1660 + }, + { + "epoch": 0.7209201388888888, + "grad_norm": 1.2029560804367065, + "learning_rate": 7.40094106592635e-06, + "loss": 0.067, + "step": 1661 + }, + { + "epoch": 0.7213541666666666, + "grad_norm": 0.5232289433479309, + "learning_rate": 7.39785734849503e-06, + "loss": 0.0389, + "step": 1662 + }, + { + "epoch": 0.7217881944444444, + "grad_norm": 0.6728552579879761, + "learning_rate": 7.394772446107335e-06, + "loss": 0.0629, + "step": 1663 + }, + { + "epoch": 0.7222222222222222, + "grad_norm": 0.64680415391922, + "learning_rate": 7.391686360287744e-06, + "loss": 0.044, + "step": 1664 + }, + { + "epoch": 0.72265625, + "grad_norm": 0.5645890831947327, + "learning_rate": 7.388599092561315e-06, + "loss": 0.0496, + "step": 1665 + }, + { + "epoch": 0.7230902777777778, + "grad_norm": 0.6096482872962952, + "learning_rate": 7.385510644453694e-06, + "loss": 0.0591, + "step": 1666 + }, + { + "epoch": 0.7235243055555556, + "grad_norm": 0.7313439846038818, + "learning_rate": 7.382421017491108e-06, + "loss": 0.0697, + "step": 1667 + }, + { + "epoch": 0.7239583333333334, + "grad_norm": 0.6447915434837341, + "learning_rate": 7.3793302132003665e-06, + "loss": 0.0465, + "step": 1668 + }, + { + "epoch": 0.7243923611111112, + "grad_norm": 0.6097691655158997, + "learning_rate": 7.376238233108863e-06, + "loss": 0.0352, + "step": 1669 + }, + { + "epoch": 0.7248263888888888, + "grad_norm": 0.7632279992103577, + "learning_rate": 7.37314507874457e-06, + "loss": 0.055, + "step": 1670 + }, + { + "epoch": 0.7252604166666666, + "grad_norm": 0.7166815996170044, + "learning_rate": 7.370050751636041e-06, + "loss": 0.0426, + "step": 1671 + }, + { + "epoch": 0.7256944444444444, + "grad_norm": 0.9069613814353943, + "learning_rate": 7.366955253312407e-06, + "loss": 0.0528, + "step": 1672 + }, + { + "epoch": 0.7261284722222222, + "grad_norm": 0.7017959356307983, + "learning_rate": 7.3638585853033826e-06, + "loss": 0.0519, + "step": 1673 + }, + { + "epoch": 0.7265625, + "grad_norm": 0.594621479511261, + "learning_rate": 7.360760749139257e-06, + "loss": 0.0357, + "step": 1674 + }, + { + "epoch": 0.7269965277777778, + "grad_norm": 0.7259089350700378, + "learning_rate": 7.357661746350895e-06, + "loss": 0.0497, + "step": 1675 + }, + { + "epoch": 0.7274305555555556, + "grad_norm": 0.4349896311759949, + "learning_rate": 7.354561578469743e-06, + "loss": 0.0321, + "step": 1676 + }, + { + "epoch": 0.7278645833333334, + "grad_norm": 1.0102602243423462, + "learning_rate": 7.351460247027818e-06, + "loss": 0.0527, + "step": 1677 + }, + { + "epoch": 0.7282986111111112, + "grad_norm": 0.8198878169059753, + "learning_rate": 7.348357753557716e-06, + "loss": 0.0447, + "step": 1678 + }, + { + "epoch": 0.7287326388888888, + "grad_norm": 0.8680114150047302, + "learning_rate": 7.345254099592605e-06, + "loss": 0.0543, + "step": 1679 + }, + { + "epoch": 0.7291666666666666, + "grad_norm": 0.9011032581329346, + "learning_rate": 7.342149286666226e-06, + "loss": 0.0598, + "step": 1680 + }, + { + "epoch": 0.7296006944444444, + "grad_norm": 0.6858422756195068, + "learning_rate": 7.339043316312895e-06, + "loss": 0.0359, + "step": 1681 + }, + { + "epoch": 0.7300347222222222, + "grad_norm": 0.875012993812561, + "learning_rate": 7.335936190067497e-06, + "loss": 0.0593, + "step": 1682 + }, + { + "epoch": 0.73046875, + "grad_norm": 0.6261489987373352, + "learning_rate": 7.332827909465493e-06, + "loss": 0.0505, + "step": 1683 + }, + { + "epoch": 0.7309027777777778, + "grad_norm": 0.7853638529777527, + "learning_rate": 7.329718476042908e-06, + "loss": 0.0413, + "step": 1684 + }, + { + "epoch": 0.7313368055555556, + "grad_norm": 0.6131523251533508, + "learning_rate": 7.326607891336342e-06, + "loss": 0.0472, + "step": 1685 + }, + { + "epoch": 0.7317708333333334, + "grad_norm": 0.9483081102371216, + "learning_rate": 7.323496156882961e-06, + "loss": 0.0606, + "step": 1686 + }, + { + "epoch": 0.7322048611111112, + "grad_norm": 0.7218665480613708, + "learning_rate": 7.320383274220502e-06, + "loss": 0.0458, + "step": 1687 + }, + { + "epoch": 0.7326388888888888, + "grad_norm": 0.6520138382911682, + "learning_rate": 7.317269244887265e-06, + "loss": 0.0495, + "step": 1688 + }, + { + "epoch": 0.7330729166666666, + "grad_norm": 0.4401592016220093, + "learning_rate": 7.314154070422121e-06, + "loss": 0.031, + "step": 1689 + }, + { + "epoch": 0.7335069444444444, + "grad_norm": 0.7516753673553467, + "learning_rate": 7.311037752364505e-06, + "loss": 0.0407, + "step": 1690 + }, + { + "epoch": 0.7339409722222222, + "grad_norm": 0.3969035744667053, + "learning_rate": 7.307920292254417e-06, + "loss": 0.0274, + "step": 1691 + }, + { + "epoch": 0.734375, + "grad_norm": 0.773316502571106, + "learning_rate": 7.304801691632421e-06, + "loss": 0.0572, + "step": 1692 + }, + { + "epoch": 0.7348090277777778, + "grad_norm": 0.665576696395874, + "learning_rate": 7.301681952039645e-06, + "loss": 0.0467, + "step": 1693 + }, + { + "epoch": 0.7352430555555556, + "grad_norm": 0.601359486579895, + "learning_rate": 7.298561075017783e-06, + "loss": 0.0457, + "step": 1694 + }, + { + "epoch": 0.7356770833333334, + "grad_norm": 0.6760073304176331, + "learning_rate": 7.295439062109085e-06, + "loss": 0.058, + "step": 1695 + }, + { + "epoch": 0.7361111111111112, + "grad_norm": 1.0305536985397339, + "learning_rate": 7.292315914856366e-06, + "loss": 0.0546, + "step": 1696 + }, + { + "epoch": 0.7365451388888888, + "grad_norm": 0.6392028331756592, + "learning_rate": 7.289191634803002e-06, + "loss": 0.0425, + "step": 1697 + }, + { + "epoch": 0.7369791666666666, + "grad_norm": 0.985002875328064, + "learning_rate": 7.286066223492928e-06, + "loss": 0.0679, + "step": 1698 + }, + { + "epoch": 0.7374131944444444, + "grad_norm": 0.6694068312644958, + "learning_rate": 7.282939682470637e-06, + "loss": 0.029, + "step": 1699 + }, + { + "epoch": 0.7378472222222222, + "grad_norm": 0.41032591462135315, + "learning_rate": 7.2798120132811815e-06, + "loss": 0.0295, + "step": 1700 + }, + { + "epoch": 0.73828125, + "grad_norm": 0.7770684957504272, + "learning_rate": 7.276683217470171e-06, + "loss": 0.064, + "step": 1701 + }, + { + "epoch": 0.7387152777777778, + "grad_norm": 0.552871584892273, + "learning_rate": 7.273553296583773e-06, + "loss": 0.0394, + "step": 1702 + }, + { + "epoch": 0.7391493055555556, + "grad_norm": 1.2697641849517822, + "learning_rate": 7.270422252168708e-06, + "loss": 0.0613, + "step": 1703 + }, + { + "epoch": 0.7395833333333334, + "grad_norm": 0.560497522354126, + "learning_rate": 7.267290085772256e-06, + "loss": 0.0303, + "step": 1704 + }, + { + "epoch": 0.7400173611111112, + "grad_norm": 0.6473203301429749, + "learning_rate": 7.264156798942247e-06, + "loss": 0.0397, + "step": 1705 + }, + { + "epoch": 0.7404513888888888, + "grad_norm": 0.6359472870826721, + "learning_rate": 7.2610223932270665e-06, + "loss": 0.0434, + "step": 1706 + }, + { + "epoch": 0.7408854166666666, + "grad_norm": 0.7363631725311279, + "learning_rate": 7.2578868701756564e-06, + "loss": 0.0445, + "step": 1707 + }, + { + "epoch": 0.7413194444444444, + "grad_norm": 0.9180881381034851, + "learning_rate": 7.254750231337503e-06, + "loss": 0.0514, + "step": 1708 + }, + { + "epoch": 0.7417534722222222, + "grad_norm": 1.7444438934326172, + "learning_rate": 7.251612478262653e-06, + "loss": 0.0694, + "step": 1709 + }, + { + "epoch": 0.7421875, + "grad_norm": 0.7270860075950623, + "learning_rate": 7.2484736125016965e-06, + "loss": 0.0561, + "step": 1710 + }, + { + "epoch": 0.7426215277777778, + "grad_norm": 0.9425753951072693, + "learning_rate": 7.245333635605778e-06, + "loss": 0.0519, + "step": 1711 + }, + { + "epoch": 0.7430555555555556, + "grad_norm": 0.7384446263313293, + "learning_rate": 7.242192549126586e-06, + "loss": 0.0374, + "step": 1712 + }, + { + "epoch": 0.7434895833333334, + "grad_norm": 0.6277630925178528, + "learning_rate": 7.239050354616365e-06, + "loss": 0.0412, + "step": 1713 + }, + { + "epoch": 0.7439236111111112, + "grad_norm": 1.0086884498596191, + "learning_rate": 7.235907053627899e-06, + "loss": 0.065, + "step": 1714 + }, + { + "epoch": 0.7443576388888888, + "grad_norm": 1.4292784929275513, + "learning_rate": 7.232762647714525e-06, + "loss": 0.092, + "step": 1715 + }, + { + "epoch": 0.7447916666666666, + "grad_norm": 0.9445415139198303, + "learning_rate": 7.229617138430123e-06, + "loss": 0.0444, + "step": 1716 + }, + { + "epoch": 0.7452256944444444, + "grad_norm": 0.821932852268219, + "learning_rate": 7.226470527329117e-06, + "loss": 0.0449, + "step": 1717 + }, + { + "epoch": 0.7456597222222222, + "grad_norm": 0.6048581600189209, + "learning_rate": 7.22332281596648e-06, + "loss": 0.0496, + "step": 1718 + }, + { + "epoch": 0.74609375, + "grad_norm": 0.5893974900245667, + "learning_rate": 7.220174005897724e-06, + "loss": 0.0543, + "step": 1719 + }, + { + "epoch": 0.7465277777777778, + "grad_norm": 0.702113687992096, + "learning_rate": 7.217024098678908e-06, + "loss": 0.0479, + "step": 1720 + }, + { + "epoch": 0.7469618055555556, + "grad_norm": 0.7570499777793884, + "learning_rate": 7.213873095866628e-06, + "loss": 0.0529, + "step": 1721 + }, + { + "epoch": 0.7473958333333334, + "grad_norm": 0.8346821069717407, + "learning_rate": 7.210720999018027e-06, + "loss": 0.0518, + "step": 1722 + }, + { + "epoch": 0.7478298611111112, + "grad_norm": 0.5654697418212891, + "learning_rate": 7.207567809690786e-06, + "loss": 0.0538, + "step": 1723 + }, + { + "epoch": 0.7482638888888888, + "grad_norm": 1.3470139503479004, + "learning_rate": 7.204413529443125e-06, + "loss": 0.0699, + "step": 1724 + }, + { + "epoch": 0.7486979166666666, + "grad_norm": 0.6355326771736145, + "learning_rate": 7.2012581598338035e-06, + "loss": 0.0601, + "step": 1725 + }, + { + "epoch": 0.7491319444444444, + "grad_norm": 0.6898193359375, + "learning_rate": 7.198101702422123e-06, + "loss": 0.0562, + "step": 1726 + }, + { + "epoch": 0.7495659722222222, + "grad_norm": 0.7929469347000122, + "learning_rate": 7.194944158767916e-06, + "loss": 0.067, + "step": 1727 + }, + { + "epoch": 0.75, + "grad_norm": 0.7281990051269531, + "learning_rate": 7.1917855304315575e-06, + "loss": 0.0668, + "step": 1728 + }, + { + "epoch": 0.7504340277777778, + "grad_norm": 0.45239919424057007, + "learning_rate": 7.188625818973955e-06, + "loss": 0.0265, + "step": 1729 + }, + { + "epoch": 0.7508680555555556, + "grad_norm": 0.8641472458839417, + "learning_rate": 7.185465025956555e-06, + "loss": 0.07, + "step": 1730 + }, + { + "epoch": 0.7513020833333334, + "grad_norm": 0.6607235670089722, + "learning_rate": 7.182303152941334e-06, + "loss": 0.0557, + "step": 1731 + }, + { + "epoch": 0.7517361111111112, + "grad_norm": 0.5443063378334045, + "learning_rate": 7.179140201490805e-06, + "loss": 0.0493, + "step": 1732 + }, + { + "epoch": 0.7521701388888888, + "grad_norm": 0.6223754286766052, + "learning_rate": 7.1759761731680135e-06, + "loss": 0.0585, + "step": 1733 + }, + { + "epoch": 0.7526041666666666, + "grad_norm": 0.5347004532814026, + "learning_rate": 7.172811069536536e-06, + "loss": 0.0506, + "step": 1734 + }, + { + "epoch": 0.7530381944444444, + "grad_norm": 0.509100079536438, + "learning_rate": 7.169644892160481e-06, + "loss": 0.0413, + "step": 1735 + }, + { + "epoch": 0.7534722222222222, + "grad_norm": 0.9070925712585449, + "learning_rate": 7.166477642604489e-06, + "loss": 0.0576, + "step": 1736 + }, + { + "epoch": 0.75390625, + "grad_norm": 0.46375447511672974, + "learning_rate": 7.163309322433728e-06, + "loss": 0.0442, + "step": 1737 + }, + { + "epoch": 0.7543402777777778, + "grad_norm": 1.7822531461715698, + "learning_rate": 7.160139933213899e-06, + "loss": 0.0693, + "step": 1738 + }, + { + "epoch": 0.7547743055555556, + "grad_norm": 0.5103685855865479, + "learning_rate": 7.156969476511226e-06, + "loss": 0.0433, + "step": 1739 + }, + { + "epoch": 0.7552083333333334, + "grad_norm": 0.5952293276786804, + "learning_rate": 7.153797953892463e-06, + "loss": 0.0592, + "step": 1740 + }, + { + "epoch": 0.7556423611111112, + "grad_norm": 0.8038705587387085, + "learning_rate": 7.150625366924893e-06, + "loss": 0.0575, + "step": 1741 + }, + { + "epoch": 0.7560763888888888, + "grad_norm": 1.7251653671264648, + "learning_rate": 7.14745171717632e-06, + "loss": 0.0634, + "step": 1742 + }, + { + "epoch": 0.7565104166666666, + "grad_norm": 0.5570972561836243, + "learning_rate": 7.144277006215079e-06, + "loss": 0.048, + "step": 1743 + }, + { + "epoch": 0.7569444444444444, + "grad_norm": 0.6353147625923157, + "learning_rate": 7.141101235610025e-06, + "loss": 0.0506, + "step": 1744 + }, + { + "epoch": 0.7573784722222222, + "grad_norm": 0.7536773085594177, + "learning_rate": 7.137924406930539e-06, + "loss": 0.0478, + "step": 1745 + }, + { + "epoch": 0.7578125, + "grad_norm": 0.6892389059066772, + "learning_rate": 7.134746521746521e-06, + "loss": 0.0473, + "step": 1746 + }, + { + "epoch": 0.7582465277777778, + "grad_norm": 1.0137925148010254, + "learning_rate": 7.131567581628398e-06, + "loss": 0.0649, + "step": 1747 + }, + { + "epoch": 0.7586805555555556, + "grad_norm": 0.7530260682106018, + "learning_rate": 7.128387588147119e-06, + "loss": 0.0453, + "step": 1748 + }, + { + "epoch": 0.7591145833333334, + "grad_norm": 0.8550921678543091, + "learning_rate": 7.125206542874146e-06, + "loss": 0.0603, + "step": 1749 + }, + { + "epoch": 0.7595486111111112, + "grad_norm": 0.42573830485343933, + "learning_rate": 7.122024447381469e-06, + "loss": 0.0342, + "step": 1750 + }, + { + "epoch": 0.7599826388888888, + "grad_norm": 0.6021143794059753, + "learning_rate": 7.118841303241594e-06, + "loss": 0.0416, + "step": 1751 + }, + { + "epoch": 0.7604166666666666, + "grad_norm": 0.6040723323822021, + "learning_rate": 7.115657112027543e-06, + "loss": 0.0495, + "step": 1752 + }, + { + "epoch": 0.7608506944444444, + "grad_norm": 0.5850236415863037, + "learning_rate": 7.112471875312858e-06, + "loss": 0.0542, + "step": 1753 + }, + { + "epoch": 0.7612847222222222, + "grad_norm": 0.5208216309547424, + "learning_rate": 7.109285594671599e-06, + "loss": 0.046, + "step": 1754 + }, + { + "epoch": 0.76171875, + "grad_norm": 0.6182832717895508, + "learning_rate": 7.106098271678337e-06, + "loss": 0.0562, + "step": 1755 + }, + { + "epoch": 0.7621527777777778, + "grad_norm": 0.6657155156135559, + "learning_rate": 7.102909907908161e-06, + "loss": 0.0701, + "step": 1756 + }, + { + "epoch": 0.7625868055555556, + "grad_norm": 0.62384033203125, + "learning_rate": 7.099720504936676e-06, + "loss": 0.06, + "step": 1757 + }, + { + "epoch": 0.7630208333333334, + "grad_norm": 0.8038895130157471, + "learning_rate": 7.096530064339997e-06, + "loss": 0.058, + "step": 1758 + }, + { + "epoch": 0.7634548611111112, + "grad_norm": 0.999763548374176, + "learning_rate": 7.0933385876947566e-06, + "loss": 0.0506, + "step": 1759 + }, + { + "epoch": 0.7638888888888888, + "grad_norm": 0.6280719637870789, + "learning_rate": 7.090146076578095e-06, + "loss": 0.053, + "step": 1760 + }, + { + "epoch": 0.7643229166666666, + "grad_norm": 0.7273646593093872, + "learning_rate": 7.086952532567664e-06, + "loss": 0.0512, + "step": 1761 + }, + { + "epoch": 0.7647569444444444, + "grad_norm": 0.822219967842102, + "learning_rate": 7.083757957241627e-06, + "loss": 0.0716, + "step": 1762 + }, + { + "epoch": 0.7651909722222222, + "grad_norm": 1.076998233795166, + "learning_rate": 7.080562352178657e-06, + "loss": 0.0819, + "step": 1763 + }, + { + "epoch": 0.765625, + "grad_norm": 0.5817064642906189, + "learning_rate": 7.0773657189579355e-06, + "loss": 0.0407, + "step": 1764 + }, + { + "epoch": 0.7660590277777778, + "grad_norm": 0.6132467985153198, + "learning_rate": 7.074168059159155e-06, + "loss": 0.0469, + "step": 1765 + }, + { + "epoch": 0.7664930555555556, + "grad_norm": 0.6296002864837646, + "learning_rate": 7.07096937436251e-06, + "loss": 0.045, + "step": 1766 + }, + { + "epoch": 0.7669270833333334, + "grad_norm": 0.6748993992805481, + "learning_rate": 7.067769666148705e-06, + "loss": 0.0394, + "step": 1767 + }, + { + "epoch": 0.7673611111111112, + "grad_norm": 0.6105886697769165, + "learning_rate": 7.064568936098948e-06, + "loss": 0.0406, + "step": 1768 + }, + { + "epoch": 0.7677951388888888, + "grad_norm": 0.8035129308700562, + "learning_rate": 7.061367185794956e-06, + "loss": 0.0612, + "step": 1769 + }, + { + "epoch": 0.7682291666666666, + "grad_norm": 1.034857153892517, + "learning_rate": 7.058164416818945e-06, + "loss": 0.0544, + "step": 1770 + }, + { + "epoch": 0.7686631944444444, + "grad_norm": 0.629572331905365, + "learning_rate": 7.054960630753639e-06, + "loss": 0.0444, + "step": 1771 + }, + { + "epoch": 0.7690972222222222, + "grad_norm": 0.9120610356330872, + "learning_rate": 7.051755829182262e-06, + "loss": 0.0683, + "step": 1772 + }, + { + "epoch": 0.76953125, + "grad_norm": 0.6496632695198059, + "learning_rate": 7.04855001368854e-06, + "loss": 0.0565, + "step": 1773 + }, + { + "epoch": 0.7699652777777778, + "grad_norm": 0.5675666928291321, + "learning_rate": 7.045343185856701e-06, + "loss": 0.0415, + "step": 1774 + }, + { + "epoch": 0.7703993055555556, + "grad_norm": 0.6537423133850098, + "learning_rate": 7.042135347271472e-06, + "loss": 0.0598, + "step": 1775 + }, + { + "epoch": 0.7708333333333334, + "grad_norm": 0.7533829808235168, + "learning_rate": 7.038926499518079e-06, + "loss": 0.0715, + "step": 1776 + }, + { + "epoch": 0.7712673611111112, + "grad_norm": 0.529955267906189, + "learning_rate": 7.035716644182251e-06, + "loss": 0.0459, + "step": 1777 + }, + { + "epoch": 0.7717013888888888, + "grad_norm": 0.5904744863510132, + "learning_rate": 7.032505782850209e-06, + "loss": 0.0398, + "step": 1778 + }, + { + "epoch": 0.7721354166666666, + "grad_norm": 0.4565054178237915, + "learning_rate": 7.029293917108678e-06, + "loss": 0.0445, + "step": 1779 + }, + { + "epoch": 0.7725694444444444, + "grad_norm": 0.5179257392883301, + "learning_rate": 7.026081048544873e-06, + "loss": 0.0374, + "step": 1780 + }, + { + "epoch": 0.7730034722222222, + "grad_norm": 0.4564122259616852, + "learning_rate": 7.022867178746506e-06, + "loss": 0.0397, + "step": 1781 + }, + { + "epoch": 0.7734375, + "grad_norm": 0.9230285286903381, + "learning_rate": 7.019652309301785e-06, + "loss": 0.0634, + "step": 1782 + }, + { + "epoch": 0.7738715277777778, + "grad_norm": 0.6400481462478638, + "learning_rate": 7.016436441799412e-06, + "loss": 0.0603, + "step": 1783 + }, + { + "epoch": 0.7743055555555556, + "grad_norm": 0.7093133330345154, + "learning_rate": 7.0132195778285806e-06, + "loss": 0.0433, + "step": 1784 + }, + { + "epoch": 0.7747395833333334, + "grad_norm": 0.5242016315460205, + "learning_rate": 7.010001718978981e-06, + "loss": 0.0424, + "step": 1785 + }, + { + "epoch": 0.7751736111111112, + "grad_norm": 0.5063726305961609, + "learning_rate": 7.006782866840791e-06, + "loss": 0.0332, + "step": 1786 + }, + { + "epoch": 0.7756076388888888, + "grad_norm": 0.8927604556083679, + "learning_rate": 7.00356302300468e-06, + "loss": 0.0459, + "step": 1787 + }, + { + "epoch": 0.7760416666666666, + "grad_norm": 0.6253120303153992, + "learning_rate": 7.0003421890618065e-06, + "loss": 0.0442, + "step": 1788 + }, + { + "epoch": 0.7764756944444444, + "grad_norm": 0.6660035848617554, + "learning_rate": 6.997120366603823e-06, + "loss": 0.0479, + "step": 1789 + }, + { + "epoch": 0.7769097222222222, + "grad_norm": 0.7846994996070862, + "learning_rate": 6.993897557222865e-06, + "loss": 0.0601, + "step": 1790 + }, + { + "epoch": 0.77734375, + "grad_norm": 0.4856984317302704, + "learning_rate": 6.990673762511557e-06, + "loss": 0.03, + "step": 1791 + }, + { + "epoch": 0.7777777777777778, + "grad_norm": 0.8482086062431335, + "learning_rate": 6.987448984063014e-06, + "loss": 0.0578, + "step": 1792 + }, + { + "epoch": 0.7782118055555556, + "grad_norm": 1.2565405368804932, + "learning_rate": 6.984223223470834e-06, + "loss": 0.0745, + "step": 1793 + }, + { + "epoch": 0.7786458333333334, + "grad_norm": 1.0909104347229004, + "learning_rate": 6.9809964823291e-06, + "loss": 0.0482, + "step": 1794 + }, + { + "epoch": 0.7790798611111112, + "grad_norm": 0.6057597398757935, + "learning_rate": 6.977768762232379e-06, + "loss": 0.0464, + "step": 1795 + }, + { + "epoch": 0.7795138888888888, + "grad_norm": 0.8874408602714539, + "learning_rate": 6.974540064775726e-06, + "loss": 0.0598, + "step": 1796 + }, + { + "epoch": 0.7799479166666666, + "grad_norm": 0.5312131643295288, + "learning_rate": 6.971310391554671e-06, + "loss": 0.0401, + "step": 1797 + }, + { + "epoch": 0.7803819444444444, + "grad_norm": 0.46000969409942627, + "learning_rate": 6.968079744165238e-06, + "loss": 0.0337, + "step": 1798 + }, + { + "epoch": 0.7808159722222222, + "grad_norm": 0.6547852158546448, + "learning_rate": 6.96484812420392e-06, + "loss": 0.0497, + "step": 1799 + }, + { + "epoch": 0.78125, + "grad_norm": 0.7223719358444214, + "learning_rate": 6.961615533267701e-06, + "loss": 0.0327, + "step": 1800 + }, + { + "epoch": 0.78125, + "eval_layout_rm_rank_test_accuracy": 0.9675206041167634, + "eval_layout_rm_rank_test_chosen_scores": 0.6561079025268555, + "eval_layout_rm_rank_test_loss": 0.04797226935625076, + "eval_layout_rm_rank_test_rejected_scores": 0.10819125175476074, + "eval_layout_rm_rank_test_runtime": 1182.7169, + "eval_layout_rm_rank_test_samples_per_second": 20.826, + "eval_layout_rm_rank_test_steps_per_second": 0.651, + "step": 1800 + }, + { + "epoch": 0.7816840277777778, + "grad_norm": 0.7490624785423279, + "learning_rate": 6.958381972954038e-06, + "loss": 0.0452, + "step": 1801 + }, + { + "epoch": 0.7821180555555556, + "grad_norm": 0.7055108547210693, + "learning_rate": 6.95514744486087e-06, + "loss": 0.0483, + "step": 1802 + }, + { + "epoch": 0.7825520833333334, + "grad_norm": 0.7814860939979553, + "learning_rate": 6.9519119505866125e-06, + "loss": 0.0674, + "step": 1803 + }, + { + "epoch": 0.7829861111111112, + "grad_norm": 0.649394690990448, + "learning_rate": 6.948675491730159e-06, + "loss": 0.0329, + "step": 1804 + }, + { + "epoch": 0.7834201388888888, + "grad_norm": 0.5071682929992676, + "learning_rate": 6.945438069890881e-06, + "loss": 0.0267, + "step": 1805 + }, + { + "epoch": 0.7838541666666666, + "grad_norm": 0.6328599452972412, + "learning_rate": 6.942199686668625e-06, + "loss": 0.0353, + "step": 1806 + }, + { + "epoch": 0.7842881944444444, + "grad_norm": 0.7604709267616272, + "learning_rate": 6.938960343663712e-06, + "loss": 0.0465, + "step": 1807 + }, + { + "epoch": 0.7847222222222222, + "grad_norm": 0.693663477897644, + "learning_rate": 6.935720042476939e-06, + "loss": 0.0404, + "step": 1808 + }, + { + "epoch": 0.78515625, + "grad_norm": 0.9226562976837158, + "learning_rate": 6.932478784709574e-06, + "loss": 0.034, + "step": 1809 + }, + { + "epoch": 0.7855902777777778, + "grad_norm": 1.2657535076141357, + "learning_rate": 6.9292365719633605e-06, + "loss": 0.0932, + "step": 1810 + }, + { + "epoch": 0.7860243055555556, + "grad_norm": 0.7438696026802063, + "learning_rate": 6.9259934058405124e-06, + "loss": 0.0393, + "step": 1811 + }, + { + "epoch": 0.7864583333333334, + "grad_norm": 0.6592585444450378, + "learning_rate": 6.922749287943711e-06, + "loss": 0.0406, + "step": 1812 + }, + { + "epoch": 0.7868923611111112, + "grad_norm": 0.8676554560661316, + "learning_rate": 6.919504219876117e-06, + "loss": 0.0396, + "step": 1813 + }, + { + "epoch": 0.7873263888888888, + "grad_norm": 0.7734280228614807, + "learning_rate": 6.916258203241355e-06, + "loss": 0.0504, + "step": 1814 + }, + { + "epoch": 0.7877604166666666, + "grad_norm": 0.8438120484352112, + "learning_rate": 6.913011239643516e-06, + "loss": 0.0368, + "step": 1815 + }, + { + "epoch": 0.7881944444444444, + "grad_norm": 0.975790798664093, + "learning_rate": 6.909763330687162e-06, + "loss": 0.0591, + "step": 1816 + }, + { + "epoch": 0.7886284722222222, + "grad_norm": 0.7018232345581055, + "learning_rate": 6.906514477977324e-06, + "loss": 0.0397, + "step": 1817 + }, + { + "epoch": 0.7890625, + "grad_norm": 0.914828360080719, + "learning_rate": 6.903264683119496e-06, + "loss": 0.0456, + "step": 1818 + }, + { + "epoch": 0.7894965277777778, + "grad_norm": 0.9112297296524048, + "learning_rate": 6.900013947719639e-06, + "loss": 0.0513, + "step": 1819 + }, + { + "epoch": 0.7899305555555556, + "grad_norm": 0.5047491788864136, + "learning_rate": 6.896762273384179e-06, + "loss": 0.0494, + "step": 1820 + }, + { + "epoch": 0.7903645833333334, + "grad_norm": 0.7444157004356384, + "learning_rate": 6.893509661720005e-06, + "loss": 0.0515, + "step": 1821 + }, + { + "epoch": 0.7907986111111112, + "grad_norm": 0.61995929479599, + "learning_rate": 6.890256114334472e-06, + "loss": 0.0615, + "step": 1822 + }, + { + "epoch": 0.7912326388888888, + "grad_norm": 1.0646355152130127, + "learning_rate": 6.887001632835392e-06, + "loss": 0.0932, + "step": 1823 + }, + { + "epoch": 0.7916666666666666, + "grad_norm": 1.0587164163589478, + "learning_rate": 6.883746218831045e-06, + "loss": 0.0614, + "step": 1824 + }, + { + "epoch": 0.7921006944444444, + "grad_norm": 0.7038317322731018, + "learning_rate": 6.8804898739301654e-06, + "loss": 0.0428, + "step": 1825 + }, + { + "epoch": 0.7925347222222222, + "grad_norm": 0.5331775546073914, + "learning_rate": 6.877232599741953e-06, + "loss": 0.0479, + "step": 1826 + }, + { + "epoch": 0.79296875, + "grad_norm": 0.43616053462028503, + "learning_rate": 6.873974397876067e-06, + "loss": 0.0347, + "step": 1827 + }, + { + "epoch": 0.7934027777777778, + "grad_norm": 0.5002326369285583, + "learning_rate": 6.8707152699426184e-06, + "loss": 0.0447, + "step": 1828 + }, + { + "epoch": 0.7938368055555556, + "grad_norm": 0.6660779714584351, + "learning_rate": 6.867455217552183e-06, + "loss": 0.0476, + "step": 1829 + }, + { + "epoch": 0.7942708333333334, + "grad_norm": 0.5385332703590393, + "learning_rate": 6.864194242315788e-06, + "loss": 0.0496, + "step": 1830 + }, + { + "epoch": 0.7947048611111112, + "grad_norm": 0.5744720697402954, + "learning_rate": 6.8609323458449236e-06, + "loss": 0.047, + "step": 1831 + }, + { + "epoch": 0.7951388888888888, + "grad_norm": 0.7200570106506348, + "learning_rate": 6.857669529751529e-06, + "loss": 0.0519, + "step": 1832 + }, + { + "epoch": 0.7955729166666666, + "grad_norm": 0.5372190475463867, + "learning_rate": 6.854405795647997e-06, + "loss": 0.0387, + "step": 1833 + }, + { + "epoch": 0.7960069444444444, + "grad_norm": 0.5767813324928284, + "learning_rate": 6.851141145147181e-06, + "loss": 0.0334, + "step": 1834 + }, + { + "epoch": 0.7964409722222222, + "grad_norm": 1.3028239011764526, + "learning_rate": 6.847875579862382e-06, + "loss": 0.0646, + "step": 1835 + }, + { + "epoch": 0.796875, + "grad_norm": 0.8541083931922913, + "learning_rate": 6.844609101407354e-06, + "loss": 0.0516, + "step": 1836 + }, + { + "epoch": 0.7973090277777778, + "grad_norm": 0.6067532300949097, + "learning_rate": 6.841341711396301e-06, + "loss": 0.0381, + "step": 1837 + }, + { + "epoch": 0.7977430555555556, + "grad_norm": 0.9461897015571594, + "learning_rate": 6.838073411443882e-06, + "loss": 0.0478, + "step": 1838 + }, + { + "epoch": 0.7981770833333334, + "grad_norm": 1.1896733045578003, + "learning_rate": 6.834804203165199e-06, + "loss": 0.0594, + "step": 1839 + }, + { + "epoch": 0.7986111111111112, + "grad_norm": 0.8283300995826721, + "learning_rate": 6.831534088175807e-06, + "loss": 0.0567, + "step": 1840 + }, + { + "epoch": 0.7990451388888888, + "grad_norm": 0.8821940422058105, + "learning_rate": 6.828263068091711e-06, + "loss": 0.0367, + "step": 1841 + }, + { + "epoch": 0.7994791666666666, + "grad_norm": 0.7667244672775269, + "learning_rate": 6.824991144529362e-06, + "loss": 0.054, + "step": 1842 + }, + { + "epoch": 0.7999131944444444, + "grad_norm": 0.5846320986747742, + "learning_rate": 6.8217183191056524e-06, + "loss": 0.0365, + "step": 1843 + }, + { + "epoch": 0.8003472222222222, + "grad_norm": 0.6740960478782654, + "learning_rate": 6.8184445934379255e-06, + "loss": 0.0477, + "step": 1844 + }, + { + "epoch": 0.80078125, + "grad_norm": 0.7076245546340942, + "learning_rate": 6.815169969143969e-06, + "loss": 0.0455, + "step": 1845 + }, + { + "epoch": 0.8012152777777778, + "grad_norm": 0.6552544236183167, + "learning_rate": 6.811894447842012e-06, + "loss": 0.0389, + "step": 1846 + }, + { + "epoch": 0.8016493055555556, + "grad_norm": 0.912534773349762, + "learning_rate": 6.80861803115073e-06, + "loss": 0.094, + "step": 1847 + }, + { + "epoch": 0.8020833333333334, + "grad_norm": 0.5448792576789856, + "learning_rate": 6.805340720689241e-06, + "loss": 0.0419, + "step": 1848 + }, + { + "epoch": 0.8025173611111112, + "grad_norm": 0.6828315258026123, + "learning_rate": 6.802062518077101e-06, + "loss": 0.0493, + "step": 1849 + }, + { + "epoch": 0.8029513888888888, + "grad_norm": 0.5973538756370544, + "learning_rate": 6.7987834249343096e-06, + "loss": 0.0523, + "step": 1850 + }, + { + "epoch": 0.8033854166666666, + "grad_norm": 0.7848672866821289, + "learning_rate": 6.795503442881307e-06, + "loss": 0.0468, + "step": 1851 + }, + { + "epoch": 0.8038194444444444, + "grad_norm": 0.4816887676715851, + "learning_rate": 6.7922225735389726e-06, + "loss": 0.0414, + "step": 1852 + }, + { + "epoch": 0.8042534722222222, + "grad_norm": 1.0304285287857056, + "learning_rate": 6.788940818528622e-06, + "loss": 0.0619, + "step": 1853 + }, + { + "epoch": 0.8046875, + "grad_norm": 0.5743433833122253, + "learning_rate": 6.78565817947201e-06, + "loss": 0.0493, + "step": 1854 + }, + { + "epoch": 0.8051215277777778, + "grad_norm": 0.6866853833198547, + "learning_rate": 6.782374657991329e-06, + "loss": 0.0394, + "step": 1855 + }, + { + "epoch": 0.8055555555555556, + "grad_norm": 0.8010555505752563, + "learning_rate": 6.779090255709205e-06, + "loss": 0.0506, + "step": 1856 + }, + { + "epoch": 0.8059895833333334, + "grad_norm": 0.6958445906639099, + "learning_rate": 6.775804974248703e-06, + "loss": 0.0535, + "step": 1857 + }, + { + "epoch": 0.8064236111111112, + "grad_norm": 0.6616341471672058, + "learning_rate": 6.77251881523332e-06, + "loss": 0.0438, + "step": 1858 + }, + { + "epoch": 0.8068576388888888, + "grad_norm": 0.6229743957519531, + "learning_rate": 6.769231780286987e-06, + "loss": 0.0401, + "step": 1859 + }, + { + "epoch": 0.8072916666666666, + "grad_norm": 0.7217589020729065, + "learning_rate": 6.7659438710340654e-06, + "loss": 0.054, + "step": 1860 + }, + { + "epoch": 0.8077256944444444, + "grad_norm": 0.5400440096855164, + "learning_rate": 6.762655089099353e-06, + "loss": 0.0401, + "step": 1861 + }, + { + "epoch": 0.8081597222222222, + "grad_norm": 1.5119469165802002, + "learning_rate": 6.759365436108079e-06, + "loss": 0.0661, + "step": 1862 + }, + { + "epoch": 0.80859375, + "grad_norm": 0.8494582772254944, + "learning_rate": 6.7560749136859005e-06, + "loss": 0.0482, + "step": 1863 + }, + { + "epoch": 0.8090277777777778, + "grad_norm": 2.100743055343628, + "learning_rate": 6.752783523458903e-06, + "loss": 0.0657, + "step": 1864 + }, + { + "epoch": 0.8094618055555556, + "grad_norm": 0.5947306156158447, + "learning_rate": 6.749491267053604e-06, + "loss": 0.0436, + "step": 1865 + }, + { + "epoch": 0.8098958333333334, + "grad_norm": 0.47346755862236023, + "learning_rate": 6.746198146096947e-06, + "loss": 0.0392, + "step": 1866 + }, + { + "epoch": 0.8103298611111112, + "grad_norm": 0.7026275992393494, + "learning_rate": 6.742904162216304e-06, + "loss": 0.0376, + "step": 1867 + }, + { + "epoch": 0.8107638888888888, + "grad_norm": 0.6437444686889648, + "learning_rate": 6.739609317039471e-06, + "loss": 0.0507, + "step": 1868 + }, + { + "epoch": 0.8111979166666666, + "grad_norm": 0.5360928177833557, + "learning_rate": 6.736313612194676e-06, + "loss": 0.0499, + "step": 1869 + }, + { + "epoch": 0.8116319444444444, + "grad_norm": 0.6127115488052368, + "learning_rate": 6.733017049310563e-06, + "loss": 0.0353, + "step": 1870 + }, + { + "epoch": 0.8120659722222222, + "grad_norm": 0.7498536109924316, + "learning_rate": 6.729719630016206e-06, + "loss": 0.0312, + "step": 1871 + }, + { + "epoch": 0.8125, + "grad_norm": 0.8961223363876343, + "learning_rate": 6.7264213559411e-06, + "loss": 0.0715, + "step": 1872 + }, + { + "epoch": 0.8129340277777778, + "grad_norm": 0.6380618214607239, + "learning_rate": 6.723122228715167e-06, + "loss": 0.0335, + "step": 1873 + }, + { + "epoch": 0.8133680555555556, + "grad_norm": 0.7774998545646667, + "learning_rate": 6.719822249968739e-06, + "loss": 0.0327, + "step": 1874 + }, + { + "epoch": 0.8138020833333334, + "grad_norm": 0.8536900281906128, + "learning_rate": 6.716521421332582e-06, + "loss": 0.0489, + "step": 1875 + }, + { + "epoch": 0.8142361111111112, + "grad_norm": 1.3191429376602173, + "learning_rate": 6.713219744437876e-06, + "loss": 0.1005, + "step": 1876 + }, + { + "epoch": 0.8146701388888888, + "grad_norm": 0.8960707783699036, + "learning_rate": 6.70991722091622e-06, + "loss": 0.0474, + "step": 1877 + }, + { + "epoch": 0.8151041666666666, + "grad_norm": 0.7469573020935059, + "learning_rate": 6.706613852399632e-06, + "loss": 0.0513, + "step": 1878 + }, + { + "epoch": 0.8155381944444444, + "grad_norm": 0.6806800365447998, + "learning_rate": 6.703309640520546e-06, + "loss": 0.0354, + "step": 1879 + }, + { + "epoch": 0.8159722222222222, + "grad_norm": 0.880378246307373, + "learning_rate": 6.7000045869118154e-06, + "loss": 0.0615, + "step": 1880 + }, + { + "epoch": 0.81640625, + "grad_norm": 0.8611704707145691, + "learning_rate": 6.696698693206708e-06, + "loss": 0.0499, + "step": 1881 + }, + { + "epoch": 0.8168402777777778, + "grad_norm": 0.6044244170188904, + "learning_rate": 6.6933919610389086e-06, + "loss": 0.0374, + "step": 1882 + }, + { + "epoch": 0.8172743055555556, + "grad_norm": 0.675598680973053, + "learning_rate": 6.690084392042514e-06, + "loss": 0.0549, + "step": 1883 + }, + { + "epoch": 0.8177083333333334, + "grad_norm": 0.8580988645553589, + "learning_rate": 6.686775987852037e-06, + "loss": 0.0593, + "step": 1884 + }, + { + "epoch": 0.8181423611111112, + "grad_norm": 0.4779009521007538, + "learning_rate": 6.683466750102399e-06, + "loss": 0.039, + "step": 1885 + }, + { + "epoch": 0.8185763888888888, + "grad_norm": 0.6775459051132202, + "learning_rate": 6.680156680428938e-06, + "loss": 0.0457, + "step": 1886 + }, + { + "epoch": 0.8190104166666666, + "grad_norm": 0.649315357208252, + "learning_rate": 6.6768457804674e-06, + "loss": 0.05, + "step": 1887 + }, + { + "epoch": 0.8194444444444444, + "grad_norm": 0.7105356454849243, + "learning_rate": 6.673534051853942e-06, + "loss": 0.0519, + "step": 1888 + }, + { + "epoch": 0.8198784722222222, + "grad_norm": 0.5776419043540955, + "learning_rate": 6.670221496225133e-06, + "loss": 0.0439, + "step": 1889 + }, + { + "epoch": 0.8203125, + "grad_norm": 0.6759132146835327, + "learning_rate": 6.666908115217947e-06, + "loss": 0.0544, + "step": 1890 + }, + { + "epoch": 0.8207465277777778, + "grad_norm": 0.6925734281539917, + "learning_rate": 6.6635939104697655e-06, + "loss": 0.039, + "step": 1891 + }, + { + "epoch": 0.8211805555555556, + "grad_norm": 0.3945949673652649, + "learning_rate": 6.660278883618384e-06, + "loss": 0.0246, + "step": 1892 + }, + { + "epoch": 0.8216145833333334, + "grad_norm": 1.0184353590011597, + "learning_rate": 6.656963036301995e-06, + "loss": 0.0539, + "step": 1893 + }, + { + "epoch": 0.8220486111111112, + "grad_norm": 0.7225958108901978, + "learning_rate": 6.653646370159204e-06, + "loss": 0.0419, + "step": 1894 + }, + { + "epoch": 0.8224826388888888, + "grad_norm": 0.8667299151420593, + "learning_rate": 6.650328886829015e-06, + "loss": 0.0489, + "step": 1895 + }, + { + "epoch": 0.8229166666666666, + "grad_norm": 0.597955584526062, + "learning_rate": 6.647010587950838e-06, + "loss": 0.0338, + "step": 1896 + }, + { + "epoch": 0.8233506944444444, + "grad_norm": 0.8427351117134094, + "learning_rate": 6.643691475164491e-06, + "loss": 0.0369, + "step": 1897 + }, + { + "epoch": 0.8237847222222222, + "grad_norm": 0.7786532640457153, + "learning_rate": 6.6403715501101875e-06, + "loss": 0.0411, + "step": 1898 + }, + { + "epoch": 0.82421875, + "grad_norm": 0.8530535101890564, + "learning_rate": 6.637050814428544e-06, + "loss": 0.0494, + "step": 1899 + }, + { + "epoch": 0.8246527777777778, + "grad_norm": 1.1018110513687134, + "learning_rate": 6.6337292697605796e-06, + "loss": 0.0582, + "step": 1900 + }, + { + "epoch": 0.8250868055555556, + "grad_norm": 0.9957329630851746, + "learning_rate": 6.63040691774771e-06, + "loss": 0.0593, + "step": 1901 + }, + { + "epoch": 0.8255208333333334, + "grad_norm": 1.0684702396392822, + "learning_rate": 6.627083760031755e-06, + "loss": 0.0597, + "step": 1902 + }, + { + "epoch": 0.8259548611111112, + "grad_norm": 1.0910362005233765, + "learning_rate": 6.623759798254927e-06, + "loss": 0.0489, + "step": 1903 + }, + { + "epoch": 0.8263888888888888, + "grad_norm": 0.6104505062103271, + "learning_rate": 6.62043503405984e-06, + "loss": 0.0385, + "step": 1904 + }, + { + "epoch": 0.8268229166666666, + "grad_norm": 1.3332817554473877, + "learning_rate": 6.617109469089502e-06, + "loss": 0.0629, + "step": 1905 + }, + { + "epoch": 0.8272569444444444, + "grad_norm": 0.4675104022026062, + "learning_rate": 6.613783104987318e-06, + "loss": 0.0468, + "step": 1906 + }, + { + "epoch": 0.8276909722222222, + "grad_norm": 0.4829596281051636, + "learning_rate": 6.610455943397086e-06, + "loss": 0.036, + "step": 1907 + }, + { + "epoch": 0.828125, + "grad_norm": 0.6301221251487732, + "learning_rate": 6.607127985963e-06, + "loss": 0.0441, + "step": 1908 + }, + { + "epoch": 0.8285590277777778, + "grad_norm": 0.8131287097930908, + "learning_rate": 6.603799234329646e-06, + "loss": 0.0526, + "step": 1909 + }, + { + "epoch": 0.8289930555555556, + "grad_norm": 0.43327629566192627, + "learning_rate": 6.600469690142005e-06, + "loss": 0.0357, + "step": 1910 + }, + { + "epoch": 0.8294270833333334, + "grad_norm": 0.3900514841079712, + "learning_rate": 6.597139355045448e-06, + "loss": 0.0245, + "step": 1911 + }, + { + "epoch": 0.8298611111111112, + "grad_norm": 0.6519450545310974, + "learning_rate": 6.593808230685733e-06, + "loss": 0.0433, + "step": 1912 + }, + { + "epoch": 0.8302951388888888, + "grad_norm": 0.6983678340911865, + "learning_rate": 6.590476318709016e-06, + "loss": 0.0445, + "step": 1913 + }, + { + "epoch": 0.8307291666666666, + "grad_norm": 1.0150282382965088, + "learning_rate": 6.587143620761835e-06, + "loss": 0.0482, + "step": 1914 + }, + { + "epoch": 0.8311631944444444, + "grad_norm": 0.6016936898231506, + "learning_rate": 6.583810138491122e-06, + "loss": 0.0355, + "step": 1915 + }, + { + "epoch": 0.8315972222222222, + "grad_norm": 0.873637318611145, + "learning_rate": 6.580475873544193e-06, + "loss": 0.0539, + "step": 1916 + }, + { + "epoch": 0.83203125, + "grad_norm": 0.7255651354789734, + "learning_rate": 6.57714082756875e-06, + "loss": 0.0502, + "step": 1917 + }, + { + "epoch": 0.8324652777777778, + "grad_norm": 0.9832391738891602, + "learning_rate": 6.573805002212884e-06, + "loss": 0.0773, + "step": 1918 + }, + { + "epoch": 0.8328993055555556, + "grad_norm": 0.8161395192146301, + "learning_rate": 6.570468399125071e-06, + "loss": 0.0571, + "step": 1919 + }, + { + "epoch": 0.8333333333333334, + "grad_norm": 0.6000357270240784, + "learning_rate": 6.567131019954168e-06, + "loss": 0.0406, + "step": 1920 + }, + { + "epoch": 0.8337673611111112, + "grad_norm": 0.5935305953025818, + "learning_rate": 6.56379286634942e-06, + "loss": 0.0428, + "step": 1921 + }, + { + "epoch": 0.8342013888888888, + "grad_norm": 0.9758431315422058, + "learning_rate": 6.56045393996045e-06, + "loss": 0.0543, + "step": 1922 + }, + { + "epoch": 0.8346354166666666, + "grad_norm": 0.7438514232635498, + "learning_rate": 6.557114242437266e-06, + "loss": 0.0494, + "step": 1923 + }, + { + "epoch": 0.8350694444444444, + "grad_norm": 0.5088344216346741, + "learning_rate": 6.553773775430256e-06, + "loss": 0.0467, + "step": 1924 + }, + { + "epoch": 0.8355034722222222, + "grad_norm": 0.7656986713409424, + "learning_rate": 6.550432540590192e-06, + "loss": 0.0609, + "step": 1925 + }, + { + "epoch": 0.8359375, + "grad_norm": 0.5904448628425598, + "learning_rate": 6.5470905395682186e-06, + "loss": 0.0364, + "step": 1926 + }, + { + "epoch": 0.8363715277777778, + "grad_norm": 0.7087653875350952, + "learning_rate": 6.543747774015863e-06, + "loss": 0.0426, + "step": 1927 + }, + { + "epoch": 0.8368055555555556, + "grad_norm": 0.838172435760498, + "learning_rate": 6.5404042455850306e-06, + "loss": 0.0612, + "step": 1928 + }, + { + "epoch": 0.8372395833333334, + "grad_norm": 0.5433200597763062, + "learning_rate": 6.537059955928003e-06, + "loss": 0.0479, + "step": 1929 + }, + { + "epoch": 0.8376736111111112, + "grad_norm": 0.5060282349586487, + "learning_rate": 6.5337149066974345e-06, + "loss": 0.0388, + "step": 1930 + }, + { + "epoch": 0.8381076388888888, + "grad_norm": 0.6122724413871765, + "learning_rate": 6.530369099546363e-06, + "loss": 0.042, + "step": 1931 + }, + { + "epoch": 0.8385416666666666, + "grad_norm": 0.9471703767776489, + "learning_rate": 6.5270225361281925e-06, + "loss": 0.0543, + "step": 1932 + }, + { + "epoch": 0.8389756944444444, + "grad_norm": 0.8253403306007385, + "learning_rate": 6.523675218096707e-06, + "loss": 0.0689, + "step": 1933 + }, + { + "epoch": 0.8394097222222222, + "grad_norm": 1.0257596969604492, + "learning_rate": 6.520327147106059e-06, + "loss": 0.062, + "step": 1934 + }, + { + "epoch": 0.83984375, + "grad_norm": 0.5009595155715942, + "learning_rate": 6.516978324810777e-06, + "loss": 0.0353, + "step": 1935 + }, + { + "epoch": 0.8402777777777778, + "grad_norm": 0.7189043760299683, + "learning_rate": 6.513628752865755e-06, + "loss": 0.0679, + "step": 1936 + }, + { + "epoch": 0.8407118055555556, + "grad_norm": 0.9218528270721436, + "learning_rate": 6.5102784329262626e-06, + "loss": 0.051, + "step": 1937 + }, + { + "epoch": 0.8411458333333334, + "grad_norm": 0.6909580230712891, + "learning_rate": 6.506927366647937e-06, + "loss": 0.0682, + "step": 1938 + }, + { + "epoch": 0.8415798611111112, + "grad_norm": 0.5428255796432495, + "learning_rate": 6.503575555686787e-06, + "loss": 0.042, + "step": 1939 + }, + { + "epoch": 0.8420138888888888, + "grad_norm": 0.5663094520568848, + "learning_rate": 6.500223001699187e-06, + "loss": 0.0321, + "step": 1940 + }, + { + "epoch": 0.8424479166666666, + "grad_norm": 0.5367885231971741, + "learning_rate": 6.496869706341875e-06, + "loss": 0.0563, + "step": 1941 + }, + { + "epoch": 0.8428819444444444, + "grad_norm": 0.682534396648407, + "learning_rate": 6.4935156712719615e-06, + "loss": 0.0369, + "step": 1942 + }, + { + "epoch": 0.8433159722222222, + "grad_norm": 0.9464619159698486, + "learning_rate": 6.490160898146919e-06, + "loss": 0.0597, + "step": 1943 + }, + { + "epoch": 0.84375, + "grad_norm": 0.6500828862190247, + "learning_rate": 6.4868053886245885e-06, + "loss": 0.0506, + "step": 1944 + }, + { + "epoch": 0.8441840277777778, + "grad_norm": 0.5619100332260132, + "learning_rate": 6.483449144363169e-06, + "loss": 0.0408, + "step": 1945 + }, + { + "epoch": 0.8446180555555556, + "grad_norm": 0.41894733905792236, + "learning_rate": 6.480092167021228e-06, + "loss": 0.0241, + "step": 1946 + }, + { + "epoch": 0.8450520833333334, + "grad_norm": 0.7337756752967834, + "learning_rate": 6.476734458257692e-06, + "loss": 0.0448, + "step": 1947 + }, + { + "epoch": 0.8454861111111112, + "grad_norm": 0.7293131947517395, + "learning_rate": 6.473376019731851e-06, + "loss": 0.0435, + "step": 1948 + }, + { + "epoch": 0.8459201388888888, + "grad_norm": 0.6051151752471924, + "learning_rate": 6.470016853103353e-06, + "loss": 0.0513, + "step": 1949 + }, + { + "epoch": 0.8463541666666666, + "grad_norm": 0.7551562190055847, + "learning_rate": 6.466656960032209e-06, + "loss": 0.0318, + "step": 1950 + }, + { + "epoch": 0.8467881944444444, + "grad_norm": 0.818610668182373, + "learning_rate": 6.463296342178787e-06, + "loss": 0.0586, + "step": 1951 + }, + { + "epoch": 0.8472222222222222, + "grad_norm": 0.9881995320320129, + "learning_rate": 6.459935001203812e-06, + "loss": 0.0387, + "step": 1952 + }, + { + "epoch": 0.84765625, + "grad_norm": 0.7172938585281372, + "learning_rate": 6.456572938768371e-06, + "loss": 0.0432, + "step": 1953 + }, + { + "epoch": 0.8480902777777778, + "grad_norm": 0.7845926880836487, + "learning_rate": 6.453210156533902e-06, + "loss": 0.0489, + "step": 1954 + }, + { + "epoch": 0.8485243055555556, + "grad_norm": 2.5467822551727295, + "learning_rate": 6.449846656162205e-06, + "loss": 0.0298, + "step": 1955 + }, + { + "epoch": 0.8489583333333334, + "grad_norm": 1.1665350198745728, + "learning_rate": 6.446482439315426e-06, + "loss": 0.0506, + "step": 1956 + }, + { + "epoch": 0.8493923611111112, + "grad_norm": 1.032641053199768, + "learning_rate": 6.443117507656072e-06, + "loss": 0.044, + "step": 1957 + }, + { + "epoch": 0.8498263888888888, + "grad_norm": 0.5748589038848877, + "learning_rate": 6.439751862847002e-06, + "loss": 0.0415, + "step": 1958 + }, + { + "epoch": 0.8502604166666666, + "grad_norm": 0.5190004706382751, + "learning_rate": 6.436385506551428e-06, + "loss": 0.0369, + "step": 1959 + }, + { + "epoch": 0.8506944444444444, + "grad_norm": 0.5892989039421082, + "learning_rate": 6.43301844043291e-06, + "loss": 0.0316, + "step": 1960 + }, + { + "epoch": 0.8511284722222222, + "grad_norm": 0.8164686560630798, + "learning_rate": 6.429650666155363e-06, + "loss": 0.0389, + "step": 1961 + }, + { + "epoch": 0.8515625, + "grad_norm": 0.7017279267311096, + "learning_rate": 6.426282185383049e-06, + "loss": 0.0376, + "step": 1962 + }, + { + "epoch": 0.8519965277777778, + "grad_norm": 0.5410183668136597, + "learning_rate": 6.422912999780579e-06, + "loss": 0.0429, + "step": 1963 + }, + { + "epoch": 0.8524305555555556, + "grad_norm": 1.0295459032058716, + "learning_rate": 6.4195431110129184e-06, + "loss": 0.0523, + "step": 1964 + }, + { + "epoch": 0.8528645833333334, + "grad_norm": 0.6248032450675964, + "learning_rate": 6.4161725207453694e-06, + "loss": 0.027, + "step": 1965 + }, + { + "epoch": 0.8532986111111112, + "grad_norm": 0.671219527721405, + "learning_rate": 6.41280123064359e-06, + "loss": 0.0543, + "step": 1966 + }, + { + "epoch": 0.8537326388888888, + "grad_norm": 1.6697989702224731, + "learning_rate": 6.409429242373581e-06, + "loss": 0.0504, + "step": 1967 + }, + { + "epoch": 0.8541666666666666, + "grad_norm": 1.516735315322876, + "learning_rate": 6.4060565576016875e-06, + "loss": 0.0404, + "step": 1968 + }, + { + "epoch": 0.8546006944444444, + "grad_norm": 3.2779290676116943, + "learning_rate": 6.402683177994599e-06, + "loss": 0.0635, + "step": 1969 + }, + { + "epoch": 0.8550347222222222, + "grad_norm": 1.1508862972259521, + "learning_rate": 6.399309105219347e-06, + "loss": 0.0595, + "step": 1970 + }, + { + "epoch": 0.85546875, + "grad_norm": 0.636896014213562, + "learning_rate": 6.395934340943311e-06, + "loss": 0.0267, + "step": 1971 + }, + { + "epoch": 0.8559027777777778, + "grad_norm": 0.9543007016181946, + "learning_rate": 6.3925588868342035e-06, + "loss": 0.0464, + "step": 1972 + }, + { + "epoch": 0.8563368055555556, + "grad_norm": 2.7885780334472656, + "learning_rate": 6.389182744560085e-06, + "loss": 0.0358, + "step": 1973 + }, + { + "epoch": 0.8567708333333334, + "grad_norm": 1.1232353448867798, + "learning_rate": 6.385805915789356e-06, + "loss": 0.0319, + "step": 1974 + }, + { + "epoch": 0.8572048611111112, + "grad_norm": 17.793001174926758, + "learning_rate": 6.382428402190753e-06, + "loss": 0.0765, + "step": 1975 + }, + { + "epoch": 0.8576388888888888, + "grad_norm": 5.3929443359375, + "learning_rate": 6.3790502054333515e-06, + "loss": 0.0385, + "step": 1976 + }, + { + "epoch": 0.8580729166666666, + "grad_norm": 2.506985902786255, + "learning_rate": 6.375671327186565e-06, + "loss": 0.0561, + "step": 1977 + }, + { + "epoch": 0.8585069444444444, + "grad_norm": 1.9946200847625732, + "learning_rate": 6.372291769120143e-06, + "loss": 0.0296, + "step": 1978 + }, + { + "epoch": 0.8589409722222222, + "grad_norm": 3.731701374053955, + "learning_rate": 6.368911532904173e-06, + "loss": 0.0384, + "step": 1979 + }, + { + "epoch": 0.859375, + "grad_norm": 1.0285123586654663, + "learning_rate": 6.365530620209077e-06, + "loss": 0.047, + "step": 1980 + }, + { + "epoch": 0.8598090277777778, + "grad_norm": 14.954253196716309, + "learning_rate": 6.36214903270561e-06, + "loss": 0.0549, + "step": 1981 + }, + { + "epoch": 0.8602430555555556, + "grad_norm": 4.2624125480651855, + "learning_rate": 6.3587667720648626e-06, + "loss": 0.0373, + "step": 1982 + }, + { + "epoch": 0.8606770833333334, + "grad_norm": 5.046442031860352, + "learning_rate": 6.355383839958255e-06, + "loss": 0.0531, + "step": 1983 + }, + { + "epoch": 0.8611111111111112, + "grad_norm": 3.881274938583374, + "learning_rate": 6.3520002380575395e-06, + "loss": 0.0815, + "step": 1984 + }, + { + "epoch": 0.8615451388888888, + "grad_norm": 5.472319602966309, + "learning_rate": 6.3486159680348046e-06, + "loss": 0.0767, + "step": 1985 + }, + { + "epoch": 0.8619791666666666, + "grad_norm": 3.93212890625, + "learning_rate": 6.345231031562464e-06, + "loss": 0.0518, + "step": 1986 + }, + { + "epoch": 0.8624131944444444, + "grad_norm": 5.016967296600342, + "learning_rate": 6.341845430313259e-06, + "loss": 0.0534, + "step": 1987 + }, + { + "epoch": 0.8628472222222222, + "grad_norm": 2.7708795070648193, + "learning_rate": 6.338459165960266e-06, + "loss": 0.0419, + "step": 1988 + }, + { + "epoch": 0.86328125, + "grad_norm": 1.2714332342147827, + "learning_rate": 6.335072240176885e-06, + "loss": 0.0386, + "step": 1989 + }, + { + "epoch": 0.8637152777777778, + "grad_norm": 3.638699769973755, + "learning_rate": 6.331684654636843e-06, + "loss": 0.0483, + "step": 1990 + }, + { + "epoch": 0.8641493055555556, + "grad_norm": 2.74619197845459, + "learning_rate": 6.328296411014191e-06, + "loss": 0.0311, + "step": 1991 + }, + { + "epoch": 0.8645833333333334, + "grad_norm": 2.6659770011901855, + "learning_rate": 6.32490751098331e-06, + "loss": 0.0388, + "step": 1992 + }, + { + "epoch": 0.8650173611111112, + "grad_norm": 15.077685356140137, + "learning_rate": 6.321517956218901e-06, + "loss": 0.0531, + "step": 1993 + }, + { + "epoch": 0.8654513888888888, + "grad_norm": 3.0722298622131348, + "learning_rate": 6.318127748395992e-06, + "loss": 0.05, + "step": 1994 + }, + { + "epoch": 0.8658854166666666, + "grad_norm": 3.180548667907715, + "learning_rate": 6.314736889189932e-06, + "loss": 0.0521, + "step": 1995 + }, + { + "epoch": 0.8663194444444444, + "grad_norm": 2.55326771736145, + "learning_rate": 6.311345380276393e-06, + "loss": 0.0493, + "step": 1996 + }, + { + "epoch": 0.8667534722222222, + "grad_norm": 3.489673137664795, + "learning_rate": 6.307953223331366e-06, + "loss": 0.0536, + "step": 1997 + }, + { + "epoch": 0.8671875, + "grad_norm": 1.6632792949676514, + "learning_rate": 6.304560420031163e-06, + "loss": 0.0673, + "step": 1998 + }, + { + "epoch": 0.8676215277777778, + "grad_norm": 3.4524927139282227, + "learning_rate": 6.301166972052416e-06, + "loss": 0.0699, + "step": 1999 + }, + { + "epoch": 0.8680555555555556, + "grad_norm": 2.278571367263794, + "learning_rate": 6.297772881072075e-06, + "loss": 0.0658, + "step": 2000 + }, + { + "epoch": 0.8684895833333334, + "grad_norm": 2.340599298477173, + "learning_rate": 6.29437814876741e-06, + "loss": 0.0627, + "step": 2001 + }, + { + "epoch": 0.8689236111111112, + "grad_norm": 1.5224546194076538, + "learning_rate": 6.290982776816005e-06, + "loss": 0.0373, + "step": 2002 + }, + { + "epoch": 0.8693576388888888, + "grad_norm": 6.141655445098877, + "learning_rate": 6.2875867668957604e-06, + "loss": 0.0631, + "step": 2003 + }, + { + "epoch": 0.8697916666666666, + "grad_norm": 1.6914006471633911, + "learning_rate": 6.284190120684892e-06, + "loss": 0.0556, + "step": 2004 + }, + { + "epoch": 0.8702256944444444, + "grad_norm": 2.2304089069366455, + "learning_rate": 6.280792839861935e-06, + "loss": 0.0217, + "step": 2005 + }, + { + "epoch": 0.8706597222222222, + "grad_norm": 1.3991873264312744, + "learning_rate": 6.277394926105729e-06, + "loss": 0.046, + "step": 2006 + }, + { + "epoch": 0.87109375, + "grad_norm": 1.3106783628463745, + "learning_rate": 6.2739963810954325e-06, + "loss": 0.0483, + "step": 2007 + }, + { + "epoch": 0.8715277777777778, + "grad_norm": 0.8954133987426758, + "learning_rate": 6.270597206510516e-06, + "loss": 0.0584, + "step": 2008 + }, + { + "epoch": 0.8719618055555556, + "grad_norm": 0.5888851881027222, + "learning_rate": 6.2671974040307585e-06, + "loss": 0.0375, + "step": 2009 + }, + { + "epoch": 0.8723958333333334, + "grad_norm": 9.936369895935059, + "learning_rate": 6.2637969753362515e-06, + "loss": 0.0443, + "step": 2010 + }, + { + "epoch": 0.8728298611111112, + "grad_norm": 1.5567840337753296, + "learning_rate": 6.260395922107397e-06, + "loss": 0.0389, + "step": 2011 + }, + { + "epoch": 0.8732638888888888, + "grad_norm": 0.9838268756866455, + "learning_rate": 6.256994246024899e-06, + "loss": 0.0556, + "step": 2012 + }, + { + "epoch": 0.8736979166666666, + "grad_norm": 5.1909708976745605, + "learning_rate": 6.253591948769777e-06, + "loss": 0.0542, + "step": 2013 + }, + { + "epoch": 0.8741319444444444, + "grad_norm": 0.4206882119178772, + "learning_rate": 6.250189032023353e-06, + "loss": 0.0216, + "step": 2014 + }, + { + "epoch": 0.8745659722222222, + "grad_norm": 1.2391408681869507, + "learning_rate": 6.246785497467257e-06, + "loss": 0.0398, + "step": 2015 + }, + { + "epoch": 0.875, + "grad_norm": 0.8819404244422913, + "learning_rate": 6.243381346783426e-06, + "loss": 0.0544, + "step": 2016 + }, + { + "epoch": 0.8754340277777778, + "grad_norm": 0.6158362627029419, + "learning_rate": 6.239976581654098e-06, + "loss": 0.0337, + "step": 2017 + }, + { + "epoch": 0.8758680555555556, + "grad_norm": 1.3593721389770508, + "learning_rate": 6.236571203761815e-06, + "loss": 0.0291, + "step": 2018 + }, + { + "epoch": 0.8763020833333334, + "grad_norm": 1.1812952756881714, + "learning_rate": 6.233165214789424e-06, + "loss": 0.0333, + "step": 2019 + }, + { + "epoch": 0.8767361111111112, + "grad_norm": 0.7267976403236389, + "learning_rate": 6.229758616420073e-06, + "loss": 0.0264, + "step": 2020 + }, + { + "epoch": 0.8771701388888888, + "grad_norm": 1.3565468788146973, + "learning_rate": 6.226351410337209e-06, + "loss": 0.0292, + "step": 2021 + }, + { + "epoch": 0.8776041666666666, + "grad_norm": 0.8340696096420288, + "learning_rate": 6.222943598224583e-06, + "loss": 0.0287, + "step": 2022 + }, + { + "epoch": 0.8780381944444444, + "grad_norm": 1.3752747774124146, + "learning_rate": 6.219535181766244e-06, + "loss": 0.0588, + "step": 2023 + }, + { + "epoch": 0.8784722222222222, + "grad_norm": 0.6889845132827759, + "learning_rate": 6.216126162646539e-06, + "loss": 0.0421, + "step": 2024 + }, + { + "epoch": 0.87890625, + "grad_norm": 1.2558727264404297, + "learning_rate": 6.212716542550112e-06, + "loss": 0.0365, + "step": 2025 + }, + { + "epoch": 0.8793402777777778, + "grad_norm": 1.1363980770111084, + "learning_rate": 6.209306323161908e-06, + "loss": 0.0447, + "step": 2026 + }, + { + "epoch": 0.8797743055555556, + "grad_norm": 0.8089574575424194, + "learning_rate": 6.2058955061671625e-06, + "loss": 0.0524, + "step": 2027 + }, + { + "epoch": 0.8802083333333334, + "grad_norm": 0.8970324397087097, + "learning_rate": 6.20248409325141e-06, + "loss": 0.0503, + "step": 2028 + }, + { + "epoch": 0.8806423611111112, + "grad_norm": 4.890934944152832, + "learning_rate": 6.199072086100479e-06, + "loss": 0.0463, + "step": 2029 + }, + { + "epoch": 0.8810763888888888, + "grad_norm": 1.5201784372329712, + "learning_rate": 6.195659486400491e-06, + "loss": 0.0384, + "step": 2030 + }, + { + "epoch": 0.8815104166666666, + "grad_norm": 0.8894219398498535, + "learning_rate": 6.192246295837863e-06, + "loss": 0.0539, + "step": 2031 + }, + { + "epoch": 0.8819444444444444, + "grad_norm": 1.5229333639144897, + "learning_rate": 6.188832516099298e-06, + "loss": 0.0447, + "step": 2032 + }, + { + "epoch": 0.8823784722222222, + "grad_norm": 0.9418681263923645, + "learning_rate": 6.185418148871795e-06, + "loss": 0.0322, + "step": 2033 + }, + { + "epoch": 0.8828125, + "grad_norm": 1.105890154838562, + "learning_rate": 6.182003195842643e-06, + "loss": 0.0517, + "step": 2034 + }, + { + "epoch": 0.8832465277777778, + "grad_norm": 0.7091906666755676, + "learning_rate": 6.178587658699418e-06, + "loss": 0.0399, + "step": 2035 + }, + { + "epoch": 0.8836805555555556, + "grad_norm": 1.3659592866897583, + "learning_rate": 6.175171539129988e-06, + "loss": 0.0459, + "step": 2036 + }, + { + "epoch": 0.8841145833333334, + "grad_norm": 0.693271815776825, + "learning_rate": 6.171754838822508e-06, + "loss": 0.0476, + "step": 2037 + }, + { + "epoch": 0.8845486111111112, + "grad_norm": 0.6045472025871277, + "learning_rate": 6.168337559465416e-06, + "loss": 0.0399, + "step": 2038 + }, + { + "epoch": 0.8849826388888888, + "grad_norm": 0.8418381810188293, + "learning_rate": 6.164919702747442e-06, + "loss": 0.049, + "step": 2039 + }, + { + "epoch": 0.8854166666666666, + "grad_norm": 0.8624420166015625, + "learning_rate": 6.161501270357597e-06, + "loss": 0.0531, + "step": 2040 + }, + { + "epoch": 0.8858506944444444, + "grad_norm": 0.5789508819580078, + "learning_rate": 6.158082263985178e-06, + "loss": 0.0421, + "step": 2041 + }, + { + "epoch": 0.8862847222222222, + "grad_norm": 0.6958301067352295, + "learning_rate": 6.154662685319764e-06, + "loss": 0.0552, + "step": 2042 + }, + { + "epoch": 0.88671875, + "grad_norm": 1.0128082036972046, + "learning_rate": 6.151242536051223e-06, + "loss": 0.0613, + "step": 2043 + }, + { + "epoch": 0.8871527777777778, + "grad_norm": 1.0239746570587158, + "learning_rate": 6.147821817869697e-06, + "loss": 0.0594, + "step": 2044 + }, + { + "epoch": 0.8875868055555556, + "grad_norm": 0.4922561049461365, + "learning_rate": 6.144400532465612e-06, + "loss": 0.0313, + "step": 2045 + }, + { + "epoch": 0.8880208333333334, + "grad_norm": 0.44895923137664795, + "learning_rate": 6.140978681529677e-06, + "loss": 0.0386, + "step": 2046 + }, + { + "epoch": 0.8884548611111112, + "grad_norm": 0.47623518109321594, + "learning_rate": 6.137556266752878e-06, + "loss": 0.0354, + "step": 2047 + }, + { + "epoch": 0.8888888888888888, + "grad_norm": 0.7948447465896606, + "learning_rate": 6.134133289826481e-06, + "loss": 0.0587, + "step": 2048 + }, + { + "epoch": 0.8893229166666666, + "grad_norm": 0.6529760956764221, + "learning_rate": 6.130709752442024e-06, + "loss": 0.045, + "step": 2049 + }, + { + "epoch": 0.8897569444444444, + "grad_norm": 1.3702458143234253, + "learning_rate": 6.127285656291333e-06, + "loss": 0.0638, + "step": 2050 + }, + { + "epoch": 0.8901909722222222, + "grad_norm": 0.9024243354797363, + "learning_rate": 6.1238610030665014e-06, + "loss": 0.0619, + "step": 2051 + }, + { + "epoch": 0.890625, + "grad_norm": 1.0158090591430664, + "learning_rate": 6.120435794459899e-06, + "loss": 0.0539, + "step": 2052 + }, + { + "epoch": 0.8910590277777778, + "grad_norm": 0.49644342064857483, + "learning_rate": 6.117010032164174e-06, + "loss": 0.0376, + "step": 2053 + }, + { + "epoch": 0.8914930555555556, + "grad_norm": 0.5801069736480713, + "learning_rate": 6.113583717872243e-06, + "loss": 0.0466, + "step": 2054 + }, + { + "epoch": 0.8919270833333334, + "grad_norm": 0.5090112686157227, + "learning_rate": 6.110156853277298e-06, + "loss": 0.0435, + "step": 2055 + }, + { + "epoch": 0.8923611111111112, + "grad_norm": 0.7452555298805237, + "learning_rate": 6.106729440072804e-06, + "loss": 0.0458, + "step": 2056 + }, + { + "epoch": 0.8927951388888888, + "grad_norm": 0.6473926305770874, + "learning_rate": 6.103301479952497e-06, + "loss": 0.0483, + "step": 2057 + }, + { + "epoch": 0.8932291666666666, + "grad_norm": 0.6398634314537048, + "learning_rate": 6.099872974610382e-06, + "loss": 0.0356, + "step": 2058 + }, + { + "epoch": 0.8936631944444444, + "grad_norm": 1.5123109817504883, + "learning_rate": 6.0964439257407335e-06, + "loss": 0.0769, + "step": 2059 + }, + { + "epoch": 0.8940972222222222, + "grad_norm": 1.6154077053070068, + "learning_rate": 6.093014335038094e-06, + "loss": 0.0598, + "step": 2060 + }, + { + "epoch": 0.89453125, + "grad_norm": 0.6833464503288269, + "learning_rate": 6.089584204197277e-06, + "loss": 0.0298, + "step": 2061 + }, + { + "epoch": 0.8949652777777778, + "grad_norm": 1.5443462133407593, + "learning_rate": 6.086153534913359e-06, + "loss": 0.0563, + "step": 2062 + }, + { + "epoch": 0.8953993055555556, + "grad_norm": 0.786438524723053, + "learning_rate": 6.082722328881682e-06, + "loss": 0.0319, + "step": 2063 + }, + { + "epoch": 0.8958333333333334, + "grad_norm": 0.7036506533622742, + "learning_rate": 6.079290587797858e-06, + "loss": 0.0528, + "step": 2064 + }, + { + "epoch": 0.8962673611111112, + "grad_norm": 0.4531269669532776, + "learning_rate": 6.075858313357762e-06, + "loss": 0.0254, + "step": 2065 + }, + { + "epoch": 0.8967013888888888, + "grad_norm": 0.7851917743682861, + "learning_rate": 6.072425507257528e-06, + "loss": 0.0423, + "step": 2066 + }, + { + "epoch": 0.8971354166666666, + "grad_norm": 0.6986308097839355, + "learning_rate": 6.068992171193558e-06, + "loss": 0.0198, + "step": 2067 + }, + { + "epoch": 0.8975694444444444, + "grad_norm": 3.0546317100524902, + "learning_rate": 6.065558306862514e-06, + "loss": 0.0408, + "step": 2068 + }, + { + "epoch": 0.8980034722222222, + "grad_norm": 0.8991721272468567, + "learning_rate": 6.06212391596132e-06, + "loss": 0.0408, + "step": 2069 + }, + { + "epoch": 0.8984375, + "grad_norm": 1.241905689239502, + "learning_rate": 6.058689000187157e-06, + "loss": 0.0447, + "step": 2070 + }, + { + "epoch": 0.8988715277777778, + "grad_norm": 1.7747160196304321, + "learning_rate": 6.055253561237469e-06, + "loss": 0.0491, + "step": 2071 + }, + { + "epoch": 0.8993055555555556, + "grad_norm": 0.8431347608566284, + "learning_rate": 6.051817600809957e-06, + "loss": 0.0391, + "step": 2072 + }, + { + "epoch": 0.8997395833333334, + "grad_norm": 1.0173333883285522, + "learning_rate": 6.0483811206025815e-06, + "loss": 0.0517, + "step": 2073 + }, + { + "epoch": 0.9001736111111112, + "grad_norm": 1.2025859355926514, + "learning_rate": 6.044944122313554e-06, + "loss": 0.0228, + "step": 2074 + }, + { + "epoch": 0.9006076388888888, + "grad_norm": 0.8313187956809998, + "learning_rate": 6.041506607641351e-06, + "loss": 0.0427, + "step": 2075 + }, + { + "epoch": 0.9010416666666666, + "grad_norm": 0.8221765160560608, + "learning_rate": 6.038068578284696e-06, + "loss": 0.0401, + "step": 2076 + }, + { + "epoch": 0.9014756944444444, + "grad_norm": 1.3112930059432983, + "learning_rate": 6.034630035942573e-06, + "loss": 0.0426, + "step": 2077 + }, + { + "epoch": 0.9019097222222222, + "grad_norm": 1.079392671585083, + "learning_rate": 6.0311909823142124e-06, + "loss": 0.0376, + "step": 2078 + }, + { + "epoch": 0.90234375, + "grad_norm": 1.1555027961730957, + "learning_rate": 6.027751419099107e-06, + "loss": 0.0456, + "step": 2079 + }, + { + "epoch": 0.9027777777777778, + "grad_norm": 4.339806079864502, + "learning_rate": 6.024311347996994e-06, + "loss": 0.0472, + "step": 2080 + }, + { + "epoch": 0.9032118055555556, + "grad_norm": 1.6129090785980225, + "learning_rate": 6.020870770707864e-06, + "loss": 0.0475, + "step": 2081 + }, + { + "epoch": 0.9036458333333334, + "grad_norm": 1.724150538444519, + "learning_rate": 6.017429688931956e-06, + "loss": 0.0534, + "step": 2082 + }, + { + "epoch": 0.9040798611111112, + "grad_norm": 0.8326488733291626, + "learning_rate": 6.013988104369762e-06, + "loss": 0.0406, + "step": 2083 + }, + { + "epoch": 0.9045138888888888, + "grad_norm": 0.8307406306266785, + "learning_rate": 6.010546018722017e-06, + "loss": 0.0391, + "step": 2084 + }, + { + "epoch": 0.9049479166666666, + "grad_norm": 0.7936608195304871, + "learning_rate": 6.007103433689709e-06, + "loss": 0.0573, + "step": 2085 + }, + { + "epoch": 0.9053819444444444, + "grad_norm": 0.5702849626541138, + "learning_rate": 6.00366035097407e-06, + "loss": 0.0295, + "step": 2086 + }, + { + "epoch": 0.9058159722222222, + "grad_norm": 1.1666098833084106, + "learning_rate": 6.000216772276579e-06, + "loss": 0.0506, + "step": 2087 + }, + { + "epoch": 0.90625, + "grad_norm": 0.600796639919281, + "learning_rate": 5.99677269929896e-06, + "loss": 0.0366, + "step": 2088 + }, + { + "epoch": 0.9066840277777778, + "grad_norm": 0.7910858988761902, + "learning_rate": 5.993328133743182e-06, + "loss": 0.0378, + "step": 2089 + }, + { + "epoch": 0.9071180555555556, + "grad_norm": 1.3797601461410522, + "learning_rate": 5.989883077311453e-06, + "loss": 0.0715, + "step": 2090 + }, + { + "epoch": 0.9075520833333334, + "grad_norm": 1.3978376388549805, + "learning_rate": 5.986437531706226e-06, + "loss": 0.075, + "step": 2091 + }, + { + "epoch": 0.9079861111111112, + "grad_norm": 0.7575281262397766, + "learning_rate": 5.982991498630201e-06, + "loss": 0.0324, + "step": 2092 + }, + { + "epoch": 0.9084201388888888, + "grad_norm": 1.9478482007980347, + "learning_rate": 5.979544979786315e-06, + "loss": 0.0549, + "step": 2093 + }, + { + "epoch": 0.9088541666666666, + "grad_norm": 1.209311604499817, + "learning_rate": 5.97609797687774e-06, + "loss": 0.0308, + "step": 2094 + }, + { + "epoch": 0.9092881944444444, + "grad_norm": 1.0717582702636719, + "learning_rate": 5.972650491607896e-06, + "loss": 0.0683, + "step": 2095 + }, + { + "epoch": 0.9097222222222222, + "grad_norm": 0.743563711643219, + "learning_rate": 5.9692025256804335e-06, + "loss": 0.0484, + "step": 2096 + }, + { + "epoch": 0.91015625, + "grad_norm": 0.5490623712539673, + "learning_rate": 5.965754080799244e-06, + "loss": 0.0312, + "step": 2097 + }, + { + "epoch": 0.9105902777777778, + "grad_norm": 0.6579229235649109, + "learning_rate": 5.96230515866846e-06, + "loss": 0.0568, + "step": 2098 + }, + { + "epoch": 0.9110243055555556, + "grad_norm": 0.5728437900543213, + "learning_rate": 5.9588557609924425e-06, + "loss": 0.0421, + "step": 2099 + }, + { + "epoch": 0.9114583333333334, + "grad_norm": 1.7242991924285889, + "learning_rate": 5.955405889475793e-06, + "loss": 0.0537, + "step": 2100 + }, + { + "epoch": 0.9114583333333334, + "eval_layout_rm_rank_test_accuracy": 0.9735698916000163, + "eval_layout_rm_rank_test_chosen_scores": 0.5979847311973572, + "eval_layout_rm_rank_test_loss": 0.04800070822238922, + "eval_layout_rm_rank_test_rejected_scores": 0.08641856908798218, + "eval_layout_rm_rank_test_runtime": 1197.1207, + "eval_layout_rm_rank_test_samples_per_second": 20.575, + "eval_layout_rm_rank_test_steps_per_second": 0.643, + "step": 2100 + } + ], + "logging_steps": 1.0, + "max_steps": 4608, + "num_input_tokens_seen": 0, + "num_train_epochs": 2, + "save_steps": 300, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 0.0, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +}