{ "best_global_step": 7600, "best_metric": 0.7774137258529663, "best_model_checkpoint": "models/MNLP_M3_rag_model_test/checkpoint-7600", "epoch": 3.97036166601384, "eval_steps": 200, "global_step": 7600, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.005222613918266093, "grad_norm": 11.394719123840332, "learning_rate": 9.000000000000001e-07, "loss": 3.5515, "step": 10 }, { "epoch": 0.010445227836532185, "grad_norm": 6.919948577880859, "learning_rate": 1.9000000000000002e-06, "loss": 3.5008, "step": 20 }, { "epoch": 0.015667841754798278, "grad_norm": 7.034717559814453, "learning_rate": 2.9e-06, "loss": 3.29, "step": 30 }, { "epoch": 0.02089045567306437, "grad_norm": 5.738417625427246, "learning_rate": 3.900000000000001e-06, "loss": 3.1144, "step": 40 }, { "epoch": 0.02611306959133046, "grad_norm": 6.416561603546143, "learning_rate": 4.9000000000000005e-06, "loss": 3.0745, "step": 50 }, { "epoch": 0.031335683509596556, "grad_norm": 5.485833168029785, "learning_rate": 5.9e-06, "loss": 2.8625, "step": 60 }, { "epoch": 0.036558297427862645, "grad_norm": 5.88957405090332, "learning_rate": 6.9e-06, "loss": 2.8214, "step": 70 }, { "epoch": 0.04178091134612874, "grad_norm": 6.4762654304504395, "learning_rate": 7.9e-06, "loss": 2.8076, "step": 80 }, { "epoch": 0.04700352526439483, "grad_norm": 6.038456916809082, "learning_rate": 8.900000000000001e-06, "loss": 2.7299, "step": 90 }, { "epoch": 0.05222613918266092, "grad_norm": 7.74363899230957, "learning_rate": 9.9e-06, "loss": 2.5875, "step": 100 }, { "epoch": 0.057448753100927015, "grad_norm": 11.824464797973633, "learning_rate": 9.990496304118268e-06, "loss": 2.5328, "step": 110 }, { "epoch": 0.06267136701919311, "grad_norm": 6.736820697784424, "learning_rate": 9.979936642027456e-06, "loss": 2.414, "step": 120 }, { "epoch": 0.0678939809374592, "grad_norm": 6.303720474243164, "learning_rate": 9.969376979936643e-06, "loss": 2.2947, "step": 130 }, { "epoch": 0.07311659485572529, "grad_norm": 6.727591037750244, "learning_rate": 9.95881731784583e-06, "loss": 2.2003, "step": 140 }, { "epoch": 0.07833920877399138, "grad_norm": 8.046416282653809, "learning_rate": 9.948257655755017e-06, "loss": 2.0726, "step": 150 }, { "epoch": 0.08356182269225748, "grad_norm": 8.67299747467041, "learning_rate": 9.937697993664203e-06, "loss": 2.2524, "step": 160 }, { "epoch": 0.08878443661052357, "grad_norm": 7.629809856414795, "learning_rate": 9.927138331573391e-06, "loss": 2.0773, "step": 170 }, { "epoch": 0.09400705052878966, "grad_norm": 10.00472640991211, "learning_rate": 9.916578669482577e-06, "loss": 1.9697, "step": 180 }, { "epoch": 0.09922966444705575, "grad_norm": 7.71968412399292, "learning_rate": 9.907074973600845e-06, "loss": 2.1101, "step": 190 }, { "epoch": 0.10445227836532184, "grad_norm": 6.909250736236572, "learning_rate": 9.896515311510033e-06, "loss": 1.866, "step": 200 }, { "epoch": 0.10445227836532184, "eval_loss": 2.0226237773895264, "eval_runtime": 46.6876, "eval_samples_per_second": 36.455, "eval_steps_per_second": 4.562, "step": 200 }, { "epoch": 0.10967489228358794, "grad_norm": 6.597925186157227, "learning_rate": 9.88595564941922e-06, "loss": 1.9556, "step": 210 }, { "epoch": 0.11489750620185403, "grad_norm": 9.504620552062988, "learning_rate": 9.875395987328407e-06, "loss": 1.9974, "step": 220 }, { "epoch": 0.12012012012012012, "grad_norm": 6.907344818115234, "learning_rate": 9.864836325237593e-06, "loss": 1.8866, "step": 230 }, { "epoch": 0.12534273403838622, "grad_norm": 9.235527038574219, "learning_rate": 9.85427666314678e-06, "loss": 1.8387, "step": 240 }, { "epoch": 0.1305653479566523, "grad_norm": 7.033239841461182, "learning_rate": 9.843717001055967e-06, "loss": 1.7348, "step": 250 }, { "epoch": 0.1357879618749184, "grad_norm": 10.7998628616333, "learning_rate": 9.833157338965154e-06, "loss": 1.7569, "step": 260 }, { "epoch": 0.1410105757931845, "grad_norm": 7.843267917633057, "learning_rate": 9.82259767687434e-06, "loss": 1.7569, "step": 270 }, { "epoch": 0.14623318971145058, "grad_norm": 6.2468953132629395, "learning_rate": 9.812038014783528e-06, "loss": 1.6301, "step": 280 }, { "epoch": 0.15145580362971667, "grad_norm": 7.654909133911133, "learning_rate": 9.801478352692714e-06, "loss": 1.7127, "step": 290 }, { "epoch": 0.15667841754798276, "grad_norm": 7.152418613433838, "learning_rate": 9.790918690601902e-06, "loss": 1.8976, "step": 300 }, { "epoch": 0.16190103146624885, "grad_norm": 7.338048458099365, "learning_rate": 9.780359028511088e-06, "loss": 1.7498, "step": 310 }, { "epoch": 0.16712364538451496, "grad_norm": 7.256304740905762, "learning_rate": 9.769799366420275e-06, "loss": 1.7501, "step": 320 }, { "epoch": 0.17234625930278105, "grad_norm": 10.67475700378418, "learning_rate": 9.759239704329462e-06, "loss": 1.6759, "step": 330 }, { "epoch": 0.17756887322104714, "grad_norm": 7.884083271026611, "learning_rate": 9.74868004223865e-06, "loss": 1.707, "step": 340 }, { "epoch": 0.18279148713931323, "grad_norm": 8.517298698425293, "learning_rate": 9.738120380147837e-06, "loss": 1.5422, "step": 350 }, { "epoch": 0.18801410105757932, "grad_norm": 6.652080059051514, "learning_rate": 9.727560718057023e-06, "loss": 1.6762, "step": 360 }, { "epoch": 0.1932367149758454, "grad_norm": 6.86594820022583, "learning_rate": 9.71700105596621e-06, "loss": 1.5937, "step": 370 }, { "epoch": 0.1984593288941115, "grad_norm": 7.43917989730835, "learning_rate": 9.707497360084478e-06, "loss": 1.4299, "step": 380 }, { "epoch": 0.2036819428123776, "grad_norm": 10.837226867675781, "learning_rate": 9.696937697993665e-06, "loss": 1.5797, "step": 390 }, { "epoch": 0.20890455673064368, "grad_norm": 10.075883865356445, "learning_rate": 9.686378035902851e-06, "loss": 1.5084, "step": 400 }, { "epoch": 0.20890455673064368, "eval_loss": 1.625764012336731, "eval_runtime": 46.2554, "eval_samples_per_second": 36.796, "eval_steps_per_second": 4.605, "step": 400 }, { "epoch": 0.2141271706489098, "grad_norm": 6.131842613220215, "learning_rate": 9.675818373812039e-06, "loss": 1.5666, "step": 410 }, { "epoch": 0.21934978456717588, "grad_norm": 8.409153938293457, "learning_rate": 9.665258711721227e-06, "loss": 1.6822, "step": 420 }, { "epoch": 0.22457239848544197, "grad_norm": 8.761375427246094, "learning_rate": 9.654699049630413e-06, "loss": 1.3924, "step": 430 }, { "epoch": 0.22979501240370806, "grad_norm": 6.627100944519043, "learning_rate": 9.6441393875396e-06, "loss": 1.6737, "step": 440 }, { "epoch": 0.23501762632197415, "grad_norm": 9.165101051330566, "learning_rate": 9.633579725448786e-06, "loss": 1.4854, "step": 450 }, { "epoch": 0.24024024024024024, "grad_norm": 8.033590316772461, "learning_rate": 9.623020063357974e-06, "loss": 1.4066, "step": 460 }, { "epoch": 0.24546285415850633, "grad_norm": 7.320120811462402, "learning_rate": 9.612460401267162e-06, "loss": 1.4489, "step": 470 }, { "epoch": 0.25068546807677244, "grad_norm": 6.342758655548096, "learning_rate": 9.601900739176348e-06, "loss": 1.4078, "step": 480 }, { "epoch": 0.2559080819950385, "grad_norm": 7.489528179168701, "learning_rate": 9.591341077085534e-06, "loss": 1.3413, "step": 490 }, { "epoch": 0.2611306959133046, "grad_norm": 6.255088806152344, "learning_rate": 9.58078141499472e-06, "loss": 1.4824, "step": 500 }, { "epoch": 0.2663533098315707, "grad_norm": 6.387566089630127, "learning_rate": 9.570221752903908e-06, "loss": 1.2526, "step": 510 }, { "epoch": 0.2715759237498368, "grad_norm": 7.210233688354492, "learning_rate": 9.559662090813095e-06, "loss": 1.4269, "step": 520 }, { "epoch": 0.27679853766810286, "grad_norm": 6.717288017272949, "learning_rate": 9.549102428722282e-06, "loss": 1.4725, "step": 530 }, { "epoch": 0.282021151586369, "grad_norm": 6.161440372467041, "learning_rate": 9.538542766631469e-06, "loss": 1.4743, "step": 540 }, { "epoch": 0.2872437655046351, "grad_norm": 6.798609733581543, "learning_rate": 9.527983104540655e-06, "loss": 1.4601, "step": 550 }, { "epoch": 0.29246637942290116, "grad_norm": 7.112600326538086, "learning_rate": 9.517423442449843e-06, "loss": 1.3427, "step": 560 }, { "epoch": 0.2976889933411673, "grad_norm": 5.958874225616455, "learning_rate": 9.50686378035903e-06, "loss": 1.5848, "step": 570 }, { "epoch": 0.30291160725943334, "grad_norm": 8.950939178466797, "learning_rate": 9.496304118268215e-06, "loss": 1.3731, "step": 580 }, { "epoch": 0.30813422117769945, "grad_norm": 7.173642635345459, "learning_rate": 9.485744456177403e-06, "loss": 1.4672, "step": 590 }, { "epoch": 0.3133568350959655, "grad_norm": 5.97654390335083, "learning_rate": 9.47518479408659e-06, "loss": 1.3207, "step": 600 }, { "epoch": 0.3133568350959655, "eval_loss": 1.425838589668274, "eval_runtime": 46.3712, "eval_samples_per_second": 36.704, "eval_steps_per_second": 4.593, "step": 600 }, { "epoch": 0.31857944901423163, "grad_norm": 6.612730503082275, "learning_rate": 9.464625131995778e-06, "loss": 1.3925, "step": 610 }, { "epoch": 0.3238020629324977, "grad_norm": 7.3144049644470215, "learning_rate": 9.454065469904964e-06, "loss": 1.494, "step": 620 }, { "epoch": 0.3290246768507638, "grad_norm": 5.725366592407227, "learning_rate": 9.44350580781415e-06, "loss": 1.4134, "step": 630 }, { "epoch": 0.3342472907690299, "grad_norm": 8.623462677001953, "learning_rate": 9.432946145723338e-06, "loss": 1.144, "step": 640 }, { "epoch": 0.339469904687296, "grad_norm": 6.790163040161133, "learning_rate": 9.422386483632524e-06, "loss": 1.4894, "step": 650 }, { "epoch": 0.3446925186055621, "grad_norm": 7.366871356964111, "learning_rate": 9.411826821541712e-06, "loss": 1.3748, "step": 660 }, { "epoch": 0.34991513252382817, "grad_norm": 5.874607086181641, "learning_rate": 9.401267159450899e-06, "loss": 1.0795, "step": 670 }, { "epoch": 0.3551377464420943, "grad_norm": 6.789367198944092, "learning_rate": 9.390707497360085e-06, "loss": 1.2088, "step": 680 }, { "epoch": 0.36036036036036034, "grad_norm": 6.885139465332031, "learning_rate": 9.380147835269273e-06, "loss": 1.1882, "step": 690 }, { "epoch": 0.36558297427862646, "grad_norm": 7.129133224487305, "learning_rate": 9.369588173178459e-06, "loss": 1.2315, "step": 700 }, { "epoch": 0.3708055881968925, "grad_norm": 6.09841775894165, "learning_rate": 9.359028511087645e-06, "loss": 1.1916, "step": 710 }, { "epoch": 0.37602820211515864, "grad_norm": 6.911228179931641, "learning_rate": 9.348468848996833e-06, "loss": 1.44, "step": 720 }, { "epoch": 0.38125081603342476, "grad_norm": 8.852502822875977, "learning_rate": 9.33790918690602e-06, "loss": 1.2999, "step": 730 }, { "epoch": 0.3864734299516908, "grad_norm": 7.887015342712402, "learning_rate": 9.327349524815207e-06, "loss": 1.3509, "step": 740 }, { "epoch": 0.39169604386995693, "grad_norm": 5.902195930480957, "learning_rate": 9.316789862724394e-06, "loss": 1.2506, "step": 750 }, { "epoch": 0.396918657788223, "grad_norm": 5.92559814453125, "learning_rate": 9.30623020063358e-06, "loss": 1.1291, "step": 760 }, { "epoch": 0.4021412717064891, "grad_norm": 7.447033405303955, "learning_rate": 9.295670538542766e-06, "loss": 1.1785, "step": 770 }, { "epoch": 0.4073638856247552, "grad_norm": 5.407535552978516, "learning_rate": 9.285110876451954e-06, "loss": 1.1236, "step": 780 }, { "epoch": 0.4125864995430213, "grad_norm": 5.1882219314575195, "learning_rate": 9.274551214361142e-06, "loss": 1.2353, "step": 790 }, { "epoch": 0.41780911346128735, "grad_norm": 7.093064785003662, "learning_rate": 9.263991552270328e-06, "loss": 1.459, "step": 800 }, { "epoch": 0.41780911346128735, "eval_loss": 1.3080272674560547, "eval_runtime": 46.2478, "eval_samples_per_second": 36.802, "eval_steps_per_second": 4.606, "step": 800 }, { "epoch": 0.42303172737955347, "grad_norm": 6.416601657867432, "learning_rate": 9.253431890179515e-06, "loss": 1.3287, "step": 810 }, { "epoch": 0.4282543412978196, "grad_norm": 6.671374797821045, "learning_rate": 9.242872228088701e-06, "loss": 1.2637, "step": 820 }, { "epoch": 0.43347695521608565, "grad_norm": 6.5349931716918945, "learning_rate": 9.232312565997889e-06, "loss": 1.2804, "step": 830 }, { "epoch": 0.43869956913435176, "grad_norm": 5.837822437286377, "learning_rate": 9.221752903907075e-06, "loss": 1.4397, "step": 840 }, { "epoch": 0.4439221830526178, "grad_norm": 6.69824743270874, "learning_rate": 9.211193241816263e-06, "loss": 1.2325, "step": 850 }, { "epoch": 0.44914479697088394, "grad_norm": 5.331833362579346, "learning_rate": 9.20063357972545e-06, "loss": 1.3133, "step": 860 }, { "epoch": 0.45436741088915, "grad_norm": 6.8653950691223145, "learning_rate": 9.190073917634637e-06, "loss": 1.0374, "step": 870 }, { "epoch": 0.4595900248074161, "grad_norm": 6.36031436920166, "learning_rate": 9.179514255543824e-06, "loss": 1.0597, "step": 880 }, { "epoch": 0.4648126387256822, "grad_norm": 3.890155553817749, "learning_rate": 9.16895459345301e-06, "loss": 1.2453, "step": 890 }, { "epoch": 0.4700352526439483, "grad_norm": 5.179451942443848, "learning_rate": 9.158394931362196e-06, "loss": 1.0745, "step": 900 }, { "epoch": 0.4752578665622144, "grad_norm": 7.032919406890869, "learning_rate": 9.147835269271384e-06, "loss": 1.1987, "step": 910 }, { "epoch": 0.4804804804804805, "grad_norm": 6.223219394683838, "learning_rate": 9.137275607180572e-06, "loss": 1.2441, "step": 920 }, { "epoch": 0.4857030943987466, "grad_norm": 8.567842483520508, "learning_rate": 9.126715945089758e-06, "loss": 1.1818, "step": 930 }, { "epoch": 0.49092570831701265, "grad_norm": 5.338006019592285, "learning_rate": 9.116156282998945e-06, "loss": 1.1527, "step": 940 }, { "epoch": 0.49614832223527877, "grad_norm": 6.287044048309326, "learning_rate": 9.10559662090813e-06, "loss": 1.3234, "step": 950 }, { "epoch": 0.5013709361535449, "grad_norm": 6.575079441070557, "learning_rate": 9.095036958817319e-06, "loss": 1.1989, "step": 960 }, { "epoch": 0.506593550071811, "grad_norm": 7.368027687072754, "learning_rate": 9.084477296726507e-06, "loss": 1.0985, "step": 970 }, { "epoch": 0.511816163990077, "grad_norm": 5.3375020027160645, "learning_rate": 9.073917634635693e-06, "loss": 1.1171, "step": 980 }, { "epoch": 0.5170387779083431, "grad_norm": 7.050992965698242, "learning_rate": 9.06335797254488e-06, "loss": 1.2531, "step": 990 }, { "epoch": 0.5222613918266092, "grad_norm": 6.947799205780029, "learning_rate": 9.052798310454066e-06, "loss": 1.0675, "step": 1000 }, { "epoch": 0.5222613918266092, "eval_loss": 1.2219752073287964, "eval_runtime": 46.2981, "eval_samples_per_second": 36.762, "eval_steps_per_second": 4.601, "step": 1000 }, { "epoch": 0.5274840057448753, "grad_norm": 10.056715965270996, "learning_rate": 9.042238648363253e-06, "loss": 1.1444, "step": 1010 }, { "epoch": 0.5327066196631414, "grad_norm": 6.277551651000977, "learning_rate": 9.03167898627244e-06, "loss": 1.118, "step": 1020 }, { "epoch": 0.5379292335814075, "grad_norm": 5.285930633544922, "learning_rate": 9.021119324181626e-06, "loss": 1.0377, "step": 1030 }, { "epoch": 0.5431518474996736, "grad_norm": 5.600802421569824, "learning_rate": 9.010559662090814e-06, "loss": 1.0265, "step": 1040 }, { "epoch": 0.5483744614179397, "grad_norm": 6.1389007568359375, "learning_rate": 9e-06, "loss": 1.1026, "step": 1050 }, { "epoch": 0.5535970753362057, "grad_norm": 7.223113536834717, "learning_rate": 8.989440337909188e-06, "loss": 1.1998, "step": 1060 }, { "epoch": 0.5588196892544719, "grad_norm": 8.13656997680664, "learning_rate": 8.978880675818374e-06, "loss": 1.0726, "step": 1070 }, { "epoch": 0.564042303172738, "grad_norm": 7.210083484649658, "learning_rate": 8.96832101372756e-06, "loss": 1.1418, "step": 1080 }, { "epoch": 0.569264917091004, "grad_norm": 7.832534313201904, "learning_rate": 8.957761351636749e-06, "loss": 1.1464, "step": 1090 }, { "epoch": 0.5744875310092702, "grad_norm": 5.135114669799805, "learning_rate": 8.947201689545935e-06, "loss": 1.0915, "step": 1100 }, { "epoch": 0.5797101449275363, "grad_norm": 8.231823921203613, "learning_rate": 8.936642027455123e-06, "loss": 1.1763, "step": 1110 }, { "epoch": 0.5849327588458023, "grad_norm": 5.530185699462891, "learning_rate": 8.926082365364309e-06, "loss": 1.1993, "step": 1120 }, { "epoch": 0.5901553727640684, "grad_norm": 5.953641414642334, "learning_rate": 8.915522703273495e-06, "loss": 1.0549, "step": 1130 }, { "epoch": 0.5953779866823345, "grad_norm": 5.919338226318359, "learning_rate": 8.904963041182683e-06, "loss": 1.0351, "step": 1140 }, { "epoch": 0.6006006006006006, "grad_norm": 6.962036609649658, "learning_rate": 8.89440337909187e-06, "loss": 1.046, "step": 1150 }, { "epoch": 0.6058232145188667, "grad_norm": 5.828774929046631, "learning_rate": 8.883843717001058e-06, "loss": 1.087, "step": 1160 }, { "epoch": 0.6110458284371327, "grad_norm": 6.581724643707275, "learning_rate": 8.873284054910244e-06, "loss": 1.0683, "step": 1170 }, { "epoch": 0.6162684423553989, "grad_norm": 7.396463394165039, "learning_rate": 8.86272439281943e-06, "loss": 1.1508, "step": 1180 }, { "epoch": 0.621491056273665, "grad_norm": 5.524245262145996, "learning_rate": 8.852164730728618e-06, "loss": 1.155, "step": 1190 }, { "epoch": 0.626713670191931, "grad_norm": 8.499662399291992, "learning_rate": 8.841605068637804e-06, "loss": 1.2246, "step": 1200 }, { "epoch": 0.626713670191931, "eval_loss": 1.154821753501892, "eval_runtime": 46.2456, "eval_samples_per_second": 36.803, "eval_steps_per_second": 4.606, "step": 1200 }, { "epoch": 0.6319362841101972, "grad_norm": 4.054498195648193, "learning_rate": 8.83104540654699e-06, "loss": 1.1299, "step": 1210 }, { "epoch": 0.6371588980284633, "grad_norm": 6.266629695892334, "learning_rate": 8.820485744456179e-06, "loss": 1.0569, "step": 1220 }, { "epoch": 0.6423815119467293, "grad_norm": 7.285578727722168, "learning_rate": 8.809926082365365e-06, "loss": 1.1132, "step": 1230 }, { "epoch": 0.6476041258649954, "grad_norm": 7.280442237854004, "learning_rate": 8.799366420274553e-06, "loss": 1.1091, "step": 1240 }, { "epoch": 0.6528267397832616, "grad_norm": 6.459787368774414, "learning_rate": 8.788806758183739e-06, "loss": 1.1433, "step": 1250 }, { "epoch": 0.6580493537015276, "grad_norm": 6.095096588134766, "learning_rate": 8.778247096092925e-06, "loss": 1.1457, "step": 1260 }, { "epoch": 0.6632719676197937, "grad_norm": 6.624663352966309, "learning_rate": 8.767687434002112e-06, "loss": 1.1755, "step": 1270 }, { "epoch": 0.6684945815380599, "grad_norm": 5.858925819396973, "learning_rate": 8.7571277719113e-06, "loss": 0.9852, "step": 1280 }, { "epoch": 0.6737171954563259, "grad_norm": 6.378047943115234, "learning_rate": 8.746568109820487e-06, "loss": 0.989, "step": 1290 }, { "epoch": 0.678939809374592, "grad_norm": 6.781316757202148, "learning_rate": 8.736008447729674e-06, "loss": 0.8984, "step": 1300 }, { "epoch": 0.684162423292858, "grad_norm": 5.8333940505981445, "learning_rate": 8.72544878563886e-06, "loss": 1.0745, "step": 1310 }, { "epoch": 0.6893850372111242, "grad_norm": 5.035146236419678, "learning_rate": 8.714889123548046e-06, "loss": 0.9417, "step": 1320 }, { "epoch": 0.6946076511293903, "grad_norm": 5.607509613037109, "learning_rate": 8.704329461457234e-06, "loss": 1.0106, "step": 1330 }, { "epoch": 0.6998302650476563, "grad_norm": 8.404295921325684, "learning_rate": 8.69376979936642e-06, "loss": 1.063, "step": 1340 }, { "epoch": 0.7050528789659224, "grad_norm": 6.693871021270752, "learning_rate": 8.683210137275608e-06, "loss": 0.9467, "step": 1350 }, { "epoch": 0.7102754928841886, "grad_norm": 8.556498527526855, "learning_rate": 8.672650475184795e-06, "loss": 0.9936, "step": 1360 }, { "epoch": 0.7154981068024546, "grad_norm": 6.516254425048828, "learning_rate": 8.662090813093983e-06, "loss": 0.958, "step": 1370 }, { "epoch": 0.7207207207207207, "grad_norm": 6.926424026489258, "learning_rate": 8.651531151003169e-06, "loss": 1.0758, "step": 1380 }, { "epoch": 0.7259433346389869, "grad_norm": 5.722439765930176, "learning_rate": 8.640971488912355e-06, "loss": 0.987, "step": 1390 }, { "epoch": 0.7311659485572529, "grad_norm": 5.813411712646484, "learning_rate": 8.630411826821541e-06, "loss": 1.0847, "step": 1400 }, { "epoch": 0.7311659485572529, "eval_loss": 1.0959590673446655, "eval_runtime": 46.2353, "eval_samples_per_second": 36.812, "eval_steps_per_second": 4.607, "step": 1400 }, { "epoch": 0.736388562475519, "grad_norm": 9.68322467803955, "learning_rate": 8.61985216473073e-06, "loss": 1.0947, "step": 1410 }, { "epoch": 0.741611176393785, "grad_norm": 5.5112762451171875, "learning_rate": 8.609292502639917e-06, "loss": 1.0718, "step": 1420 }, { "epoch": 0.7468337903120512, "grad_norm": 6.487595558166504, "learning_rate": 8.598732840549104e-06, "loss": 1.008, "step": 1430 }, { "epoch": 0.7520564042303173, "grad_norm": 6.72760534286499, "learning_rate": 8.58817317845829e-06, "loss": 1.0198, "step": 1440 }, { "epoch": 0.7572790181485833, "grad_norm": 6.073751449584961, "learning_rate": 8.577613516367476e-06, "loss": 1.0065, "step": 1450 }, { "epoch": 0.7625016320668495, "grad_norm": 6.201911449432373, "learning_rate": 8.567053854276664e-06, "loss": 1.2082, "step": 1460 }, { "epoch": 0.7677242459851156, "grad_norm": 7.030183792114258, "learning_rate": 8.55649419218585e-06, "loss": 1.0877, "step": 1470 }, { "epoch": 0.7729468599033816, "grad_norm": 6.393901348114014, "learning_rate": 8.545934530095038e-06, "loss": 0.9444, "step": 1480 }, { "epoch": 0.7781694738216477, "grad_norm": 7.0087571144104, "learning_rate": 8.535374868004225e-06, "loss": 0.9064, "step": 1490 }, { "epoch": 0.7833920877399139, "grad_norm": 7.8241777420043945, "learning_rate": 8.52481520591341e-06, "loss": 1.0552, "step": 1500 }, { "epoch": 0.7886147016581799, "grad_norm": 6.263652801513672, "learning_rate": 8.514255543822599e-06, "loss": 1.1847, "step": 1510 }, { "epoch": 0.793837315576446, "grad_norm": 7.5798797607421875, "learning_rate": 8.503695881731785e-06, "loss": 1.076, "step": 1520 }, { "epoch": 0.799059929494712, "grad_norm": 5.389642238616943, "learning_rate": 8.493136219640971e-06, "loss": 1.0102, "step": 1530 }, { "epoch": 0.8042825434129782, "grad_norm": 5.936399459838867, "learning_rate": 8.48257655755016e-06, "loss": 1.0314, "step": 1540 }, { "epoch": 0.8095051573312443, "grad_norm": 8.437224388122559, "learning_rate": 8.472016895459345e-06, "loss": 1.0985, "step": 1550 }, { "epoch": 0.8147277712495103, "grad_norm": 5.470308303833008, "learning_rate": 8.461457233368533e-06, "loss": 1.0003, "step": 1560 }, { "epoch": 0.8199503851677765, "grad_norm": 10.188332557678223, "learning_rate": 8.45089757127772e-06, "loss": 1.0165, "step": 1570 }, { "epoch": 0.8251729990860426, "grad_norm": 8.477367401123047, "learning_rate": 8.440337909186906e-06, "loss": 1.0532, "step": 1580 }, { "epoch": 0.8303956130043086, "grad_norm": 4.078097820281982, "learning_rate": 8.429778247096094e-06, "loss": 1.0828, "step": 1590 }, { "epoch": 0.8356182269225747, "grad_norm": 5.285001277923584, "learning_rate": 8.41921858500528e-06, "loss": 0.9931, "step": 1600 }, { "epoch": 0.8356182269225747, "eval_loss": 1.0581225156784058, "eval_runtime": 46.2517, "eval_samples_per_second": 36.799, "eval_steps_per_second": 4.605, "step": 1600 }, { "epoch": 0.8408408408408409, "grad_norm": 4.965864181518555, "learning_rate": 8.408658922914468e-06, "loss": 1.1802, "step": 1610 }, { "epoch": 0.8460634547591069, "grad_norm": 6.969324588775635, "learning_rate": 8.398099260823654e-06, "loss": 1.0296, "step": 1620 }, { "epoch": 0.851286068677373, "grad_norm": 8.806953430175781, "learning_rate": 8.38753959873284e-06, "loss": 0.9611, "step": 1630 }, { "epoch": 0.8565086825956392, "grad_norm": 7.901791572570801, "learning_rate": 8.376979936642029e-06, "loss": 0.9345, "step": 1640 }, { "epoch": 0.8617312965139052, "grad_norm": 6.055725574493408, "learning_rate": 8.366420274551215e-06, "loss": 1.0747, "step": 1650 }, { "epoch": 0.8669539104321713, "grad_norm": 7.076270580291748, "learning_rate": 8.355860612460403e-06, "loss": 0.9801, "step": 1660 }, { "epoch": 0.8721765243504374, "grad_norm": 5.258996963500977, "learning_rate": 8.345300950369589e-06, "loss": 1.0273, "step": 1670 }, { "epoch": 0.8773991382687035, "grad_norm": 6.378342151641846, "learning_rate": 8.334741288278775e-06, "loss": 0.8198, "step": 1680 }, { "epoch": 0.8826217521869696, "grad_norm": 6.257429599761963, "learning_rate": 8.324181626187963e-06, "loss": 0.9199, "step": 1690 }, { "epoch": 0.8878443661052356, "grad_norm": 8.555800437927246, "learning_rate": 8.31362196409715e-06, "loss": 0.9679, "step": 1700 }, { "epoch": 0.8930669800235017, "grad_norm": 5.934536933898926, "learning_rate": 8.303062302006336e-06, "loss": 0.9922, "step": 1710 }, { "epoch": 0.8982895939417679, "grad_norm": 5.284457206726074, "learning_rate": 8.292502639915522e-06, "loss": 1.0682, "step": 1720 }, { "epoch": 0.903512207860034, "grad_norm": 6.454044342041016, "learning_rate": 8.28194297782471e-06, "loss": 1.1901, "step": 1730 }, { "epoch": 0.9087348217783, "grad_norm": 6.973818778991699, "learning_rate": 8.271383315733898e-06, "loss": 1.01, "step": 1740 }, { "epoch": 0.9139574356965662, "grad_norm": 7.165948390960693, "learning_rate": 8.260823653643084e-06, "loss": 0.926, "step": 1750 }, { "epoch": 0.9191800496148322, "grad_norm": 5.891210556030273, "learning_rate": 8.25026399155227e-06, "loss": 1.0527, "step": 1760 }, { "epoch": 0.9244026635330983, "grad_norm": 6.440408229827881, "learning_rate": 8.239704329461457e-06, "loss": 1.0271, "step": 1770 }, { "epoch": 0.9296252774513644, "grad_norm": 6.762996673583984, "learning_rate": 8.229144667370645e-06, "loss": 1.1016, "step": 1780 }, { "epoch": 0.9348478913696305, "grad_norm": 7.777276515960693, "learning_rate": 8.218585005279833e-06, "loss": 1.2933, "step": 1790 }, { "epoch": 0.9400705052878966, "grad_norm": 6.9960713386535645, "learning_rate": 8.208025343189019e-06, "loss": 1.1484, "step": 1800 }, { "epoch": 0.9400705052878966, "eval_loss": 1.0149155855178833, "eval_runtime": 46.3161, "eval_samples_per_second": 36.748, "eval_steps_per_second": 4.599, "step": 1800 }, { "epoch": 0.9452931192061627, "grad_norm": 5.822863578796387, "learning_rate": 8.197465681098205e-06, "loss": 0.952, "step": 1810 }, { "epoch": 0.9505157331244288, "grad_norm": 5.783324718475342, "learning_rate": 8.186906019007393e-06, "loss": 0.9129, "step": 1820 }, { "epoch": 0.9557383470426949, "grad_norm": 7.200591564178467, "learning_rate": 8.17634635691658e-06, "loss": 1.0411, "step": 1830 }, { "epoch": 0.960960960960961, "grad_norm": 6.707890033721924, "learning_rate": 8.165786694825766e-06, "loss": 1.0755, "step": 1840 }, { "epoch": 0.966183574879227, "grad_norm": 4.4951372146606445, "learning_rate": 8.155227032734954e-06, "loss": 1.1139, "step": 1850 }, { "epoch": 0.9714061887974932, "grad_norm": 7.841273307800293, "learning_rate": 8.14466737064414e-06, "loss": 0.9171, "step": 1860 }, { "epoch": 0.9766288027157592, "grad_norm": 8.396512985229492, "learning_rate": 8.134107708553328e-06, "loss": 0.8624, "step": 1870 }, { "epoch": 0.9818514166340253, "grad_norm": 7.903951168060303, "learning_rate": 8.123548046462514e-06, "loss": 0.9832, "step": 1880 }, { "epoch": 0.9870740305522914, "grad_norm": 5.722747325897217, "learning_rate": 8.1129883843717e-06, "loss": 0.9931, "step": 1890 }, { "epoch": 0.9922966444705575, "grad_norm": 6.809545993804932, "learning_rate": 8.102428722280887e-06, "loss": 0.987, "step": 1900 }, { "epoch": 0.9975192583888236, "grad_norm": 5.796718120574951, "learning_rate": 8.091869060190075e-06, "loss": 0.9521, "step": 1910 }, { "epoch": 1.0031335683509597, "grad_norm": 4.707700252532959, "learning_rate": 8.081309398099263e-06, "loss": 1.0939, "step": 1920 }, { "epoch": 1.0083561822692257, "grad_norm": 5.458223342895508, "learning_rate": 8.070749736008449e-06, "loss": 0.8501, "step": 1930 }, { "epoch": 1.0135787961874918, "grad_norm": 7.022110939025879, "learning_rate": 8.060190073917635e-06, "loss": 0.8575, "step": 1940 }, { "epoch": 1.0188014101057579, "grad_norm": 6.051275730133057, "learning_rate": 8.049630411826821e-06, "loss": 0.7803, "step": 1950 }, { "epoch": 1.024024024024024, "grad_norm": 13.546333312988281, "learning_rate": 8.03907074973601e-06, "loss": 0.691, "step": 1960 }, { "epoch": 1.0292466379422902, "grad_norm": 6.829512596130371, "learning_rate": 8.028511087645196e-06, "loss": 0.7195, "step": 1970 }, { "epoch": 1.0344692518605563, "grad_norm": 6.821556091308594, "learning_rate": 8.017951425554384e-06, "loss": 0.7773, "step": 1980 }, { "epoch": 1.0396918657788223, "grad_norm": 4.730713844299316, "learning_rate": 8.00739176346357e-06, "loss": 0.6783, "step": 1990 }, { "epoch": 1.0449144796970884, "grad_norm": 6.354138374328613, "learning_rate": 7.996832101372756e-06, "loss": 0.9788, "step": 2000 }, { "epoch": 1.0449144796970884, "eval_loss": 0.9896802306175232, "eval_runtime": 46.2765, "eval_samples_per_second": 36.779, "eval_steps_per_second": 4.603, "step": 2000 }, { "epoch": 1.0501370936153545, "grad_norm": 6.299434185028076, "learning_rate": 7.986272439281944e-06, "loss": 0.7521, "step": 2010 }, { "epoch": 1.0553597075336205, "grad_norm": 8.378788948059082, "learning_rate": 7.97571277719113e-06, "loss": 0.7983, "step": 2020 }, { "epoch": 1.0605823214518866, "grad_norm": 5.674183368682861, "learning_rate": 7.965153115100317e-06, "loss": 0.8165, "step": 2030 }, { "epoch": 1.0658049353701529, "grad_norm": 5.855607032775879, "learning_rate": 7.954593453009504e-06, "loss": 1.0197, "step": 2040 }, { "epoch": 1.071027549288419, "grad_norm": 5.06273078918457, "learning_rate": 7.94403379091869e-06, "loss": 0.799, "step": 2050 }, { "epoch": 1.076250163206685, "grad_norm": 4.809935092926025, "learning_rate": 7.933474128827879e-06, "loss": 0.8601, "step": 2060 }, { "epoch": 1.081472777124951, "grad_norm": 4.642035007476807, "learning_rate": 7.922914466737065e-06, "loss": 0.8228, "step": 2070 }, { "epoch": 1.086695391043217, "grad_norm": 6.76859188079834, "learning_rate": 7.912354804646251e-06, "loss": 0.8457, "step": 2080 }, { "epoch": 1.0919180049614832, "grad_norm": 7.555065155029297, "learning_rate": 7.901795142555439e-06, "loss": 0.8369, "step": 2090 }, { "epoch": 1.0971406188797492, "grad_norm": 5.144375324249268, "learning_rate": 7.891235480464627e-06, "loss": 0.6828, "step": 2100 }, { "epoch": 1.1023632327980155, "grad_norm": 6.584686756134033, "learning_rate": 7.880675818373813e-06, "loss": 0.7198, "step": 2110 }, { "epoch": 1.1075858467162816, "grad_norm": 5.92726469039917, "learning_rate": 7.870116156283e-06, "loss": 0.8571, "step": 2120 }, { "epoch": 1.1128084606345476, "grad_norm": 5.866957187652588, "learning_rate": 7.859556494192186e-06, "loss": 0.8838, "step": 2130 }, { "epoch": 1.1180310745528137, "grad_norm": 6.889613151550293, "learning_rate": 7.848996832101374e-06, "loss": 0.7764, "step": 2140 }, { "epoch": 1.1232536884710798, "grad_norm": 7.770586013793945, "learning_rate": 7.83843717001056e-06, "loss": 0.7697, "step": 2150 }, { "epoch": 1.1284763023893458, "grad_norm": 6.084799766540527, "learning_rate": 7.827877507919746e-06, "loss": 0.8748, "step": 2160 }, { "epoch": 1.1336989163076119, "grad_norm": 8.996906280517578, "learning_rate": 7.817317845828934e-06, "loss": 0.6519, "step": 2170 }, { "epoch": 1.1389215302258782, "grad_norm": 4.936269283294678, "learning_rate": 7.80675818373812e-06, "loss": 0.9594, "step": 2180 }, { "epoch": 1.1441441441441442, "grad_norm": 5.769779205322266, "learning_rate": 7.796198521647309e-06, "loss": 0.7806, "step": 2190 }, { "epoch": 1.1493667580624103, "grad_norm": 7.1322808265686035, "learning_rate": 7.785638859556495e-06, "loss": 0.8086, "step": 2200 }, { "epoch": 1.1493667580624103, "eval_loss": 0.969113826751709, "eval_runtime": 46.2762, "eval_samples_per_second": 36.779, "eval_steps_per_second": 4.603, "step": 2200 }, { "epoch": 1.1545893719806763, "grad_norm": 6.716241359710693, "learning_rate": 7.775079197465681e-06, "loss": 0.9122, "step": 2210 }, { "epoch": 1.1598119858989424, "grad_norm": 5.767160892486572, "learning_rate": 7.764519535374867e-06, "loss": 0.7598, "step": 2220 }, { "epoch": 1.1650345998172085, "grad_norm": 7.989006519317627, "learning_rate": 7.753959873284055e-06, "loss": 0.6704, "step": 2230 }, { "epoch": 1.1702572137354745, "grad_norm": 5.272390365600586, "learning_rate": 7.743400211193243e-06, "loss": 0.8521, "step": 2240 }, { "epoch": 1.1754798276537408, "grad_norm": 6.191717147827148, "learning_rate": 7.73284054910243e-06, "loss": 0.8905, "step": 2250 }, { "epoch": 1.1807024415720069, "grad_norm": 5.682114124298096, "learning_rate": 7.722280887011616e-06, "loss": 0.8386, "step": 2260 }, { "epoch": 1.185925055490273, "grad_norm": 6.549655914306641, "learning_rate": 7.711721224920804e-06, "loss": 0.691, "step": 2270 }, { "epoch": 1.191147669408539, "grad_norm": 6.487022876739502, "learning_rate": 7.70116156282999e-06, "loss": 0.7769, "step": 2280 }, { "epoch": 1.196370283326805, "grad_norm": 7.281522274017334, "learning_rate": 7.690601900739178e-06, "loss": 0.7235, "step": 2290 }, { "epoch": 1.2015928972450711, "grad_norm": 7.294795513153076, "learning_rate": 7.680042238648364e-06, "loss": 0.7813, "step": 2300 }, { "epoch": 1.2068155111633372, "grad_norm": 6.67874813079834, "learning_rate": 7.66948257655755e-06, "loss": 0.7375, "step": 2310 }, { "epoch": 1.2120381250816032, "grad_norm": 3.83427357673645, "learning_rate": 7.658922914466738e-06, "loss": 0.7533, "step": 2320 }, { "epoch": 1.2172607389998695, "grad_norm": 5.654359340667725, "learning_rate": 7.648363252375925e-06, "loss": 0.8276, "step": 2330 }, { "epoch": 1.2224833529181356, "grad_norm": 5.315032482147217, "learning_rate": 7.637803590285111e-06, "loss": 0.7584, "step": 2340 }, { "epoch": 1.2277059668364017, "grad_norm": 6.630548477172852, "learning_rate": 7.627243928194299e-06, "loss": 0.6388, "step": 2350 }, { "epoch": 1.2329285807546677, "grad_norm": 5.981212615966797, "learning_rate": 7.616684266103486e-06, "loss": 0.8596, "step": 2360 }, { "epoch": 1.2381511946729338, "grad_norm": 5.186179161071777, "learning_rate": 7.606124604012672e-06, "loss": 0.7127, "step": 2370 }, { "epoch": 1.2433738085911998, "grad_norm": 6.0210747718811035, "learning_rate": 7.595564941921859e-06, "loss": 0.8277, "step": 2380 }, { "epoch": 1.248596422509466, "grad_norm": 6.917499542236328, "learning_rate": 7.585005279831046e-06, "loss": 0.6321, "step": 2390 }, { "epoch": 1.253819036427732, "grad_norm": 6.423802852630615, "learning_rate": 7.574445617740233e-06, "loss": 0.8517, "step": 2400 }, { "epoch": 1.253819036427732, "eval_loss": 0.9506328701972961, "eval_runtime": 46.3505, "eval_samples_per_second": 36.72, "eval_steps_per_second": 4.595, "step": 2400 }, { "epoch": 1.2590416503459982, "grad_norm": 6.889662742614746, "learning_rate": 7.563885955649419e-06, "loss": 0.7603, "step": 2410 }, { "epoch": 1.2642642642642643, "grad_norm": 4.960092544555664, "learning_rate": 7.553326293558607e-06, "loss": 0.7804, "step": 2420 }, { "epoch": 1.2694868781825304, "grad_norm": 5.164410591125488, "learning_rate": 7.542766631467794e-06, "loss": 0.6938, "step": 2430 }, { "epoch": 1.2747094921007964, "grad_norm": 5.916507720947266, "learning_rate": 7.53220696937698e-06, "loss": 0.6964, "step": 2440 }, { "epoch": 1.2799321060190625, "grad_norm": 6.940438747406006, "learning_rate": 7.521647307286167e-06, "loss": 0.7605, "step": 2450 }, { "epoch": 1.2851547199373288, "grad_norm": 6.659502983093262, "learning_rate": 7.511087645195354e-06, "loss": 0.8735, "step": 2460 }, { "epoch": 1.2903773338555946, "grad_norm": 5.659145355224609, "learning_rate": 7.500527983104541e-06, "loss": 0.726, "step": 2470 }, { "epoch": 1.295599947773861, "grad_norm": 3.484576463699341, "learning_rate": 7.489968321013729e-06, "loss": 0.6992, "step": 2480 }, { "epoch": 1.300822561692127, "grad_norm": 6.005791664123535, "learning_rate": 7.479408658922915e-06, "loss": 0.8257, "step": 2490 }, { "epoch": 1.306045175610393, "grad_norm": 4.005056381225586, "learning_rate": 7.468848996832102e-06, "loss": 0.8418, "step": 2500 }, { "epoch": 1.311267789528659, "grad_norm": 6.585374355316162, "learning_rate": 7.458289334741288e-06, "loss": 0.7168, "step": 2510 }, { "epoch": 1.3164904034469251, "grad_norm": 6.849618434906006, "learning_rate": 7.4477296726504755e-06, "loss": 0.8348, "step": 2520 }, { "epoch": 1.3217130173651912, "grad_norm": 4.997506618499756, "learning_rate": 7.437170010559663e-06, "loss": 0.7155, "step": 2530 }, { "epoch": 1.3269356312834573, "grad_norm": 6.247817516326904, "learning_rate": 7.42661034846885e-06, "loss": 0.7999, "step": 2540 }, { "epoch": 1.3321582452017235, "grad_norm": 6.865342617034912, "learning_rate": 7.416050686378037e-06, "loss": 0.8178, "step": 2550 }, { "epoch": 1.3373808591199896, "grad_norm": 7.5182695388793945, "learning_rate": 7.405491024287224e-06, "loss": 0.8535, "step": 2560 }, { "epoch": 1.3426034730382557, "grad_norm": 5.786922454833984, "learning_rate": 7.39493136219641e-06, "loss": 0.8538, "step": 2570 }, { "epoch": 1.3478260869565217, "grad_norm": 5.576653480529785, "learning_rate": 7.384371700105597e-06, "loss": 0.6142, "step": 2580 }, { "epoch": 1.3530487008747878, "grad_norm": 6.3509135246276855, "learning_rate": 7.3738120380147835e-06, "loss": 0.872, "step": 2590 }, { "epoch": 1.3582713147930539, "grad_norm": 6.3292131423950195, "learning_rate": 7.3632523759239715e-06, "loss": 0.7452, "step": 2600 }, { "epoch": 1.3582713147930539, "eval_loss": 0.928360104560852, "eval_runtime": 46.2783, "eval_samples_per_second": 36.778, "eval_steps_per_second": 4.603, "step": 2600 }, { "epoch": 1.36349392871132, "grad_norm": 5.916106700897217, "learning_rate": 7.352692713833159e-06, "loss": 0.7897, "step": 2610 }, { "epoch": 1.3687165426295862, "grad_norm": 6.225005149841309, "learning_rate": 7.342133051742345e-06, "loss": 0.7989, "step": 2620 }, { "epoch": 1.3739391565478523, "grad_norm": 7.300755500793457, "learning_rate": 7.331573389651532e-06, "loss": 0.8093, "step": 2630 }, { "epoch": 1.3791617704661183, "grad_norm": 6.355301380157471, "learning_rate": 7.321013727560718e-06, "loss": 0.6979, "step": 2640 }, { "epoch": 1.3843843843843844, "grad_norm": 6.439295291900635, "learning_rate": 7.310454065469905e-06, "loss": 0.6481, "step": 2650 }, { "epoch": 1.3896069983026504, "grad_norm": 5.19166374206543, "learning_rate": 7.299894403379092e-06, "loss": 0.6853, "step": 2660 }, { "epoch": 1.3948296122209165, "grad_norm": 7.574211120605469, "learning_rate": 7.2893347412882796e-06, "loss": 0.6786, "step": 2670 }, { "epoch": 1.4000522261391826, "grad_norm": 6.01971435546875, "learning_rate": 7.278775079197467e-06, "loss": 0.7968, "step": 2680 }, { "epoch": 1.4052748400574488, "grad_norm": 4.888395309448242, "learning_rate": 7.268215417106653e-06, "loss": 0.8298, "step": 2690 }, { "epoch": 1.410497453975715, "grad_norm": 4.738596439361572, "learning_rate": 7.25765575501584e-06, "loss": 0.6968, "step": 2700 }, { "epoch": 1.415720067893981, "grad_norm": 6.128376483917236, "learning_rate": 7.247096092925026e-06, "loss": 0.7189, "step": 2710 }, { "epoch": 1.420942681812247, "grad_norm": 6.777405738830566, "learning_rate": 7.236536430834213e-06, "loss": 0.7324, "step": 2720 }, { "epoch": 1.426165295730513, "grad_norm": 5.1269402503967285, "learning_rate": 7.225976768743401e-06, "loss": 0.6326, "step": 2730 }, { "epoch": 1.4313879096487792, "grad_norm": 7.080173492431641, "learning_rate": 7.215417106652588e-06, "loss": 0.7193, "step": 2740 }, { "epoch": 1.4366105235670452, "grad_norm": 6.149571895599365, "learning_rate": 7.204857444561775e-06, "loss": 0.815, "step": 2750 }, { "epoch": 1.4418331374853115, "grad_norm": 4.2188849449157715, "learning_rate": 7.194297782470962e-06, "loss": 0.8526, "step": 2760 }, { "epoch": 1.4470557514035776, "grad_norm": 6.189548492431641, "learning_rate": 7.183738120380148e-06, "loss": 0.7041, "step": 2770 }, { "epoch": 1.4522783653218436, "grad_norm": 8.304208755493164, "learning_rate": 7.173178458289335e-06, "loss": 0.8734, "step": 2780 }, { "epoch": 1.4575009792401097, "grad_norm": 8.095356941223145, "learning_rate": 7.162618796198522e-06, "loss": 0.733, "step": 2790 }, { "epoch": 1.4627235931583757, "grad_norm": 5.834177017211914, "learning_rate": 7.1520591341077094e-06, "loss": 0.7102, "step": 2800 }, { "epoch": 1.4627235931583757, "eval_loss": 0.9070786237716675, "eval_runtime": 46.2298, "eval_samples_per_second": 36.816, "eval_steps_per_second": 4.607, "step": 2800 }, { "epoch": 1.4679462070766418, "grad_norm": 7.127483367919922, "learning_rate": 7.1414994720168965e-06, "loss": 0.8149, "step": 2810 }, { "epoch": 1.4731688209949079, "grad_norm": 5.216626167297363, "learning_rate": 7.131995776135164e-06, "loss": 0.7184, "step": 2820 }, { "epoch": 1.4783914349131742, "grad_norm": 5.421391487121582, "learning_rate": 7.121436114044351e-06, "loss": 0.757, "step": 2830 }, { "epoch": 1.4836140488314402, "grad_norm": 5.557046413421631, "learning_rate": 7.110876451953538e-06, "loss": 0.7887, "step": 2840 }, { "epoch": 1.4888366627497063, "grad_norm": 4.0539870262146, "learning_rate": 7.1003167898627245e-06, "loss": 0.7473, "step": 2850 }, { "epoch": 1.4940592766679723, "grad_norm": 5.317719459533691, "learning_rate": 7.0897571277719124e-06, "loss": 0.7603, "step": 2860 }, { "epoch": 1.4992818905862384, "grad_norm": 7.20483922958374, "learning_rate": 7.079197465681099e-06, "loss": 0.618, "step": 2870 }, { "epoch": 1.5045045045045045, "grad_norm": 5.992430686950684, "learning_rate": 7.068637803590286e-06, "loss": 0.7101, "step": 2880 }, { "epoch": 1.5097271184227705, "grad_norm": 5.599936008453369, "learning_rate": 7.058078141499473e-06, "loss": 0.7421, "step": 2890 }, { "epoch": 1.5149497323410368, "grad_norm": 3.7422847747802734, "learning_rate": 7.047518479408659e-06, "loss": 0.57, "step": 2900 }, { "epoch": 1.5201723462593026, "grad_norm": 6.516021251678467, "learning_rate": 7.036958817317846e-06, "loss": 0.7308, "step": 2910 }, { "epoch": 1.525394960177569, "grad_norm": 6.263866901397705, "learning_rate": 7.026399155227034e-06, "loss": 0.8289, "step": 2920 }, { "epoch": 1.530617574095835, "grad_norm": 4.577157974243164, "learning_rate": 7.0158394931362205e-06, "loss": 0.8547, "step": 2930 }, { "epoch": 1.535840188014101, "grad_norm": 4.392026424407959, "learning_rate": 7.006335797254489e-06, "loss": 0.7407, "step": 2940 }, { "epoch": 1.541062801932367, "grad_norm": 4.726680755615234, "learning_rate": 6.995776135163675e-06, "loss": 0.7082, "step": 2950 }, { "epoch": 1.5462854158506332, "grad_norm": 6.287652492523193, "learning_rate": 6.985216473072862e-06, "loss": 0.7867, "step": 2960 }, { "epoch": 1.5515080297688995, "grad_norm": 4.373517990112305, "learning_rate": 6.974656810982049e-06, "loss": 0.6765, "step": 2970 }, { "epoch": 1.5567306436871653, "grad_norm": 6.31062126159668, "learning_rate": 6.9640971488912356e-06, "loss": 0.6577, "step": 2980 }, { "epoch": 1.5619532576054316, "grad_norm": 4.2869415283203125, "learning_rate": 6.9535374868004235e-06, "loss": 0.6042, "step": 2990 }, { "epoch": 1.5671758715236976, "grad_norm": 4.132930278778076, "learning_rate": 6.942977824709611e-06, "loss": 0.7663, "step": 3000 }, { "epoch": 1.5671758715236976, "eval_loss": 0.8875888586044312, "eval_runtime": 46.2332, "eval_samples_per_second": 36.813, "eval_steps_per_second": 4.607, "step": 3000 }, { "epoch": 1.5723984854419637, "grad_norm": 4.136049270629883, "learning_rate": 6.932418162618797e-06, "loss": 0.6862, "step": 3010 }, { "epoch": 1.5776210993602298, "grad_norm": 4.138570308685303, "learning_rate": 6.921858500527984e-06, "loss": 0.6622, "step": 3020 }, { "epoch": 1.5828437132784958, "grad_norm": 6.920501708984375, "learning_rate": 6.91129883843717e-06, "loss": 0.6052, "step": 3030 }, { "epoch": 1.588066327196762, "grad_norm": 5.639624118804932, "learning_rate": 6.900739176346357e-06, "loss": 0.6701, "step": 3040 }, { "epoch": 1.593288941115028, "grad_norm": 5.700570106506348, "learning_rate": 6.890179514255544e-06, "loss": 0.8079, "step": 3050 }, { "epoch": 1.5985115550332942, "grad_norm": 4.964538097381592, "learning_rate": 6.8796198521647316e-06, "loss": 0.6963, "step": 3060 }, { "epoch": 1.6037341689515603, "grad_norm": 4.319785118103027, "learning_rate": 6.869060190073919e-06, "loss": 0.6664, "step": 3070 }, { "epoch": 1.6089567828698264, "grad_norm": 6.524580478668213, "learning_rate": 6.858500527983105e-06, "loss": 0.7526, "step": 3080 }, { "epoch": 1.6141793967880924, "grad_norm": 5.287715911865234, "learning_rate": 6.847940865892292e-06, "loss": 0.6971, "step": 3090 }, { "epoch": 1.6194020107063585, "grad_norm": 6.37127161026001, "learning_rate": 6.837381203801478e-06, "loss": 0.6264, "step": 3100 }, { "epoch": 1.6246246246246248, "grad_norm": 5.0084123611450195, "learning_rate": 6.826821541710665e-06, "loss": 0.7568, "step": 3110 }, { "epoch": 1.6298472385428906, "grad_norm": 4.985651969909668, "learning_rate": 6.816261879619853e-06, "loss": 0.7554, "step": 3120 }, { "epoch": 1.6350698524611569, "grad_norm": 6.405339241027832, "learning_rate": 6.80570221752904e-06, "loss": 0.6628, "step": 3130 }, { "epoch": 1.640292466379423, "grad_norm": 5.617925643920898, "learning_rate": 6.795142555438227e-06, "loss": 0.6515, "step": 3140 }, { "epoch": 1.645515080297689, "grad_norm": 2.850438117980957, "learning_rate": 6.784582893347413e-06, "loss": 0.6283, "step": 3150 }, { "epoch": 1.650737694215955, "grad_norm": 7.605482578277588, "learning_rate": 6.7740232312566e-06, "loss": 0.6881, "step": 3160 }, { "epoch": 1.6559603081342211, "grad_norm": 4.773893356323242, "learning_rate": 6.763463569165787e-06, "loss": 0.6669, "step": 3170 }, { "epoch": 1.6611829220524874, "grad_norm": 4.882478713989258, "learning_rate": 6.752903907074974e-06, "loss": 0.6374, "step": 3180 }, { "epoch": 1.6664055359707532, "grad_norm": 6.456390380859375, "learning_rate": 6.7423442449841614e-06, "loss": 0.7759, "step": 3190 }, { "epoch": 1.6716281498890195, "grad_norm": 7.474002361297607, "learning_rate": 6.7317845828933485e-06, "loss": 0.6657, "step": 3200 }, { "epoch": 1.6716281498890195, "eval_loss": 0.8759788870811462, "eval_runtime": 46.2024, "eval_samples_per_second": 36.838, "eval_steps_per_second": 4.61, "step": 3200 }, { "epoch": 1.6768507638072856, "grad_norm": 4.953747272491455, "learning_rate": 6.721224920802535e-06, "loss": 0.7443, "step": 3210 }, { "epoch": 1.6820733777255517, "grad_norm": 5.401075839996338, "learning_rate": 6.710665258711722e-06, "loss": 0.7067, "step": 3220 }, { "epoch": 1.6872959916438177, "grad_norm": 5.775487899780273, "learning_rate": 6.700105596620908e-06, "loss": 0.78, "step": 3230 }, { "epoch": 1.6925186055620838, "grad_norm": 6.578312397003174, "learning_rate": 6.689545934530095e-06, "loss": 0.6206, "step": 3240 }, { "epoch": 1.69774121948035, "grad_norm": 7.354413986206055, "learning_rate": 6.678986272439283e-06, "loss": 0.7265, "step": 3250 }, { "epoch": 1.702963833398616, "grad_norm": 6.80817985534668, "learning_rate": 6.6684266103484695e-06, "loss": 0.8696, "step": 3260 }, { "epoch": 1.7081864473168822, "grad_norm": 6.331092357635498, "learning_rate": 6.657866948257657e-06, "loss": 0.7226, "step": 3270 }, { "epoch": 1.7134090612351482, "grad_norm": 6.063718795776367, "learning_rate": 6.647307286166843e-06, "loss": 0.6339, "step": 3280 }, { "epoch": 1.7186316751534143, "grad_norm": 4.693406105041504, "learning_rate": 6.63674762407603e-06, "loss": 0.7316, "step": 3290 }, { "epoch": 1.7238542890716804, "grad_norm": 6.732961654663086, "learning_rate": 6.626187961985216e-06, "loss": 0.7185, "step": 3300 }, { "epoch": 1.7290769029899464, "grad_norm": 4.482574939727783, "learning_rate": 6.615628299894404e-06, "loss": 0.7814, "step": 3310 }, { "epoch": 1.7342995169082127, "grad_norm": 7.299856662750244, "learning_rate": 6.605068637803591e-06, "loss": 0.7275, "step": 3320 }, { "epoch": 1.7395221308264786, "grad_norm": 4.190903186798096, "learning_rate": 6.5945089757127776e-06, "loss": 0.6967, "step": 3330 }, { "epoch": 1.7447447447447448, "grad_norm": 5.144697666168213, "learning_rate": 6.583949313621965e-06, "loss": 0.7279, "step": 3340 }, { "epoch": 1.7499673586630107, "grad_norm": 8.06899642944336, "learning_rate": 6.573389651531151e-06, "loss": 0.7133, "step": 3350 }, { "epoch": 1.755189972581277, "grad_norm": 5.388707637786865, "learning_rate": 6.562829989440338e-06, "loss": 0.7695, "step": 3360 }, { "epoch": 1.760412586499543, "grad_norm": 5.485361576080322, "learning_rate": 6.552270327349526e-06, "loss": 0.6475, "step": 3370 }, { "epoch": 1.765635200417809, "grad_norm": 5.023000717163086, "learning_rate": 6.541710665258712e-06, "loss": 0.6575, "step": 3380 }, { "epoch": 1.7708578143360754, "grad_norm": 5.406675815582275, "learning_rate": 6.531151003167899e-06, "loss": 0.7263, "step": 3390 }, { "epoch": 1.7760804282543412, "grad_norm": 3.564267873764038, "learning_rate": 6.520591341077086e-06, "loss": 0.598, "step": 3400 }, { "epoch": 1.7760804282543412, "eval_loss": 0.8584678769111633, "eval_runtime": 46.2248, "eval_samples_per_second": 36.82, "eval_steps_per_second": 4.608, "step": 3400 }, { "epoch": 1.7813030421726075, "grad_norm": 4.055863380432129, "learning_rate": 6.510031678986273e-06, "loss": 0.6674, "step": 3410 }, { "epoch": 1.7865256560908733, "grad_norm": 5.625813007354736, "learning_rate": 6.49947201689546e-06, "loss": 0.8424, "step": 3420 }, { "epoch": 1.7917482700091396, "grad_norm": 6.47999906539917, "learning_rate": 6.488912354804647e-06, "loss": 0.6039, "step": 3430 }, { "epoch": 1.7969708839274057, "grad_norm": 5.702643871307373, "learning_rate": 6.478352692713834e-06, "loss": 0.588, "step": 3440 }, { "epoch": 1.8021934978456717, "grad_norm": 6.600216388702393, "learning_rate": 6.467793030623021e-06, "loss": 0.7704, "step": 3450 }, { "epoch": 1.807416111763938, "grad_norm": 5.743258476257324, "learning_rate": 6.4572333685322074e-06, "loss": 0.6381, "step": 3460 }, { "epoch": 1.8126387256822039, "grad_norm": 7.323511123657227, "learning_rate": 6.4466737064413945e-06, "loss": 0.8124, "step": 3470 }, { "epoch": 1.8178613396004701, "grad_norm": 4.996503829956055, "learning_rate": 6.436114044350581e-06, "loss": 0.6816, "step": 3480 }, { "epoch": 1.823083953518736, "grad_norm": 6.126676559448242, "learning_rate": 6.425554382259768e-06, "loss": 0.6773, "step": 3490 }, { "epoch": 1.8283065674370023, "grad_norm": 4.893184185028076, "learning_rate": 6.414994720168956e-06, "loss": 0.684, "step": 3500 }, { "epoch": 1.8335291813552683, "grad_norm": 6.5332841873168945, "learning_rate": 6.404435058078142e-06, "loss": 0.771, "step": 3510 }, { "epoch": 1.8387517952735344, "grad_norm": 8.364972114562988, "learning_rate": 6.393875395987329e-06, "loss": 0.7224, "step": 3520 }, { "epoch": 1.8439744091918004, "grad_norm": 5.508388042449951, "learning_rate": 6.3833157338965155e-06, "loss": 0.7867, "step": 3530 }, { "epoch": 1.8491970231100665, "grad_norm": 6.582828044891357, "learning_rate": 6.372756071805703e-06, "loss": 0.5697, "step": 3540 }, { "epoch": 1.8544196370283328, "grad_norm": 6.311943531036377, "learning_rate": 6.362196409714889e-06, "loss": 0.6919, "step": 3550 }, { "epoch": 1.8596422509465986, "grad_norm": 8.718938827514648, "learning_rate": 6.351636747624077e-06, "loss": 0.685, "step": 3560 }, { "epoch": 1.864864864864865, "grad_norm": 7.9847846031188965, "learning_rate": 6.341077085533264e-06, "loss": 0.5817, "step": 3570 }, { "epoch": 1.870087478783131, "grad_norm": 5.257252216339111, "learning_rate": 6.33051742344245e-06, "loss": 0.7489, "step": 3580 }, { "epoch": 1.875310092701397, "grad_norm": 6.155455589294434, "learning_rate": 6.319957761351637e-06, "loss": 0.6976, "step": 3590 }, { "epoch": 1.880532706619663, "grad_norm": 5.958306789398193, "learning_rate": 6.3093980992608236e-06, "loss": 0.6385, "step": 3600 }, { "epoch": 1.880532706619663, "eval_loss": 0.8434246778488159, "eval_runtime": 46.2141, "eval_samples_per_second": 36.829, "eval_steps_per_second": 4.609, "step": 3600 }, { "epoch": 1.8857553205379292, "grad_norm": 4.219689846038818, "learning_rate": 6.298838437170011e-06, "loss": 0.6891, "step": 3610 }, { "epoch": 1.8909779344561954, "grad_norm": 6.840215682983398, "learning_rate": 6.288278775079199e-06, "loss": 0.5082, "step": 3620 }, { "epoch": 1.8962005483744613, "grad_norm": 4.201242923736572, "learning_rate": 6.277719112988385e-06, "loss": 0.6311, "step": 3630 }, { "epoch": 1.9014231622927276, "grad_norm": 4.635916709899902, "learning_rate": 6.267159450897572e-06, "loss": 0.5849, "step": 3640 }, { "epoch": 1.9066457762109936, "grad_norm": 9.082616806030273, "learning_rate": 6.256599788806758e-06, "loss": 0.7785, "step": 3650 }, { "epoch": 1.9118683901292597, "grad_norm": 7.005586624145508, "learning_rate": 6.246040126715945e-06, "loss": 0.5691, "step": 3660 }, { "epoch": 1.9170910040475257, "grad_norm": 6.064583778381348, "learning_rate": 6.2354804646251325e-06, "loss": 0.7137, "step": 3670 }, { "epoch": 1.9223136179657918, "grad_norm": 8.23308277130127, "learning_rate": 6.2249208025343196e-06, "loss": 0.7679, "step": 3680 }, { "epoch": 1.927536231884058, "grad_norm": 3.684671401977539, "learning_rate": 6.214361140443507e-06, "loss": 0.652, "step": 3690 }, { "epoch": 1.932758845802324, "grad_norm": 3.7538697719573975, "learning_rate": 6.203801478352694e-06, "loss": 0.7469, "step": 3700 }, { "epoch": 1.9379814597205902, "grad_norm": 6.125692844390869, "learning_rate": 6.19324181626188e-06, "loss": 0.775, "step": 3710 }, { "epoch": 1.9432040736388563, "grad_norm": 6.9215989112854, "learning_rate": 6.182682154171067e-06, "loss": 0.7417, "step": 3720 }, { "epoch": 1.9484266875571223, "grad_norm": 5.37251615524292, "learning_rate": 6.172122492080253e-06, "loss": 0.6806, "step": 3730 }, { "epoch": 1.9536493014753884, "grad_norm": 6.892059803009033, "learning_rate": 6.1615628299894405e-06, "loss": 0.72, "step": 3740 }, { "epoch": 1.9588719153936545, "grad_norm": 4.601454257965088, "learning_rate": 6.1510031678986285e-06, "loss": 0.5358, "step": 3750 }, { "epoch": 1.9640945293119207, "grad_norm": 6.319756984710693, "learning_rate": 6.140443505807815e-06, "loss": 0.6178, "step": 3760 }, { "epoch": 1.9693171432301866, "grad_norm": 5.619534015655518, "learning_rate": 6.129883843717002e-06, "loss": 0.6324, "step": 3770 }, { "epoch": 1.9745397571484529, "grad_norm": 6.824273109436035, "learning_rate": 6.119324181626188e-06, "loss": 0.6543, "step": 3780 }, { "epoch": 1.979762371066719, "grad_norm": 4.03241491317749, "learning_rate": 6.108764519535375e-06, "loss": 0.6577, "step": 3790 }, { "epoch": 1.984984984984985, "grad_norm": 8.365711212158203, "learning_rate": 6.0982048574445615e-06, "loss": 0.7096, "step": 3800 }, { "epoch": 1.984984984984985, "eval_loss": 0.8318305611610413, "eval_runtime": 46.2366, "eval_samples_per_second": 36.811, "eval_steps_per_second": 4.607, "step": 3800 }, { "epoch": 1.990207598903251, "grad_norm": 4.457459926605225, "learning_rate": 6.0876451953537494e-06, "loss": 0.7665, "step": 3810 }, { "epoch": 1.9954302128215171, "grad_norm": 5.687203407287598, "learning_rate": 6.0770855332629365e-06, "loss": 0.6722, "step": 3820 }, { "epoch": 2.001044522783653, "grad_norm": 6.511070251464844, "learning_rate": 6.066525871172123e-06, "loss": 0.7669, "step": 3830 }, { "epoch": 2.0062671367019194, "grad_norm": 5.137298583984375, "learning_rate": 6.05596620908131e-06, "loss": 0.5114, "step": 3840 }, { "epoch": 2.011489750620185, "grad_norm": 7.063135623931885, "learning_rate": 6.045406546990496e-06, "loss": 0.5532, "step": 3850 }, { "epoch": 2.0167123645384515, "grad_norm": 4.833804607391357, "learning_rate": 6.034846884899683e-06, "loss": 0.4304, "step": 3860 }, { "epoch": 2.0219349784567178, "grad_norm": 6.820064544677734, "learning_rate": 6.024287222808871e-06, "loss": 0.5066, "step": 3870 }, { "epoch": 2.0271575923749836, "grad_norm": 6.833749771118164, "learning_rate": 6.0137275607180575e-06, "loss": 0.5043, "step": 3880 }, { "epoch": 2.03238020629325, "grad_norm": 4.371280670166016, "learning_rate": 6.003167898627245e-06, "loss": 0.6015, "step": 3890 }, { "epoch": 2.0376028202115157, "grad_norm": 4.963273048400879, "learning_rate": 5.992608236536432e-06, "loss": 0.5881, "step": 3900 }, { "epoch": 2.042825434129782, "grad_norm": 5.74879264831543, "learning_rate": 5.982048574445618e-06, "loss": 0.5776, "step": 3910 }, { "epoch": 2.048048048048048, "grad_norm": 3.3885014057159424, "learning_rate": 5.971488912354805e-06, "loss": 0.6508, "step": 3920 }, { "epoch": 2.053270661966314, "grad_norm": 3.875781774520874, "learning_rate": 5.960929250263991e-06, "loss": 0.5135, "step": 3930 }, { "epoch": 2.0584932758845804, "grad_norm": 5.869786739349365, "learning_rate": 5.950369588173179e-06, "loss": 0.5343, "step": 3940 }, { "epoch": 2.0637158898028463, "grad_norm": 5.666866779327393, "learning_rate": 5.939809926082366e-06, "loss": 0.5258, "step": 3950 }, { "epoch": 2.0689385037211125, "grad_norm": 5.5713276863098145, "learning_rate": 5.929250263991553e-06, "loss": 0.6209, "step": 3960 }, { "epoch": 2.0741611176393784, "grad_norm": 5.73265266418457, "learning_rate": 5.91869060190074e-06, "loss": 0.4707, "step": 3970 }, { "epoch": 2.0793837315576447, "grad_norm": 5.312356948852539, "learning_rate": 5.908130939809926e-06, "loss": 0.5429, "step": 3980 }, { "epoch": 2.0846063454759105, "grad_norm": 5.636459827423096, "learning_rate": 5.897571277719113e-06, "loss": 0.5524, "step": 3990 }, { "epoch": 2.089828959394177, "grad_norm": 5.541628360748291, "learning_rate": 5.887011615628301e-06, "loss": 0.6338, "step": 4000 }, { "epoch": 2.089828959394177, "eval_loss": 0.8418287038803101, "eval_runtime": 46.2756, "eval_samples_per_second": 36.78, "eval_steps_per_second": 4.603, "step": 4000 }, { "epoch": 2.095051573312443, "grad_norm": 7.190380096435547, "learning_rate": 5.876451953537487e-06, "loss": 0.5818, "step": 4010 }, { "epoch": 2.100274187230709, "grad_norm": 6.190309047698975, "learning_rate": 5.8658922914466745e-06, "loss": 0.5268, "step": 4020 }, { "epoch": 2.105496801148975, "grad_norm": 5.186282157897949, "learning_rate": 5.855332629355861e-06, "loss": 0.4396, "step": 4030 }, { "epoch": 2.110719415067241, "grad_norm": 3.6927499771118164, "learning_rate": 5.844772967265048e-06, "loss": 0.6601, "step": 4040 }, { "epoch": 2.1159420289855073, "grad_norm": 5.906070232391357, "learning_rate": 5.834213305174234e-06, "loss": 0.5902, "step": 4050 }, { "epoch": 2.121164642903773, "grad_norm": 5.765960216522217, "learning_rate": 5.823653643083422e-06, "loss": 0.6127, "step": 4060 }, { "epoch": 2.1263872568220394, "grad_norm": 6.663849830627441, "learning_rate": 5.813093980992609e-06, "loss": 0.5044, "step": 4070 }, { "epoch": 2.1316098707403057, "grad_norm": 5.963075637817383, "learning_rate": 5.8025343189017954e-06, "loss": 0.4473, "step": 4080 }, { "epoch": 2.1368324846585716, "grad_norm": 5.396392345428467, "learning_rate": 5.7919746568109825e-06, "loss": 0.4896, "step": 4090 }, { "epoch": 2.142055098576838, "grad_norm": 6.087408542633057, "learning_rate": 5.781414994720169e-06, "loss": 0.5368, "step": 4100 }, { "epoch": 2.1472777124951037, "grad_norm": 4.697368144989014, "learning_rate": 5.770855332629356e-06, "loss": 0.6034, "step": 4110 }, { "epoch": 2.15250032641337, "grad_norm": 5.101240634918213, "learning_rate": 5.760295670538544e-06, "loss": 0.4699, "step": 4120 }, { "epoch": 2.157722940331636, "grad_norm": 6.5133891105651855, "learning_rate": 5.74973600844773e-06, "loss": 0.4589, "step": 4130 }, { "epoch": 2.162945554249902, "grad_norm": 7.932409763336182, "learning_rate": 5.739176346356917e-06, "loss": 0.5443, "step": 4140 }, { "epoch": 2.1681681681681684, "grad_norm": 4.897655010223389, "learning_rate": 5.728616684266104e-06, "loss": 0.5367, "step": 4150 }, { "epoch": 2.173390782086434, "grad_norm": 6.237987041473389, "learning_rate": 5.718057022175291e-06, "loss": 0.5477, "step": 4160 }, { "epoch": 2.1786133960047005, "grad_norm": 5.690924167633057, "learning_rate": 5.707497360084478e-06, "loss": 0.6216, "step": 4170 }, { "epoch": 2.1838360099229663, "grad_norm": 5.274245738983154, "learning_rate": 5.696937697993664e-06, "loss": 0.5441, "step": 4180 }, { "epoch": 2.1890586238412326, "grad_norm": 6.222249984741211, "learning_rate": 5.686378035902852e-06, "loss": 0.5579, "step": 4190 }, { "epoch": 2.1942812377594985, "grad_norm": 6.638361930847168, "learning_rate": 5.675818373812039e-06, "loss": 0.6108, "step": 4200 }, { "epoch": 2.1942812377594985, "eval_loss": 0.8388937711715698, "eval_runtime": 46.2559, "eval_samples_per_second": 36.795, "eval_steps_per_second": 4.605, "step": 4200 }, { "epoch": 2.1995038516777647, "grad_norm": 5.303590297698975, "learning_rate": 5.665258711721225e-06, "loss": 0.5602, "step": 4210 }, { "epoch": 2.204726465596031, "grad_norm": 4.8176727294921875, "learning_rate": 5.654699049630412e-06, "loss": 0.5329, "step": 4220 }, { "epoch": 2.209949079514297, "grad_norm": 7.120988368988037, "learning_rate": 5.644139387539599e-06, "loss": 0.4963, "step": 4230 }, { "epoch": 2.215171693432563, "grad_norm": 5.514588832855225, "learning_rate": 5.633579725448786e-06, "loss": 0.6124, "step": 4240 }, { "epoch": 2.220394307350829, "grad_norm": 5.315512657165527, "learning_rate": 5.623020063357974e-06, "loss": 0.5925, "step": 4250 }, { "epoch": 2.2256169212690953, "grad_norm": 5.721431732177734, "learning_rate": 5.61246040126716e-06, "loss": 0.6081, "step": 4260 }, { "epoch": 2.230839535187361, "grad_norm": 7.8369832038879395, "learning_rate": 5.601900739176347e-06, "loss": 0.4506, "step": 4270 }, { "epoch": 2.2360621491056274, "grad_norm": 5.328855514526367, "learning_rate": 5.591341077085533e-06, "loss": 0.61, "step": 4280 }, { "epoch": 2.2412847630238932, "grad_norm": 4.6545891761779785, "learning_rate": 5.5807814149947205e-06, "loss": 0.5102, "step": 4290 }, { "epoch": 2.2465073769421595, "grad_norm": 4.399157524108887, "learning_rate": 5.570221752903907e-06, "loss": 0.5495, "step": 4300 }, { "epoch": 2.251729990860426, "grad_norm": 8.38592529296875, "learning_rate": 5.559662090813095e-06, "loss": 0.5832, "step": 4310 }, { "epoch": 2.2569526047786916, "grad_norm": 3.9857163429260254, "learning_rate": 5.549102428722282e-06, "loss": 0.4725, "step": 4320 }, { "epoch": 2.262175218696958, "grad_norm": 5.648230075836182, "learning_rate": 5.538542766631468e-06, "loss": 0.5457, "step": 4330 }, { "epoch": 2.2673978326152238, "grad_norm": 3.6229002475738525, "learning_rate": 5.527983104540655e-06, "loss": 0.5997, "step": 4340 }, { "epoch": 2.27262044653349, "grad_norm": 6.094500541687012, "learning_rate": 5.517423442449842e-06, "loss": 0.5271, "step": 4350 }, { "epoch": 2.2778430604517563, "grad_norm": 7.447030544281006, "learning_rate": 5.5068637803590285e-06, "loss": 0.6592, "step": 4360 }, { "epoch": 2.283065674370022, "grad_norm": 6.060546398162842, "learning_rate": 5.4963041182682165e-06, "loss": 0.5148, "step": 4370 }, { "epoch": 2.2882882882882885, "grad_norm": 6.3843092918396, "learning_rate": 5.485744456177403e-06, "loss": 0.5469, "step": 4380 }, { "epoch": 2.2935109022065543, "grad_norm": 5.431898593902588, "learning_rate": 5.47518479408659e-06, "loss": 0.5922, "step": 4390 }, { "epoch": 2.2987335161248206, "grad_norm": 7.113710403442383, "learning_rate": 5.464625131995777e-06, "loss": 0.6052, "step": 4400 }, { "epoch": 2.2987335161248206, "eval_loss": 0.8255796432495117, "eval_runtime": 46.2225, "eval_samples_per_second": 36.822, "eval_steps_per_second": 4.608, "step": 4400 }, { "epoch": 2.3039561300430864, "grad_norm": 5.431846618652344, "learning_rate": 5.454065469904963e-06, "loss": 0.5565, "step": 4410 }, { "epoch": 2.3091787439613527, "grad_norm": 4.409207344055176, "learning_rate": 5.44350580781415e-06, "loss": 0.6042, "step": 4420 }, { "epoch": 2.3144013578796185, "grad_norm": 6.498474597930908, "learning_rate": 5.432946145723337e-06, "loss": 0.526, "step": 4430 }, { "epoch": 2.319623971797885, "grad_norm": 5.55870246887207, "learning_rate": 5.4223864836325246e-06, "loss": 0.6467, "step": 4440 }, { "epoch": 2.324846585716151, "grad_norm": 5.727528095245361, "learning_rate": 5.411826821541712e-06, "loss": 0.5701, "step": 4450 }, { "epoch": 2.330069199634417, "grad_norm": 5.781177520751953, "learning_rate": 5.401267159450898e-06, "loss": 0.5244, "step": 4460 }, { "epoch": 2.3352918135526832, "grad_norm": 3.1030118465423584, "learning_rate": 5.390707497360085e-06, "loss": 0.4523, "step": 4470 }, { "epoch": 2.340514427470949, "grad_norm": 3.780679941177368, "learning_rate": 5.380147835269271e-06, "loss": 0.5043, "step": 4480 }, { "epoch": 2.3457370413892153, "grad_norm": 7.008651256561279, "learning_rate": 5.369588173178458e-06, "loss": 0.4744, "step": 4490 }, { "epoch": 2.3509596553074816, "grad_norm": 5.0416693687438965, "learning_rate": 5.359028511087646e-06, "loss": 0.485, "step": 4500 }, { "epoch": 2.3561822692257475, "grad_norm": 6.809351921081543, "learning_rate": 5.348468848996833e-06, "loss": 0.6127, "step": 4510 }, { "epoch": 2.3614048831440138, "grad_norm": 5.149819374084473, "learning_rate": 5.33790918690602e-06, "loss": 0.5398, "step": 4520 }, { "epoch": 2.3666274970622796, "grad_norm": 6.1692962646484375, "learning_rate": 5.327349524815206e-06, "loss": 0.5371, "step": 4530 }, { "epoch": 2.371850110980546, "grad_norm": 7.224261283874512, "learning_rate": 5.316789862724393e-06, "loss": 0.5493, "step": 4540 }, { "epoch": 2.3770727248988117, "grad_norm": 6.257209777832031, "learning_rate": 5.306230200633579e-06, "loss": 0.5177, "step": 4550 }, { "epoch": 2.382295338817078, "grad_norm": 5.789205074310303, "learning_rate": 5.295670538542767e-06, "loss": 0.5016, "step": 4560 }, { "epoch": 2.387517952735344, "grad_norm": 6.116456508636475, "learning_rate": 5.285110876451954e-06, "loss": 0.6088, "step": 4570 }, { "epoch": 2.39274056665361, "grad_norm": 5.818353176116943, "learning_rate": 5.274551214361141e-06, "loss": 0.5105, "step": 4580 }, { "epoch": 2.3979631805718764, "grad_norm": 5.000683784484863, "learning_rate": 5.263991552270328e-06, "loss": 0.5899, "step": 4590 }, { "epoch": 2.4031857944901422, "grad_norm": 6.238855838775635, "learning_rate": 5.253431890179515e-06, "loss": 0.6326, "step": 4600 }, { "epoch": 2.4031857944901422, "eval_loss": 0.8141046166419983, "eval_runtime": 46.2243, "eval_samples_per_second": 36.82, "eval_steps_per_second": 4.608, "step": 4600 }, { "epoch": 2.4084084084084085, "grad_norm": 4.040828704833984, "learning_rate": 5.242872228088701e-06, "loss": 0.4504, "step": 4610 }, { "epoch": 2.4136310223266744, "grad_norm": 4.401372909545898, "learning_rate": 5.232312565997888e-06, "loss": 0.4497, "step": 4620 }, { "epoch": 2.4188536362449407, "grad_norm": 5.584476947784424, "learning_rate": 5.221752903907075e-06, "loss": 0.6214, "step": 4630 }, { "epoch": 2.4240762501632065, "grad_norm": 3.0335025787353516, "learning_rate": 5.2111932418162625e-06, "loss": 0.5114, "step": 4640 }, { "epoch": 2.4292988640814728, "grad_norm": 6.4493727684021, "learning_rate": 5.20063357972545e-06, "loss": 0.6065, "step": 4650 }, { "epoch": 2.434521477999739, "grad_norm": 4.674168109893799, "learning_rate": 5.190073917634636e-06, "loss": 0.4778, "step": 4660 }, { "epoch": 2.439744091918005, "grad_norm": 6.156538963317871, "learning_rate": 5.179514255543823e-06, "loss": 0.489, "step": 4670 }, { "epoch": 2.444966705836271, "grad_norm": 5.803397178649902, "learning_rate": 5.168954593453009e-06, "loss": 0.5948, "step": 4680 }, { "epoch": 2.450189319754537, "grad_norm": 6.375555992126465, "learning_rate": 5.158394931362197e-06, "loss": 0.4768, "step": 4690 }, { "epoch": 2.4554119336728033, "grad_norm": 6.442558288574219, "learning_rate": 5.147835269271384e-06, "loss": 0.4996, "step": 4700 }, { "epoch": 2.460634547591069, "grad_norm": 8.586763381958008, "learning_rate": 5.1372756071805705e-06, "loss": 0.7869, "step": 4710 }, { "epoch": 2.4658571615093354, "grad_norm": 6.680276393890381, "learning_rate": 5.126715945089758e-06, "loss": 0.5526, "step": 4720 }, { "epoch": 2.4710797754276017, "grad_norm": 3.914189100265503, "learning_rate": 5.116156282998944e-06, "loss": 0.6078, "step": 4730 }, { "epoch": 2.4763023893458675, "grad_norm": 7.448416233062744, "learning_rate": 5.105596620908131e-06, "loss": 0.4982, "step": 4740 }, { "epoch": 2.481525003264134, "grad_norm": 7.687001705169678, "learning_rate": 5.095036958817319e-06, "loss": 0.6684, "step": 4750 }, { "epoch": 2.4867476171823997, "grad_norm": 6.067854404449463, "learning_rate": 5.084477296726505e-06, "loss": 0.5189, "step": 4760 }, { "epoch": 2.491970231100666, "grad_norm": 5.405977725982666, "learning_rate": 5.073917634635692e-06, "loss": 0.5212, "step": 4770 }, { "epoch": 2.497192845018932, "grad_norm": 7.773893356323242, "learning_rate": 5.063357972544879e-06, "loss": 0.6681, "step": 4780 }, { "epoch": 2.502415458937198, "grad_norm": 9.031746864318848, "learning_rate": 5.052798310454066e-06, "loss": 0.6579, "step": 4790 }, { "epoch": 2.507638072855464, "grad_norm": 5.8720245361328125, "learning_rate": 5.042238648363252e-06, "loss": 0.5891, "step": 4800 }, { "epoch": 2.507638072855464, "eval_loss": 0.8107092380523682, "eval_runtime": 46.2852, "eval_samples_per_second": 36.772, "eval_steps_per_second": 4.602, "step": 4800 }, { "epoch": 2.51286068677373, "grad_norm": 4.402786731719971, "learning_rate": 5.03167898627244e-06, "loss": 0.4713, "step": 4810 }, { "epoch": 2.5180833006919965, "grad_norm": 5.443326473236084, "learning_rate": 5.021119324181627e-06, "loss": 0.6362, "step": 4820 }, { "epoch": 2.5233059146102623, "grad_norm": 6.188055515289307, "learning_rate": 5.010559662090813e-06, "loss": 0.6042, "step": 4830 }, { "epoch": 2.5285285285285286, "grad_norm": 5.5944600105285645, "learning_rate": 5e-06, "loss": 0.5093, "step": 4840 }, { "epoch": 2.5337511424467944, "grad_norm": 6.214510917663574, "learning_rate": 4.9894403379091875e-06, "loss": 0.5405, "step": 4850 }, { "epoch": 2.5389737563650607, "grad_norm": 4.829537868499756, "learning_rate": 4.978880675818375e-06, "loss": 0.52, "step": 4860 }, { "epoch": 2.544196370283327, "grad_norm": 5.498637676239014, "learning_rate": 4.968321013727561e-06, "loss": 0.4998, "step": 4870 }, { "epoch": 2.549418984201593, "grad_norm": 6.1551361083984375, "learning_rate": 4.957761351636748e-06, "loss": 0.5842, "step": 4880 }, { "epoch": 2.554641598119859, "grad_norm": 5.512228488922119, "learning_rate": 4.947201689545935e-06, "loss": 0.5877, "step": 4890 }, { "epoch": 2.559864212038125, "grad_norm": 6.113735675811768, "learning_rate": 4.936642027455122e-06, "loss": 0.4979, "step": 4900 }, { "epoch": 2.5650868259563913, "grad_norm": 9.69180965423584, "learning_rate": 4.9260823653643085e-06, "loss": 0.5595, "step": 4910 }, { "epoch": 2.5703094398746575, "grad_norm": 6.716381072998047, "learning_rate": 4.915522703273496e-06, "loss": 0.577, "step": 4920 }, { "epoch": 2.5755320537929234, "grad_norm": 6.1616926193237305, "learning_rate": 4.904963041182683e-06, "loss": 0.4457, "step": 4930 }, { "epoch": 2.580754667711189, "grad_norm": 5.395188331604004, "learning_rate": 4.894403379091869e-06, "loss": 0.4934, "step": 4940 }, { "epoch": 2.5859772816294555, "grad_norm": 5.993736743927002, "learning_rate": 4.883843717001057e-06, "loss": 0.5133, "step": 4950 }, { "epoch": 2.591199895547722, "grad_norm": 4.6023054122924805, "learning_rate": 4.873284054910243e-06, "loss": 0.5084, "step": 4960 }, { "epoch": 2.5964225094659876, "grad_norm": 4.444733619689941, "learning_rate": 4.86272439281943e-06, "loss": 0.5717, "step": 4970 }, { "epoch": 2.601645123384254, "grad_norm": 6.304750442504883, "learning_rate": 4.8521647307286165e-06, "loss": 0.5583, "step": 4980 }, { "epoch": 2.6068677373025197, "grad_norm": 5.222369194030762, "learning_rate": 4.8416050686378045e-06, "loss": 0.5425, "step": 4990 }, { "epoch": 2.612090351220786, "grad_norm": 5.956515789031982, "learning_rate": 4.831045406546991e-06, "loss": 0.5626, "step": 5000 }, { "epoch": 2.612090351220786, "eval_loss": 0.8068883419036865, "eval_runtime": 46.2364, "eval_samples_per_second": 36.811, "eval_steps_per_second": 4.607, "step": 5000 }, { "epoch": 2.6173129651390523, "grad_norm": 6.04465913772583, "learning_rate": 4.820485744456178e-06, "loss": 0.6118, "step": 5010 }, { "epoch": 2.622535579057318, "grad_norm": 4.643867015838623, "learning_rate": 4.809926082365365e-06, "loss": 0.6346, "step": 5020 }, { "epoch": 2.6277581929755844, "grad_norm": 5.192962646484375, "learning_rate": 4.799366420274551e-06, "loss": 0.539, "step": 5030 }, { "epoch": 2.6329808068938503, "grad_norm": 5.151241779327393, "learning_rate": 4.788806758183738e-06, "loss": 0.552, "step": 5040 }, { "epoch": 2.6382034208121166, "grad_norm": 4.308994293212891, "learning_rate": 4.7782470960929254e-06, "loss": 0.5292, "step": 5050 }, { "epoch": 2.6434260347303824, "grad_norm": 5.595186233520508, "learning_rate": 4.7676874340021126e-06, "loss": 0.4777, "step": 5060 }, { "epoch": 2.6486486486486487, "grad_norm": 6.7816643714904785, "learning_rate": 4.757127771911299e-06, "loss": 0.4999, "step": 5070 }, { "epoch": 2.6538712625669145, "grad_norm": 7.1444315910339355, "learning_rate": 4.746568109820486e-06, "loss": 0.5028, "step": 5080 }, { "epoch": 2.659093876485181, "grad_norm": 5.04287576675415, "learning_rate": 4.736008447729673e-06, "loss": 0.5156, "step": 5090 }, { "epoch": 2.664316490403447, "grad_norm": 4.410764694213867, "learning_rate": 4.72544878563886e-06, "loss": 0.5241, "step": 5100 }, { "epoch": 2.669539104321713, "grad_norm": 4.788335800170898, "learning_rate": 4.714889123548047e-06, "loss": 0.5099, "step": 5110 }, { "epoch": 2.674761718239979, "grad_norm": 5.737407207489014, "learning_rate": 4.7043294614572335e-06, "loss": 0.4395, "step": 5120 }, { "epoch": 2.679984332158245, "grad_norm": 6.255344867706299, "learning_rate": 4.693769799366421e-06, "loss": 0.5994, "step": 5130 }, { "epoch": 2.6852069460765113, "grad_norm": 5.028295516967773, "learning_rate": 4.683210137275608e-06, "loss": 0.5518, "step": 5140 }, { "epoch": 2.6904295599947776, "grad_norm": 4.391537189483643, "learning_rate": 4.672650475184795e-06, "loss": 0.4991, "step": 5150 }, { "epoch": 2.6956521739130435, "grad_norm": 5.823568820953369, "learning_rate": 4.662090813093981e-06, "loss": 0.5885, "step": 5160 }, { "epoch": 2.7008747878313093, "grad_norm": 2.928971529006958, "learning_rate": 4.651531151003168e-06, "loss": 0.6942, "step": 5170 }, { "epoch": 2.7060974017495756, "grad_norm": 9.696499824523926, "learning_rate": 4.640971488912355e-06, "loss": 0.5699, "step": 5180 }, { "epoch": 2.711320015667842, "grad_norm": 4.2604217529296875, "learning_rate": 4.630411826821542e-06, "loss": 0.4676, "step": 5190 }, { "epoch": 2.7165426295861077, "grad_norm": 5.672421932220459, "learning_rate": 4.619852164730729e-06, "loss": 0.4576, "step": 5200 }, { "epoch": 2.7165426295861077, "eval_loss": 0.7933436632156372, "eval_runtime": 46.2281, "eval_samples_per_second": 36.817, "eval_steps_per_second": 4.608, "step": 5200 }, { "epoch": 2.721765243504374, "grad_norm": 4.114403247833252, "learning_rate": 4.609292502639916e-06, "loss": 0.4742, "step": 5210 }, { "epoch": 2.72698785742264, "grad_norm": 5.642977237701416, "learning_rate": 4.598732840549103e-06, "loss": 0.5453, "step": 5220 }, { "epoch": 2.732210471340906, "grad_norm": 5.690392017364502, "learning_rate": 4.588173178458289e-06, "loss": 0.5248, "step": 5230 }, { "epoch": 2.7374330852591724, "grad_norm": 7.28243350982666, "learning_rate": 4.577613516367477e-06, "loss": 0.5059, "step": 5240 }, { "epoch": 2.7426556991774382, "grad_norm": 5.291462421417236, "learning_rate": 4.567053854276663e-06, "loss": 0.4494, "step": 5250 }, { "epoch": 2.7478783130957045, "grad_norm": 2.6974287033081055, "learning_rate": 4.5564941921858505e-06, "loss": 0.5439, "step": 5260 }, { "epoch": 2.7531009270139704, "grad_norm": 4.0036940574646, "learning_rate": 4.545934530095038e-06, "loss": 0.4797, "step": 5270 }, { "epoch": 2.7583235409322366, "grad_norm": 5.800724506378174, "learning_rate": 4.535374868004224e-06, "loss": 0.5122, "step": 5280 }, { "epoch": 2.763546154850503, "grad_norm": 6.420878887176514, "learning_rate": 4.524815205913411e-06, "loss": 0.4956, "step": 5290 }, { "epoch": 2.7687687687687688, "grad_norm": 6.129545211791992, "learning_rate": 4.514255543822598e-06, "loss": 0.5465, "step": 5300 }, { "epoch": 2.7739913826870346, "grad_norm": 5.964089870452881, "learning_rate": 4.503695881731785e-06, "loss": 0.5012, "step": 5310 }, { "epoch": 2.779213996605301, "grad_norm": 5.476171493530273, "learning_rate": 4.4931362196409714e-06, "loss": 0.4842, "step": 5320 }, { "epoch": 2.784436610523567, "grad_norm": 3.6184587478637695, "learning_rate": 4.4825765575501585e-06, "loss": 0.464, "step": 5330 }, { "epoch": 2.789659224441833, "grad_norm": 6.052497863769531, "learning_rate": 4.472016895459346e-06, "loss": 0.5625, "step": 5340 }, { "epoch": 2.7948818383600993, "grad_norm": 8.540828704833984, "learning_rate": 4.461457233368533e-06, "loss": 0.581, "step": 5350 }, { "epoch": 2.800104452278365, "grad_norm": 6.093939781188965, "learning_rate": 4.45089757127772e-06, "loss": 0.5907, "step": 5360 }, { "epoch": 2.8053270661966314, "grad_norm": 4.6257405281066895, "learning_rate": 4.440337909186906e-06, "loss": 0.4679, "step": 5370 }, { "epoch": 2.8105496801148977, "grad_norm": 4.155122756958008, "learning_rate": 4.429778247096093e-06, "loss": 0.4799, "step": 5380 }, { "epoch": 2.8157722940331635, "grad_norm": 6.194579601287842, "learning_rate": 4.41921858500528e-06, "loss": 0.5603, "step": 5390 }, { "epoch": 2.82099490795143, "grad_norm": 7.508232593536377, "learning_rate": 4.4086589229144675e-06, "loss": 0.6672, "step": 5400 }, { "epoch": 2.82099490795143, "eval_loss": 0.7919074296951294, "eval_runtime": 46.2409, "eval_samples_per_second": 36.807, "eval_steps_per_second": 4.606, "step": 5400 }, { "epoch": 2.8262175218696957, "grad_norm": 6.648965358734131, "learning_rate": 4.398099260823654e-06, "loss": 0.6096, "step": 5410 }, { "epoch": 2.831440135787962, "grad_norm": 6.033213138580322, "learning_rate": 4.387539598732841e-06, "loss": 0.5375, "step": 5420 }, { "epoch": 2.8366627497062282, "grad_norm": 5.509268760681152, "learning_rate": 4.376979936642028e-06, "loss": 0.4289, "step": 5430 }, { "epoch": 2.841885363624494, "grad_norm": 4.335843563079834, "learning_rate": 4.366420274551215e-06, "loss": 0.443, "step": 5440 }, { "epoch": 2.84710797754276, "grad_norm": 3.9269707202911377, "learning_rate": 4.355860612460401e-06, "loss": 0.4666, "step": 5450 }, { "epoch": 2.852330591461026, "grad_norm": 7.474977970123291, "learning_rate": 4.345300950369588e-06, "loss": 0.5793, "step": 5460 }, { "epoch": 2.8575532053792925, "grad_norm": 3.6517860889434814, "learning_rate": 4.3347412882787755e-06, "loss": 0.4276, "step": 5470 }, { "epoch": 2.8627758192975583, "grad_norm": 4.662909030914307, "learning_rate": 4.324181626187962e-06, "loss": 0.5647, "step": 5480 }, { "epoch": 2.8679984332158246, "grad_norm": 6.30706787109375, "learning_rate": 4.31362196409715e-06, "loss": 0.5985, "step": 5490 }, { "epoch": 2.8732210471340904, "grad_norm": 3.8878538608551025, "learning_rate": 4.303062302006336e-06, "loss": 0.4561, "step": 5500 }, { "epoch": 2.8784436610523567, "grad_norm": 4.7928667068481445, "learning_rate": 4.292502639915523e-06, "loss": 0.5941, "step": 5510 }, { "epoch": 2.883666274970623, "grad_norm": 7.024383544921875, "learning_rate": 4.28194297782471e-06, "loss": 0.51, "step": 5520 }, { "epoch": 2.888888888888889, "grad_norm": 4.680624961853027, "learning_rate": 4.2713833157338965e-06, "loss": 0.5751, "step": 5530 }, { "epoch": 2.894111502807155, "grad_norm": 5.982644081115723, "learning_rate": 4.260823653643084e-06, "loss": 0.5139, "step": 5540 }, { "epoch": 2.899334116725421, "grad_norm": 5.7597784996032715, "learning_rate": 4.250263991552271e-06, "loss": 0.5367, "step": 5550 }, { "epoch": 2.9045567306436872, "grad_norm": 5.401406764984131, "learning_rate": 4.239704329461458e-06, "loss": 0.5542, "step": 5560 }, { "epoch": 2.9097793445619535, "grad_norm": 8.459036827087402, "learning_rate": 4.229144667370644e-06, "loss": 0.5354, "step": 5570 }, { "epoch": 2.9150019584802194, "grad_norm": 6.203250885009766, "learning_rate": 4.218585005279832e-06, "loss": 0.5376, "step": 5580 }, { "epoch": 2.920224572398485, "grad_norm": 6.801321983337402, "learning_rate": 4.208025343189018e-06, "loss": 0.4566, "step": 5590 }, { "epoch": 2.9254471863167515, "grad_norm": 2.9718496799468994, "learning_rate": 4.197465681098205e-06, "loss": 0.483, "step": 5600 }, { "epoch": 2.9254471863167515, "eval_loss": 0.7785268425941467, "eval_runtime": 46.2314, "eval_samples_per_second": 36.815, "eval_steps_per_second": 4.607, "step": 5600 }, { "epoch": 2.9306698002350178, "grad_norm": 3.2528350353240967, "learning_rate": 4.1869060190073925e-06, "loss": 0.5526, "step": 5610 }, { "epoch": 2.9358924141532836, "grad_norm": 3.9217379093170166, "learning_rate": 4.176346356916579e-06, "loss": 0.558, "step": 5620 }, { "epoch": 2.94111502807155, "grad_norm": 4.15424919128418, "learning_rate": 4.165786694825766e-06, "loss": 0.4264, "step": 5630 }, { "epoch": 2.9463376419898157, "grad_norm": 5.303351402282715, "learning_rate": 4.155227032734953e-06, "loss": 0.5739, "step": 5640 }, { "epoch": 2.951560255908082, "grad_norm": 7.237427234649658, "learning_rate": 4.14466737064414e-06, "loss": 0.5442, "step": 5650 }, { "epoch": 2.9567828698263483, "grad_norm": 4.967709541320801, "learning_rate": 4.134107708553326e-06, "loss": 0.4926, "step": 5660 }, { "epoch": 2.962005483744614, "grad_norm": 5.920149326324463, "learning_rate": 4.1235480464625134e-06, "loss": 0.5655, "step": 5670 }, { "epoch": 2.9672280976628804, "grad_norm": 5.823659896850586, "learning_rate": 4.1129883843717006e-06, "loss": 0.5124, "step": 5680 }, { "epoch": 2.9724507115811463, "grad_norm": 3.851020336151123, "learning_rate": 4.102428722280888e-06, "loss": 0.423, "step": 5690 }, { "epoch": 2.9776733254994125, "grad_norm": 4.632102012634277, "learning_rate": 4.091869060190074e-06, "loss": 0.42, "step": 5700 }, { "epoch": 2.9828959394176784, "grad_norm": 6.207057476043701, "learning_rate": 4.081309398099261e-06, "loss": 0.4569, "step": 5710 }, { "epoch": 2.9881185533359447, "grad_norm": 4.414632797241211, "learning_rate": 4.070749736008448e-06, "loss": 0.4628, "step": 5720 }, { "epoch": 2.9933411672542105, "grad_norm": 5.721477508544922, "learning_rate": 4.060190073917634e-06, "loss": 0.4518, "step": 5730 }, { "epoch": 2.998563781172477, "grad_norm": 5.436526298522949, "learning_rate": 4.049630411826822e-06, "loss": 0.5313, "step": 5740 }, { "epoch": 3.0041780911346128, "grad_norm": 7.654147148132324, "learning_rate": 4.039070749736009e-06, "loss": 0.494, "step": 5750 }, { "epoch": 3.009400705052879, "grad_norm": 5.68324089050293, "learning_rate": 4.028511087645196e-06, "loss": 0.3699, "step": 5760 }, { "epoch": 3.014623318971145, "grad_norm": 5.219386577606201, "learning_rate": 4.017951425554383e-06, "loss": 0.4355, "step": 5770 }, { "epoch": 3.019845932889411, "grad_norm": 6.570154190063477, "learning_rate": 4.007391763463569e-06, "loss": 0.4075, "step": 5780 }, { "epoch": 3.0250685468076774, "grad_norm": 7.014920234680176, "learning_rate": 3.996832101372756e-06, "loss": 0.3904, "step": 5790 }, { "epoch": 3.0302911607259433, "grad_norm": 4.148968696594238, "learning_rate": 3.986272439281943e-06, "loss": 0.3938, "step": 5800 }, { "epoch": 3.0302911607259433, "eval_loss": 0.8061103224754333, "eval_runtime": 46.2431, "eval_samples_per_second": 36.806, "eval_steps_per_second": 4.606, "step": 5800 }, { "epoch": 3.0355137746442096, "grad_norm": 8.123299598693848, "learning_rate": 3.97571277719113e-06, "loss": 0.4201, "step": 5810 }, { "epoch": 3.0407363885624754, "grad_norm": 8.127564430236816, "learning_rate": 3.965153115100317e-06, "loss": 0.416, "step": 5820 }, { "epoch": 3.0459590024807417, "grad_norm": 3.643094539642334, "learning_rate": 3.954593453009505e-06, "loss": 0.36, "step": 5830 }, { "epoch": 3.0511816163990075, "grad_norm": 3.6215567588806152, "learning_rate": 3.944033790918691e-06, "loss": 0.354, "step": 5840 }, { "epoch": 3.056404230317274, "grad_norm": 6.820983409881592, "learning_rate": 3.933474128827878e-06, "loss": 0.4113, "step": 5850 }, { "epoch": 3.06162684423554, "grad_norm": 5.715891361236572, "learning_rate": 3.922914466737065e-06, "loss": 0.3628, "step": 5860 }, { "epoch": 3.066849458153806, "grad_norm": 7.616763114929199, "learning_rate": 3.912354804646251e-06, "loss": 0.4049, "step": 5870 }, { "epoch": 3.0720720720720722, "grad_norm": 4.177463531494141, "learning_rate": 3.9017951425554385e-06, "loss": 0.4452, "step": 5880 }, { "epoch": 3.077294685990338, "grad_norm": 4.512898921966553, "learning_rate": 3.891235480464626e-06, "loss": 0.384, "step": 5890 }, { "epoch": 3.0825172999086043, "grad_norm": 3.9176089763641357, "learning_rate": 3.880675818373813e-06, "loss": 0.4524, "step": 5900 }, { "epoch": 3.08773991382687, "grad_norm": 5.587650299072266, "learning_rate": 3.870116156282999e-06, "loss": 0.3638, "step": 5910 }, { "epoch": 3.0929625277451365, "grad_norm": 5.886160373687744, "learning_rate": 3.859556494192186e-06, "loss": 0.3664, "step": 5920 }, { "epoch": 3.0981851416634028, "grad_norm": 3.332893133163452, "learning_rate": 3.848996832101373e-06, "loss": 0.4352, "step": 5930 }, { "epoch": 3.1034077555816686, "grad_norm": 4.1508097648620605, "learning_rate": 3.83843717001056e-06, "loss": 0.485, "step": 5940 }, { "epoch": 3.108630369499935, "grad_norm": 6.934416770935059, "learning_rate": 3.8278775079197465e-06, "loss": 0.4333, "step": 5950 }, { "epoch": 3.1138529834182007, "grad_norm": 5.893505573272705, "learning_rate": 3.817317845828934e-06, "loss": 0.4209, "step": 5960 }, { "epoch": 3.119075597336467, "grad_norm": 6.001057147979736, "learning_rate": 3.8067581837381208e-06, "loss": 0.4274, "step": 5970 }, { "epoch": 3.124298211254733, "grad_norm": 4.873240947723389, "learning_rate": 3.7961985216473074e-06, "loss": 0.3278, "step": 5980 }, { "epoch": 3.129520825172999, "grad_norm": 6.454697132110596, "learning_rate": 3.7856388595564946e-06, "loss": 0.4673, "step": 5990 }, { "epoch": 3.134743439091265, "grad_norm": 5.828022003173828, "learning_rate": 3.7750791974656812e-06, "loss": 0.4288, "step": 6000 }, { "epoch": 3.134743439091265, "eval_loss": 0.8229681849479675, "eval_runtime": 46.2578, "eval_samples_per_second": 36.794, "eval_steps_per_second": 4.605, "step": 6000 }, { "epoch": 3.1399660530095312, "grad_norm": 6.981528282165527, "learning_rate": 3.7645195353748684e-06, "loss": 0.3955, "step": 6010 }, { "epoch": 3.1451886669277975, "grad_norm": 4.903995990753174, "learning_rate": 3.7539598732840555e-06, "loss": 0.3907, "step": 6020 }, { "epoch": 3.1504112808460634, "grad_norm": 4.399137496948242, "learning_rate": 3.743400211193242e-06, "loss": 0.4785, "step": 6030 }, { "epoch": 3.1556338947643296, "grad_norm": 3.4194514751434326, "learning_rate": 3.732840549102429e-06, "loss": 0.4086, "step": 6040 }, { "epoch": 3.1608565086825955, "grad_norm": 6.683743476867676, "learning_rate": 3.7222808870116164e-06, "loss": 0.4418, "step": 6050 }, { "epoch": 3.1660791226008618, "grad_norm": 4.924780368804932, "learning_rate": 3.711721224920803e-06, "loss": 0.4035, "step": 6060 }, { "epoch": 3.171301736519128, "grad_norm": 6.23117733001709, "learning_rate": 3.7011615628299897e-06, "loss": 0.3554, "step": 6070 }, { "epoch": 3.176524350437394, "grad_norm": 7.044112682342529, "learning_rate": 3.690601900739177e-06, "loss": 0.4143, "step": 6080 }, { "epoch": 3.18174696435566, "grad_norm": 6.9131059646606445, "learning_rate": 3.6800422386483635e-06, "loss": 0.4045, "step": 6090 }, { "epoch": 3.186969578273926, "grad_norm": 5.362022876739502, "learning_rate": 3.66948257655755e-06, "loss": 0.4318, "step": 6100 }, { "epoch": 3.1921921921921923, "grad_norm": 5.799452781677246, "learning_rate": 3.6589229144667377e-06, "loss": 0.4649, "step": 6110 }, { "epoch": 3.197414806110458, "grad_norm": 7.595244884490967, "learning_rate": 3.6483632523759244e-06, "loss": 0.4466, "step": 6120 }, { "epoch": 3.2026374200287244, "grad_norm": 2.4312336444854736, "learning_rate": 3.637803590285111e-06, "loss": 0.4557, "step": 6130 }, { "epoch": 3.2078600339469903, "grad_norm": 5.092735767364502, "learning_rate": 3.627243928194298e-06, "loss": 0.4255, "step": 6140 }, { "epoch": 3.2130826478652565, "grad_norm": 5.3324151039123535, "learning_rate": 3.616684266103485e-06, "loss": 0.432, "step": 6150 }, { "epoch": 3.218305261783523, "grad_norm": 4.059586524963379, "learning_rate": 3.606124604012672e-06, "loss": 0.4766, "step": 6160 }, { "epoch": 3.2235278757017887, "grad_norm": 6.850623607635498, "learning_rate": 3.5955649419218587e-06, "loss": 0.4387, "step": 6170 }, { "epoch": 3.228750489620055, "grad_norm": 5.995065212249756, "learning_rate": 3.585005279831046e-06, "loss": 0.4257, "step": 6180 }, { "epoch": 3.233973103538321, "grad_norm": 3.885401487350464, "learning_rate": 3.5744456177402325e-06, "loss": 0.4423, "step": 6190 }, { "epoch": 3.239195717456587, "grad_norm": 4.709335803985596, "learning_rate": 3.563885955649419e-06, "loss": 0.3812, "step": 6200 }, { "epoch": 3.239195717456587, "eval_loss": 0.810655415058136, "eval_runtime": 46.276, "eval_samples_per_second": 36.779, "eval_steps_per_second": 4.603, "step": 6200 }, { "epoch": 3.244418331374853, "grad_norm": 7.61454963684082, "learning_rate": 3.5533262935586067e-06, "loss": 0.4508, "step": 6210 }, { "epoch": 3.249640945293119, "grad_norm": 7.8873066902160645, "learning_rate": 3.5427666314677934e-06, "loss": 0.4208, "step": 6220 }, { "epoch": 3.2548635592113855, "grad_norm": 5.000669956207275, "learning_rate": 3.53220696937698e-06, "loss": 0.4103, "step": 6230 }, { "epoch": 3.2600861731296513, "grad_norm": 4.964175701141357, "learning_rate": 3.521647307286167e-06, "loss": 0.4562, "step": 6240 }, { "epoch": 3.2653087870479176, "grad_norm": 4.287696838378906, "learning_rate": 3.511087645195354e-06, "loss": 0.4661, "step": 6250 }, { "epoch": 3.2705314009661834, "grad_norm": 5.519683837890625, "learning_rate": 3.500527983104541e-06, "loss": 0.4882, "step": 6260 }, { "epoch": 3.2757540148844497, "grad_norm": 5.700749397277832, "learning_rate": 3.489968321013728e-06, "loss": 0.4463, "step": 6270 }, { "epoch": 3.2809766288027156, "grad_norm": 5.7745466232299805, "learning_rate": 3.4794086589229148e-06, "loss": 0.4626, "step": 6280 }, { "epoch": 3.286199242720982, "grad_norm": 8.0064058303833, "learning_rate": 3.4688489968321015e-06, "loss": 0.3982, "step": 6290 }, { "epoch": 3.291421856639248, "grad_norm": 5.860507488250732, "learning_rate": 3.458289334741289e-06, "loss": 0.3835, "step": 6300 }, { "epoch": 3.296644470557514, "grad_norm": 7.413349628448486, "learning_rate": 3.4477296726504757e-06, "loss": 0.4338, "step": 6310 }, { "epoch": 3.3018670844757803, "grad_norm": 4.818141937255859, "learning_rate": 3.4371700105596624e-06, "loss": 0.497, "step": 6320 }, { "epoch": 3.307089698394046, "grad_norm": 3.987377405166626, "learning_rate": 3.4266103484688495e-06, "loss": 0.4099, "step": 6330 }, { "epoch": 3.3123123123123124, "grad_norm": 7.0202860832214355, "learning_rate": 3.416050686378036e-06, "loss": 0.4791, "step": 6340 }, { "epoch": 3.317534926230578, "grad_norm": 6.688587665557861, "learning_rate": 3.405491024287223e-06, "loss": 0.427, "step": 6350 }, { "epoch": 3.3227575401488445, "grad_norm": 6.508810997009277, "learning_rate": 3.39493136219641e-06, "loss": 0.431, "step": 6360 }, { "epoch": 3.3279801540671103, "grad_norm": 7.402127265930176, "learning_rate": 3.384371700105597e-06, "loss": 0.4145, "step": 6370 }, { "epoch": 3.3332027679853766, "grad_norm": 4.331240177154541, "learning_rate": 3.3738120380147837e-06, "loss": 0.3799, "step": 6380 }, { "epoch": 3.338425381903643, "grad_norm": 6.24545431137085, "learning_rate": 3.3632523759239704e-06, "loss": 0.4339, "step": 6390 }, { "epoch": 3.3436479958219087, "grad_norm": 6.327270030975342, "learning_rate": 3.352692713833158e-06, "loss": 0.4506, "step": 6400 }, { "epoch": 3.3436479958219087, "eval_loss": 0.8060568571090698, "eval_runtime": 46.2458, "eval_samples_per_second": 36.803, "eval_steps_per_second": 4.606, "step": 6400 }, { "epoch": 3.348870609740175, "grad_norm": 2.5744376182556152, "learning_rate": 3.3421330517423446e-06, "loss": 0.484, "step": 6410 }, { "epoch": 3.354093223658441, "grad_norm": 3.4706344604492188, "learning_rate": 3.3315733896515313e-06, "loss": 0.3897, "step": 6420 }, { "epoch": 3.359315837576707, "grad_norm": 6.175302028656006, "learning_rate": 3.3210137275607184e-06, "loss": 0.4054, "step": 6430 }, { "epoch": 3.3645384514949734, "grad_norm": 5.064645767211914, "learning_rate": 3.310454065469905e-06, "loss": 0.4458, "step": 6440 }, { "epoch": 3.3697610654132393, "grad_norm": 8.018420219421387, "learning_rate": 3.299894403379092e-06, "loss": 0.5056, "step": 6450 }, { "epoch": 3.3749836793315056, "grad_norm": 6.1567301750183105, "learning_rate": 3.2893347412882793e-06, "loss": 0.3735, "step": 6460 }, { "epoch": 3.3802062932497714, "grad_norm": 5.155027866363525, "learning_rate": 3.278775079197466e-06, "loss": 0.3715, "step": 6470 }, { "epoch": 3.3854289071680377, "grad_norm": 5.396885395050049, "learning_rate": 3.2682154171066527e-06, "loss": 0.3597, "step": 6480 }, { "epoch": 3.3906515210863035, "grad_norm": 3.0646581649780273, "learning_rate": 3.25765575501584e-06, "loss": 0.4777, "step": 6490 }, { "epoch": 3.39587413500457, "grad_norm": 5.380611419677734, "learning_rate": 3.2470960929250265e-06, "loss": 0.4137, "step": 6500 }, { "epoch": 3.4010967489228356, "grad_norm": 6.225546360015869, "learning_rate": 3.2365364308342136e-06, "loss": 0.313, "step": 6510 }, { "epoch": 3.406319362841102, "grad_norm": 4.760247707366943, "learning_rate": 3.2259767687434007e-06, "loss": 0.3674, "step": 6520 }, { "epoch": 3.411541976759368, "grad_norm": 7.30977725982666, "learning_rate": 3.2154171066525874e-06, "loss": 0.3591, "step": 6530 }, { "epoch": 3.416764590677634, "grad_norm": 7.618262767791748, "learning_rate": 3.204857444561774e-06, "loss": 0.4384, "step": 6540 }, { "epoch": 3.4219872045959003, "grad_norm": 3.0289359092712402, "learning_rate": 3.1942977824709616e-06, "loss": 0.3818, "step": 6550 }, { "epoch": 3.427209818514166, "grad_norm": 3.5998988151550293, "learning_rate": 3.1837381203801483e-06, "loss": 0.4318, "step": 6560 }, { "epoch": 3.4324324324324325, "grad_norm": 4.631134033203125, "learning_rate": 3.173178458289335e-06, "loss": 0.3956, "step": 6570 }, { "epoch": 3.4376550463506987, "grad_norm": 8.129390716552734, "learning_rate": 3.1626187961985217e-06, "loss": 0.3928, "step": 6580 }, { "epoch": 3.4428776602689646, "grad_norm": 7.39198637008667, "learning_rate": 3.1520591341077088e-06, "loss": 0.4827, "step": 6590 }, { "epoch": 3.448100274187231, "grad_norm": 4.935920238494873, "learning_rate": 3.1414994720168955e-06, "loss": 0.4126, "step": 6600 }, { "epoch": 3.448100274187231, "eval_loss": 0.8010614514350891, "eval_runtime": 46.3186, "eval_samples_per_second": 36.746, "eval_steps_per_second": 4.599, "step": 6600 }, { "epoch": 3.4533228881054967, "grad_norm": 4.4114460945129395, "learning_rate": 3.1309398099260826e-06, "loss": 0.4439, "step": 6610 }, { "epoch": 3.458545502023763, "grad_norm": 6.824039459228516, "learning_rate": 3.1203801478352697e-06, "loss": 0.3552, "step": 6620 }, { "epoch": 3.463768115942029, "grad_norm": 4.671921253204346, "learning_rate": 3.1098204857444564e-06, "loss": 0.3289, "step": 6630 }, { "epoch": 3.468990729860295, "grad_norm": 3.557352304458618, "learning_rate": 3.099260823653643e-06, "loss": 0.4342, "step": 6640 }, { "epoch": 3.474213343778561, "grad_norm": 5.833057403564453, "learning_rate": 3.0887011615628306e-06, "loss": 0.4985, "step": 6650 }, { "epoch": 3.4794359576968272, "grad_norm": 3.8494341373443604, "learning_rate": 3.0781414994720173e-06, "loss": 0.4175, "step": 6660 }, { "epoch": 3.4846585716150935, "grad_norm": 5.017399311065674, "learning_rate": 3.067581837381204e-06, "loss": 0.3396, "step": 6670 }, { "epoch": 3.4898811855333594, "grad_norm": 8.080399513244629, "learning_rate": 3.057022175290391e-06, "loss": 0.412, "step": 6680 }, { "epoch": 3.4951037994516256, "grad_norm": 4.017096996307373, "learning_rate": 3.0464625131995777e-06, "loss": 0.4273, "step": 6690 }, { "epoch": 3.5003264133698915, "grad_norm": 5.441287517547607, "learning_rate": 3.0359028511087644e-06, "loss": 0.3789, "step": 6700 }, { "epoch": 3.5055490272881578, "grad_norm": 6.666896343231201, "learning_rate": 3.025343189017952e-06, "loss": 0.4841, "step": 6710 }, { "epoch": 3.510771641206424, "grad_norm": 6.075514316558838, "learning_rate": 3.0147835269271386e-06, "loss": 0.4189, "step": 6720 }, { "epoch": 3.51599425512469, "grad_norm": 4.0529303550720215, "learning_rate": 3.0042238648363253e-06, "loss": 0.442, "step": 6730 }, { "epoch": 3.521216869042956, "grad_norm": 4.397035598754883, "learning_rate": 2.9936642027455124e-06, "loss": 0.3511, "step": 6740 }, { "epoch": 3.526439482961222, "grad_norm": 4.687301158905029, "learning_rate": 2.9831045406546995e-06, "loss": 0.4178, "step": 6750 }, { "epoch": 3.5316620968794883, "grad_norm": 7.033337593078613, "learning_rate": 2.9725448785638862e-06, "loss": 0.4109, "step": 6760 }, { "epoch": 3.536884710797754, "grad_norm": 5.542331218719482, "learning_rate": 2.9619852164730733e-06, "loss": 0.4457, "step": 6770 }, { "epoch": 3.5421073247160204, "grad_norm": 2.987297534942627, "learning_rate": 2.95142555438226e-06, "loss": 0.4106, "step": 6780 }, { "epoch": 3.5473299386342863, "grad_norm": 4.037609100341797, "learning_rate": 2.9408658922914467e-06, "loss": 0.3515, "step": 6790 }, { "epoch": 3.5525525525525525, "grad_norm": 6.087532997131348, "learning_rate": 2.9303062302006342e-06, "loss": 0.4127, "step": 6800 }, { "epoch": 3.5525525525525525, "eval_loss": 0.799282431602478, "eval_runtime": 46.2878, "eval_samples_per_second": 36.77, "eval_steps_per_second": 4.602, "step": 6800 }, { "epoch": 3.557775166470819, "grad_norm": 5.302961826324463, "learning_rate": 2.919746568109821e-06, "loss": 0.3994, "step": 6810 }, { "epoch": 3.5629977803890847, "grad_norm": 6.701052188873291, "learning_rate": 2.9091869060190076e-06, "loss": 0.3903, "step": 6820 }, { "epoch": 3.568220394307351, "grad_norm": 5.228213310241699, "learning_rate": 2.8986272439281943e-06, "loss": 0.4006, "step": 6830 }, { "epoch": 3.573443008225617, "grad_norm": 5.282093524932861, "learning_rate": 2.8880675818373814e-06, "loss": 0.4515, "step": 6840 }, { "epoch": 3.578665622143883, "grad_norm": 3.786198616027832, "learning_rate": 2.8775079197465685e-06, "loss": 0.3696, "step": 6850 }, { "epoch": 3.5838882360621493, "grad_norm": 5.512637138366699, "learning_rate": 2.866948257655755e-06, "loss": 0.4546, "step": 6860 }, { "epoch": 3.589110849980415, "grad_norm": 7.117464542388916, "learning_rate": 2.8563885955649423e-06, "loss": 0.4562, "step": 6870 }, { "epoch": 3.594333463898681, "grad_norm": 4.943199634552002, "learning_rate": 2.845828933474129e-06, "loss": 0.4894, "step": 6880 }, { "epoch": 3.5995560778169473, "grad_norm": 8.774984359741211, "learning_rate": 2.8352692713833157e-06, "loss": 0.3925, "step": 6890 }, { "epoch": 3.6047786917352136, "grad_norm": 7.126657009124756, "learning_rate": 2.824709609292503e-06, "loss": 0.4375, "step": 6900 }, { "epoch": 3.6100013056534794, "grad_norm": 5.460080146789551, "learning_rate": 2.81414994720169e-06, "loss": 0.3847, "step": 6910 }, { "epoch": 3.6152239195717457, "grad_norm": 5.7454833984375, "learning_rate": 2.8035902851108766e-06, "loss": 0.3669, "step": 6920 }, { "epoch": 3.6204465334900116, "grad_norm": 7.132731914520264, "learning_rate": 2.7930306230200637e-06, "loss": 0.4013, "step": 6930 }, { "epoch": 3.625669147408278, "grad_norm": 4.874327659606934, "learning_rate": 2.7824709609292504e-06, "loss": 0.3866, "step": 6940 }, { "epoch": 3.630891761326544, "grad_norm": 6.133016586303711, "learning_rate": 2.771911298838437e-06, "loss": 0.3879, "step": 6950 }, { "epoch": 3.63611437524481, "grad_norm": 7.170290470123291, "learning_rate": 2.7613516367476246e-06, "loss": 0.374, "step": 6960 }, { "epoch": 3.6413369891630762, "grad_norm": 4.124912738800049, "learning_rate": 2.7507919746568113e-06, "loss": 0.5092, "step": 6970 }, { "epoch": 3.646559603081342, "grad_norm": 6.091069221496582, "learning_rate": 2.740232312565998e-06, "loss": 0.4028, "step": 6980 }, { "epoch": 3.6517822169996084, "grad_norm": 3.907172203063965, "learning_rate": 2.729672650475185e-06, "loss": 0.4105, "step": 6990 }, { "epoch": 3.6570048309178746, "grad_norm": 4.004384517669678, "learning_rate": 2.719112988384372e-06, "loss": 0.3923, "step": 7000 }, { "epoch": 3.6570048309178746, "eval_loss": 0.7925397157669067, "eval_runtime": 46.2922, "eval_samples_per_second": 36.766, "eval_steps_per_second": 4.601, "step": 7000 }, { "epoch": 3.6622274448361405, "grad_norm": 4.95105504989624, "learning_rate": 2.708553326293559e-06, "loss": 0.4407, "step": 7010 }, { "epoch": 3.6674500587544063, "grad_norm": 5.319108963012695, "learning_rate": 2.697993664202746e-06, "loss": 0.4652, "step": 7020 }, { "epoch": 3.6726726726726726, "grad_norm": 7.128709316253662, "learning_rate": 2.6874340021119326e-06, "loss": 0.4353, "step": 7030 }, { "epoch": 3.677895286590939, "grad_norm": 4.808097839355469, "learning_rate": 2.6768743400211193e-06, "loss": 0.345, "step": 7040 }, { "epoch": 3.6831179005092047, "grad_norm": 6.002725601196289, "learning_rate": 2.666314677930306e-06, "loss": 0.425, "step": 7050 }, { "epoch": 3.688340514427471, "grad_norm": 4.452878952026367, "learning_rate": 2.6557550158394935e-06, "loss": 0.4487, "step": 7060 }, { "epoch": 3.693563128345737, "grad_norm": 5.096455097198486, "learning_rate": 2.6451953537486802e-06, "loss": 0.4158, "step": 7070 }, { "epoch": 3.698785742264003, "grad_norm": 6.426013946533203, "learning_rate": 2.634635691657867e-06, "loss": 0.3701, "step": 7080 }, { "epoch": 3.7040083561822694, "grad_norm": 7.101649284362793, "learning_rate": 2.624076029567054e-06, "loss": 0.4757, "step": 7090 }, { "epoch": 3.7092309701005353, "grad_norm": 5.96986198425293, "learning_rate": 2.613516367476241e-06, "loss": 0.4475, "step": 7100 }, { "epoch": 3.7144535840188015, "grad_norm": 6.221879482269287, "learning_rate": 2.602956705385428e-06, "loss": 0.3911, "step": 7110 }, { "epoch": 3.7196761979370674, "grad_norm": 6.827433109283447, "learning_rate": 2.592397043294615e-06, "loss": 0.4751, "step": 7120 }, { "epoch": 3.7248988118553337, "grad_norm": 5.734457015991211, "learning_rate": 2.5818373812038016e-06, "loss": 0.4343, "step": 7130 }, { "epoch": 3.7301214257736, "grad_norm": 3.825587034225464, "learning_rate": 2.5712777191129883e-06, "loss": 0.3712, "step": 7140 }, { "epoch": 3.735344039691866, "grad_norm": 7.958340644836426, "learning_rate": 2.560718057022176e-06, "loss": 0.4076, "step": 7150 }, { "epoch": 3.7405666536101316, "grad_norm": 6.706486701965332, "learning_rate": 2.5501583949313625e-06, "loss": 0.4379, "step": 7160 }, { "epoch": 3.745789267528398, "grad_norm": 3.6787171363830566, "learning_rate": 2.539598732840549e-06, "loss": 0.3937, "step": 7170 }, { "epoch": 3.751011881446664, "grad_norm": 3.7710745334625244, "learning_rate": 2.5290390707497363e-06, "loss": 0.4023, "step": 7180 }, { "epoch": 3.75623449536493, "grad_norm": 9.986141204833984, "learning_rate": 2.518479408658923e-06, "loss": 0.4278, "step": 7190 }, { "epoch": 3.7614571092831963, "grad_norm": 5.851425647735596, "learning_rate": 2.50791974656811e-06, "loss": 0.3561, "step": 7200 }, { "epoch": 3.7614571092831963, "eval_loss": 0.7960723042488098, "eval_runtime": 46.2236, "eval_samples_per_second": 36.821, "eval_steps_per_second": 4.608, "step": 7200 }, { "epoch": 3.766679723201462, "grad_norm": 5.080770969390869, "learning_rate": 2.4973600844772968e-06, "loss": 0.4494, "step": 7210 }, { "epoch": 3.7719023371197284, "grad_norm": 6.7447190284729, "learning_rate": 2.486800422386484e-06, "loss": 0.4839, "step": 7220 }, { "epoch": 3.7771249510379947, "grad_norm": 5.529577255249023, "learning_rate": 2.4762407602956706e-06, "loss": 0.3799, "step": 7230 }, { "epoch": 3.7823475649562606, "grad_norm": 3.1499006748199463, "learning_rate": 2.4656810982048577e-06, "loss": 0.3959, "step": 7240 }, { "epoch": 3.787570178874527, "grad_norm": 7.214032173156738, "learning_rate": 2.4551214361140448e-06, "loss": 0.4165, "step": 7250 }, { "epoch": 3.7927927927927927, "grad_norm": 3.6615257263183594, "learning_rate": 2.4445617740232315e-06, "loss": 0.3899, "step": 7260 }, { "epoch": 3.798015406711059, "grad_norm": 8.117351531982422, "learning_rate": 2.4340021119324186e-06, "loss": 0.3904, "step": 7270 }, { "epoch": 3.803238020629325, "grad_norm": 6.586986064910889, "learning_rate": 2.4234424498416053e-06, "loss": 0.4566, "step": 7280 }, { "epoch": 3.808460634547591, "grad_norm": 2.746188163757324, "learning_rate": 2.412882787750792e-06, "loss": 0.3717, "step": 7290 }, { "epoch": 3.813683248465857, "grad_norm": 3.323824167251587, "learning_rate": 2.402323125659979e-06, "loss": 0.4007, "step": 7300 }, { "epoch": 3.818905862384123, "grad_norm": 4.206129550933838, "learning_rate": 2.391763463569166e-06, "loss": 0.3699, "step": 7310 }, { "epoch": 3.8241284763023895, "grad_norm": 4.980790615081787, "learning_rate": 2.381203801478353e-06, "loss": 0.4213, "step": 7320 }, { "epoch": 3.8293510902206553, "grad_norm": 4.453920841217041, "learning_rate": 2.37064413938754e-06, "loss": 0.3974, "step": 7330 }, { "epoch": 3.8345737041389216, "grad_norm": 4.445418834686279, "learning_rate": 2.3600844772967266e-06, "loss": 0.3761, "step": 7340 }, { "epoch": 3.8397963180571875, "grad_norm": 5.032138347625732, "learning_rate": 2.3495248152059137e-06, "loss": 0.3759, "step": 7350 }, { "epoch": 3.8450189319754537, "grad_norm": 6.470358371734619, "learning_rate": 2.3389651531151004e-06, "loss": 0.4194, "step": 7360 }, { "epoch": 3.85024154589372, "grad_norm": 3.676422119140625, "learning_rate": 2.328405491024287e-06, "loss": 0.4073, "step": 7370 }, { "epoch": 3.855464159811986, "grad_norm": 2.7682676315307617, "learning_rate": 2.3178458289334742e-06, "loss": 0.4572, "step": 7380 }, { "epoch": 3.860686773730252, "grad_norm": 2.9539079666137695, "learning_rate": 2.3072861668426613e-06, "loss": 0.4329, "step": 7390 }, { "epoch": 3.865909387648518, "grad_norm": 3.254023551940918, "learning_rate": 2.296726504751848e-06, "loss": 0.3963, "step": 7400 }, { "epoch": 3.865909387648518, "eval_loss": 0.7841590046882629, "eval_runtime": 46.2618, "eval_samples_per_second": 36.791, "eval_steps_per_second": 4.604, "step": 7400 }, { "epoch": 3.8711320015667843, "grad_norm": 3.7601821422576904, "learning_rate": 2.286166842661035e-06, "loss": 0.4022, "step": 7410 }, { "epoch": 3.87635461548505, "grad_norm": 6.364165782928467, "learning_rate": 2.2756071805702222e-06, "loss": 0.5135, "step": 7420 }, { "epoch": 3.8815772294033164, "grad_norm": 4.010589599609375, "learning_rate": 2.265047518479409e-06, "loss": 0.433, "step": 7430 }, { "epoch": 3.8867998433215822, "grad_norm": 9.321678161621094, "learning_rate": 2.2544878563885956e-06, "loss": 0.4228, "step": 7440 }, { "epoch": 3.8920224572398485, "grad_norm": 7.4074273109436035, "learning_rate": 2.2439281942977827e-06, "loss": 0.4568, "step": 7450 }, { "epoch": 3.897245071158115, "grad_norm": 6.130796432495117, "learning_rate": 2.2333685322069694e-06, "loss": 0.3895, "step": 7460 }, { "epoch": 3.9024676850763806, "grad_norm": 6.908585071563721, "learning_rate": 2.2228088701161565e-06, "loss": 0.35, "step": 7470 }, { "epoch": 3.907690298994647, "grad_norm": 2.8046581745147705, "learning_rate": 2.212249208025343e-06, "loss": 0.4573, "step": 7480 }, { "epoch": 3.9129129129129128, "grad_norm": 4.019318580627441, "learning_rate": 2.2016895459345303e-06, "loss": 0.4703, "step": 7490 }, { "epoch": 3.918135526831179, "grad_norm": 5.776391983032227, "learning_rate": 2.1911298838437174e-06, "loss": 0.3865, "step": 7500 }, { "epoch": 3.9233581407494453, "grad_norm": 5.176472187042236, "learning_rate": 2.180570221752904e-06, "loss": 0.3992, "step": 7510 }, { "epoch": 3.928580754667711, "grad_norm": 5.863769054412842, "learning_rate": 2.170010559662091e-06, "loss": 0.4149, "step": 7520 }, { "epoch": 3.933803368585977, "grad_norm": 4.269286632537842, "learning_rate": 2.159450897571278e-06, "loss": 0.3281, "step": 7530 }, { "epoch": 3.9390259825042433, "grad_norm": 5.141351699829102, "learning_rate": 2.1488912354804646e-06, "loss": 0.3818, "step": 7540 }, { "epoch": 3.9442485964225096, "grad_norm": 7.267117977142334, "learning_rate": 2.1383315733896517e-06, "loss": 0.4357, "step": 7550 }, { "epoch": 3.9494712103407754, "grad_norm": 2.134504556655884, "learning_rate": 2.1277719112988384e-06, "loss": 0.408, "step": 7560 }, { "epoch": 3.9546938242590417, "grad_norm": 2.806506872177124, "learning_rate": 2.1172122492080255e-06, "loss": 0.4074, "step": 7570 }, { "epoch": 3.9599164381773075, "grad_norm": 7.443352699279785, "learning_rate": 2.1066525871172126e-06, "loss": 0.4055, "step": 7580 }, { "epoch": 3.965139052095574, "grad_norm": 4.180816650390625, "learning_rate": 2.0960929250263993e-06, "loss": 0.4754, "step": 7590 }, { "epoch": 3.97036166601384, "grad_norm": 7.595555305480957, "learning_rate": 2.0855332629355864e-06, "loss": 0.4441, "step": 7600 }, { "epoch": 3.97036166601384, "eval_loss": 0.7774137258529663, "eval_runtime": 46.2875, "eval_samples_per_second": 36.77, "eval_steps_per_second": 4.602, "step": 7600 } ], "logging_steps": 10, "max_steps": 9570, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 200, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 8.228569170208358e+16, "train_batch_size": 2, "trial_name": null, "trial_params": null }