diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,6292 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.0, + "eval_steps": 500, + "global_step": 894, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0011185682326621924, + "grad_norm": 4.768556118011475, + "learning_rate": 4.99998456401763e-05, + "loss": 4.817, + "step": 1 + }, + { + "epoch": 0.0022371364653243847, + "grad_norm": 11.73538875579834, + "learning_rate": 4.9999382562611344e-05, + "loss": 4.6414, + "step": 2 + }, + { + "epoch": 0.003355704697986577, + "grad_norm": 4.775832176208496, + "learning_rate": 4.999861077302358e-05, + "loss": 4.6803, + "step": 3 + }, + { + "epoch": 0.0044742729306487695, + "grad_norm": 7.621553421020508, + "learning_rate": 4.9997530280943684e-05, + "loss": 4.8222, + "step": 4 + }, + { + "epoch": 0.005592841163310962, + "grad_norm": 4.997500896453857, + "learning_rate": 4.9996141099714405e-05, + "loss": 4.581, + "step": 5 + }, + { + "epoch": 0.006711409395973154, + "grad_norm": 4.785826683044434, + "learning_rate": 4.999444324649045e-05, + "loss": 4.4922, + "step": 6 + }, + { + "epoch": 0.007829977628635347, + "grad_norm": 6.647776126861572, + "learning_rate": 4.999243674223826e-05, + "loss": 4.8455, + "step": 7 + }, + { + "epoch": 0.008948545861297539, + "grad_norm": 6.596885681152344, + "learning_rate": 4.9990121611735704e-05, + "loss": 4.5977, + "step": 8 + }, + { + "epoch": 0.010067114093959731, + "grad_norm": 3.9681015014648438, + "learning_rate": 4.998749788357184e-05, + "loss": 4.5543, + "step": 9 + }, + { + "epoch": 0.011185682326621925, + "grad_norm": 3.7588727474212646, + "learning_rate": 4.998456559014653e-05, + "loss": 4.5815, + "step": 10 + }, + { + "epoch": 0.012304250559284116, + "grad_norm": 4.819014072418213, + "learning_rate": 4.9981324767670034e-05, + "loss": 4.5079, + "step": 11 + }, + { + "epoch": 0.013422818791946308, + "grad_norm": 3.4477319717407227, + "learning_rate": 4.997777545616258e-05, + "loss": 4.299, + "step": 12 + }, + { + "epoch": 0.0145413870246085, + "grad_norm": 4.113775730133057, + "learning_rate": 4.997391769945385e-05, + "loss": 4.3105, + "step": 13 + }, + { + "epoch": 0.015659955257270694, + "grad_norm": 3.848585844039917, + "learning_rate": 4.996975154518245e-05, + "loss": 4.4019, + "step": 14 + }, + { + "epoch": 0.016778523489932886, + "grad_norm": 3.787041425704956, + "learning_rate": 4.996527704479535e-05, + "loss": 4.4983, + "step": 15 + }, + { + "epoch": 0.017897091722595078, + "grad_norm": 4.073084354400635, + "learning_rate": 4.996049425354718e-05, + "loss": 4.4303, + "step": 16 + }, + { + "epoch": 0.01901565995525727, + "grad_norm": 3.8887457847595215, + "learning_rate": 4.99554032304996e-05, + "loss": 4.3752, + "step": 17 + }, + { + "epoch": 0.020134228187919462, + "grad_norm": 4.539124011993408, + "learning_rate": 4.995000403852057e-05, + "loss": 4.4398, + "step": 18 + }, + { + "epoch": 0.021252796420581657, + "grad_norm": 4.554653167724609, + "learning_rate": 4.994429674428356e-05, + "loss": 4.5068, + "step": 19 + }, + { + "epoch": 0.02237136465324385, + "grad_norm": 4.2434892654418945, + "learning_rate": 4.9938281418266717e-05, + "loss": 4.2128, + "step": 20 + }, + { + "epoch": 0.02348993288590604, + "grad_norm": 3.5199990272521973, + "learning_rate": 4.993195813475202e-05, + "loss": 4.3802, + "step": 21 + }, + { + "epoch": 0.024608501118568233, + "grad_norm": 4.401162624359131, + "learning_rate": 4.9925326971824345e-05, + "loss": 4.3791, + "step": 22 + }, + { + "epoch": 0.025727069351230425, + "grad_norm": 3.330929756164551, + "learning_rate": 4.9918388011370496e-05, + "loss": 4.3586, + "step": 23 + }, + { + "epoch": 0.026845637583892617, + "grad_norm": 3.9935824871063232, + "learning_rate": 4.9911141339078215e-05, + "loss": 4.4506, + "step": 24 + }, + { + "epoch": 0.02796420581655481, + "grad_norm": 5.406892776489258, + "learning_rate": 4.990358704443511e-05, + "loss": 4.1189, + "step": 25 + }, + { + "epoch": 0.029082774049217, + "grad_norm": 3.9377341270446777, + "learning_rate": 4.989572522072753e-05, + "loss": 4.4841, + "step": 26 + }, + { + "epoch": 0.030201342281879196, + "grad_norm": 3.408912181854248, + "learning_rate": 4.988755596503948e-05, + "loss": 4.4523, + "step": 27 + }, + { + "epoch": 0.03131991051454139, + "grad_norm": 3.9062108993530273, + "learning_rate": 4.987907937825133e-05, + "loss": 4.2521, + "step": 28 + }, + { + "epoch": 0.03243847874720358, + "grad_norm": 3.1871280670166016, + "learning_rate": 4.987029556503864e-05, + "loss": 3.9971, + "step": 29 + }, + { + "epoch": 0.03355704697986577, + "grad_norm": 3.458665609359741, + "learning_rate": 4.986120463387084e-05, + "loss": 4.3549, + "step": 30 + }, + { + "epoch": 0.03467561521252797, + "grad_norm": 3.2417380809783936, + "learning_rate": 4.985180669700989e-05, + "loss": 4.452, + "step": 31 + }, + { + "epoch": 0.035794183445190156, + "grad_norm": 2.737922430038452, + "learning_rate": 4.9842101870508904e-05, + "loss": 4.5525, + "step": 32 + }, + { + "epoch": 0.03691275167785235, + "grad_norm": 3.5653915405273438, + "learning_rate": 4.9832090274210714e-05, + "loss": 4.3004, + "step": 33 + }, + { + "epoch": 0.03803131991051454, + "grad_norm": 3.541865348815918, + "learning_rate": 4.982177203174636e-05, + "loss": 4.1722, + "step": 34 + }, + { + "epoch": 0.039149888143176735, + "grad_norm": 3.923772096633911, + "learning_rate": 4.981114727053362e-05, + "loss": 4.4216, + "step": 35 + }, + { + "epoch": 0.040268456375838924, + "grad_norm": 4.278195381164551, + "learning_rate": 4.98002161217754e-05, + "loss": 4.2207, + "step": 36 + }, + { + "epoch": 0.04138702460850112, + "grad_norm": 2.4082083702087402, + "learning_rate": 4.9788978720458104e-05, + "loss": 4.2682, + "step": 37 + }, + { + "epoch": 0.042505592841163314, + "grad_norm": 3.5256547927856445, + "learning_rate": 4.977743520535001e-05, + "loss": 4.2313, + "step": 38 + }, + { + "epoch": 0.0436241610738255, + "grad_norm": 3.8092055320739746, + "learning_rate": 4.9765585718999495e-05, + "loss": 4.3788, + "step": 39 + }, + { + "epoch": 0.0447427293064877, + "grad_norm": 3.763134241104126, + "learning_rate": 4.975343040773335e-05, + "loss": 4.1468, + "step": 40 + }, + { + "epoch": 0.04586129753914989, + "grad_norm": 4.660019874572754, + "learning_rate": 4.974096942165489e-05, + "loss": 4.1907, + "step": 41 + }, + { + "epoch": 0.04697986577181208, + "grad_norm": 4.472564220428467, + "learning_rate": 4.9728202914642183e-05, + "loss": 4.5382, + "step": 42 + }, + { + "epoch": 0.04809843400447427, + "grad_norm": 4.786443710327148, + "learning_rate": 4.9715131044346084e-05, + "loss": 4.2274, + "step": 43 + }, + { + "epoch": 0.049217002237136466, + "grad_norm": 4.188578128814697, + "learning_rate": 4.970175397218832e-05, + "loss": 4.0529, + "step": 44 + }, + { + "epoch": 0.050335570469798654, + "grad_norm": 3.1671926975250244, + "learning_rate": 4.9688071863359484e-05, + "loss": 4.1644, + "step": 45 + }, + { + "epoch": 0.05145413870246085, + "grad_norm": 3.790748357772827, + "learning_rate": 4.9674084886817016e-05, + "loss": 3.9051, + "step": 46 + }, + { + "epoch": 0.052572706935123045, + "grad_norm": 3.404773473739624, + "learning_rate": 4.965979321528309e-05, + "loss": 4.5011, + "step": 47 + }, + { + "epoch": 0.053691275167785234, + "grad_norm": 3.7244088649749756, + "learning_rate": 4.9645197025242506e-05, + "loss": 4.3524, + "step": 48 + }, + { + "epoch": 0.05480984340044743, + "grad_norm": 2.824608564376831, + "learning_rate": 4.963029649694049e-05, + "loss": 4.3211, + "step": 49 + }, + { + "epoch": 0.05592841163310962, + "grad_norm": 3.457947254180908, + "learning_rate": 4.9615091814380465e-05, + "loss": 4.3325, + "step": 50 + }, + { + "epoch": 0.05704697986577181, + "grad_norm": 2.658190965652466, + "learning_rate": 4.959958316532181e-05, + "loss": 4.2936, + "step": 51 + }, + { + "epoch": 0.058165548098434, + "grad_norm": 3.7741482257843018, + "learning_rate": 4.9583770741277505e-05, + "loss": 4.184, + "step": 52 + }, + { + "epoch": 0.0592841163310962, + "grad_norm": 2.5784506797790527, + "learning_rate": 4.9567654737511794e-05, + "loss": 4.2936, + "step": 53 + }, + { + "epoch": 0.06040268456375839, + "grad_norm": 3.018603563308716, + "learning_rate": 4.955123535303776e-05, + "loss": 4.1786, + "step": 54 + }, + { + "epoch": 0.06152125279642058, + "grad_norm": 3.7826101779937744, + "learning_rate": 4.953451279061485e-05, + "loss": 4.207, + "step": 55 + }, + { + "epoch": 0.06263982102908278, + "grad_norm": 4.488926887512207, + "learning_rate": 4.951748725674643e-05, + "loss": 4.6036, + "step": 56 + }, + { + "epoch": 0.06375838926174497, + "grad_norm": 4.047707557678223, + "learning_rate": 4.950015896167716e-05, + "loss": 3.9239, + "step": 57 + }, + { + "epoch": 0.06487695749440715, + "grad_norm": 2.5932610034942627, + "learning_rate": 4.9482528119390435e-05, + "loss": 4.1098, + "step": 58 + }, + { + "epoch": 0.06599552572706935, + "grad_norm": 2.828861951828003, + "learning_rate": 4.946459494760578e-05, + "loss": 4.2837, + "step": 59 + }, + { + "epoch": 0.06711409395973154, + "grad_norm": 3.5818116664886475, + "learning_rate": 4.9446359667776065e-05, + "loss": 4.3673, + "step": 60 + }, + { + "epoch": 0.06823266219239374, + "grad_norm": 2.7391154766082764, + "learning_rate": 4.9427822505084874e-05, + "loss": 4.5594, + "step": 61 + }, + { + "epoch": 0.06935123042505593, + "grad_norm": 2.6267666816711426, + "learning_rate": 4.9408983688443654e-05, + "loss": 4.3283, + "step": 62 + }, + { + "epoch": 0.07046979865771812, + "grad_norm": 2.574420690536499, + "learning_rate": 4.938984345048892e-05, + "loss": 4.2056, + "step": 63 + }, + { + "epoch": 0.07158836689038031, + "grad_norm": 3.568028450012207, + "learning_rate": 4.937040202757937e-05, + "loss": 4.2298, + "step": 64 + }, + { + "epoch": 0.07270693512304251, + "grad_norm": 3.778529405593872, + "learning_rate": 4.9350659659792976e-05, + "loss": 4.1133, + "step": 65 + }, + { + "epoch": 0.0738255033557047, + "grad_norm": 2.817169427871704, + "learning_rate": 4.933061659092401e-05, + "loss": 4.183, + "step": 66 + }, + { + "epoch": 0.07494407158836688, + "grad_norm": 3.1611924171447754, + "learning_rate": 4.931027306848004e-05, + "loss": 4.1715, + "step": 67 + }, + { + "epoch": 0.07606263982102908, + "grad_norm": 3.813957929611206, + "learning_rate": 4.9289629343678864e-05, + "loss": 4.2382, + "step": 68 + }, + { + "epoch": 0.07718120805369127, + "grad_norm": 2.7094550132751465, + "learning_rate": 4.926868567144543e-05, + "loss": 4.107, + "step": 69 + }, + { + "epoch": 0.07829977628635347, + "grad_norm": 2.958317995071411, + "learning_rate": 4.924744231040864e-05, + "loss": 4.2052, + "step": 70 + }, + { + "epoch": 0.07941834451901567, + "grad_norm": 2.8046460151672363, + "learning_rate": 4.9225899522898236e-05, + "loss": 4.1191, + "step": 71 + }, + { + "epoch": 0.08053691275167785, + "grad_norm": 2.8792548179626465, + "learning_rate": 4.920405757494147e-05, + "loss": 3.9608, + "step": 72 + }, + { + "epoch": 0.08165548098434004, + "grad_norm": 2.953927516937256, + "learning_rate": 4.91819167362599e-05, + "loss": 4.6505, + "step": 73 + }, + { + "epoch": 0.08277404921700224, + "grad_norm": 2.8335180282592773, + "learning_rate": 4.915947728026598e-05, + "loss": 4.5291, + "step": 74 + }, + { + "epoch": 0.08389261744966443, + "grad_norm": 2.0205912590026855, + "learning_rate": 4.9136739484059766e-05, + "loss": 4.1919, + "step": 75 + }, + { + "epoch": 0.08501118568232663, + "grad_norm": 4.005690097808838, + "learning_rate": 4.911370362842543e-05, + "loss": 4.2847, + "step": 76 + }, + { + "epoch": 0.08612975391498881, + "grad_norm": 2.3989546298980713, + "learning_rate": 4.9090369997827826e-05, + "loss": 4.2773, + "step": 77 + }, + { + "epoch": 0.087248322147651, + "grad_norm": 2.6297130584716797, + "learning_rate": 4.9066738880408945e-05, + "loss": 4.1242, + "step": 78 + }, + { + "epoch": 0.0883668903803132, + "grad_norm": 3.310232639312744, + "learning_rate": 4.904281056798441e-05, + "loss": 4.4331, + "step": 79 + }, + { + "epoch": 0.0894854586129754, + "grad_norm": 1.7795904874801636, + "learning_rate": 4.901858535603983e-05, + "loss": 4.1656, + "step": 80 + }, + { + "epoch": 0.09060402684563758, + "grad_norm": 2.2798268795013428, + "learning_rate": 4.899406354372715e-05, + "loss": 4.1374, + "step": 81 + }, + { + "epoch": 0.09172259507829977, + "grad_norm": 2.3790340423583984, + "learning_rate": 4.896924543386099e-05, + "loss": 4.1786, + "step": 82 + }, + { + "epoch": 0.09284116331096197, + "grad_norm": 2.345724105834961, + "learning_rate": 4.894413133291488e-05, + "loss": 3.9511, + "step": 83 + }, + { + "epoch": 0.09395973154362416, + "grad_norm": 3.110872507095337, + "learning_rate": 4.891872155101746e-05, + "loss": 4.2832, + "step": 84 + }, + { + "epoch": 0.09507829977628636, + "grad_norm": 2.6279842853546143, + "learning_rate": 4.889301640194869e-05, + "loss": 4.4677, + "step": 85 + }, + { + "epoch": 0.09619686800894854, + "grad_norm": 2.8604133129119873, + "learning_rate": 4.886701620313595e-05, + "loss": 4.3158, + "step": 86 + }, + { + "epoch": 0.09731543624161074, + "grad_norm": 4.6412858963012695, + "learning_rate": 4.884072127565014e-05, + "loss": 4.4612, + "step": 87 + }, + { + "epoch": 0.09843400447427293, + "grad_norm": 2.649590492248535, + "learning_rate": 4.881413194420169e-05, + "loss": 3.9387, + "step": 88 + }, + { + "epoch": 0.09955257270693513, + "grad_norm": 3.56644344329834, + "learning_rate": 4.878724853713655e-05, + "loss": 4.2099, + "step": 89 + }, + { + "epoch": 0.10067114093959731, + "grad_norm": 2.599231243133545, + "learning_rate": 4.876007138643216e-05, + "loss": 4.3014, + "step": 90 + }, + { + "epoch": 0.1017897091722595, + "grad_norm": 2.8511576652526855, + "learning_rate": 4.8732600827693344e-05, + "loss": 3.8494, + "step": 91 + }, + { + "epoch": 0.1029082774049217, + "grad_norm": 2.3662049770355225, + "learning_rate": 4.870483720014814e-05, + "loss": 3.9008, + "step": 92 + }, + { + "epoch": 0.1040268456375839, + "grad_norm": 3.010821580886841, + "learning_rate": 4.8676780846643644e-05, + "loss": 4.1022, + "step": 93 + }, + { + "epoch": 0.10514541387024609, + "grad_norm": 3.158817768096924, + "learning_rate": 4.864843211364176e-05, + "loss": 4.2633, + "step": 94 + }, + { + "epoch": 0.10626398210290827, + "grad_norm": 2.9291787147521973, + "learning_rate": 4.861979135121493e-05, + "loss": 4.0874, + "step": 95 + }, + { + "epoch": 0.10738255033557047, + "grad_norm": 3.55708909034729, + "learning_rate": 4.859085891304178e-05, + "loss": 4.1598, + "step": 96 + }, + { + "epoch": 0.10850111856823266, + "grad_norm": 2.482081651687622, + "learning_rate": 4.85616351564028e-05, + "loss": 4.4488, + "step": 97 + }, + { + "epoch": 0.10961968680089486, + "grad_norm": 2.2329485416412354, + "learning_rate": 4.853212044217591e-05, + "loss": 4.1849, + "step": 98 + }, + { + "epoch": 0.11073825503355705, + "grad_norm": 2.4485504627227783, + "learning_rate": 4.8502315134832e-05, + "loss": 4.2842, + "step": 99 + }, + { + "epoch": 0.11185682326621924, + "grad_norm": 2.239509344100952, + "learning_rate": 4.847221960243041e-05, + "loss": 4.4354, + "step": 100 + }, + { + "epoch": 0.11297539149888143, + "grad_norm": 2.5421688556671143, + "learning_rate": 4.8441834216614454e-05, + "loss": 4.4723, + "step": 101 + }, + { + "epoch": 0.11409395973154363, + "grad_norm": 3.123340129852295, + "learning_rate": 4.8411159352606734e-05, + "loss": 3.812, + "step": 102 + }, + { + "epoch": 0.11521252796420582, + "grad_norm": 1.9788379669189453, + "learning_rate": 4.838019538920458e-05, + "loss": 4.2419, + "step": 103 + }, + { + "epoch": 0.116331096196868, + "grad_norm": 2.201848268508911, + "learning_rate": 4.834894270877536e-05, + "loss": 4.1965, + "step": 104 + }, + { + "epoch": 0.1174496644295302, + "grad_norm": 2.6100194454193115, + "learning_rate": 4.831740169725172e-05, + "loss": 4.3533, + "step": 105 + }, + { + "epoch": 0.1185682326621924, + "grad_norm": 2.866046905517578, + "learning_rate": 4.8285572744126854e-05, + "loss": 3.9282, + "step": 106 + }, + { + "epoch": 0.11968680089485459, + "grad_norm": 2.9005253314971924, + "learning_rate": 4.8253456242449704e-05, + "loss": 4.2894, + "step": 107 + }, + { + "epoch": 0.12080536912751678, + "grad_norm": 2.173900842666626, + "learning_rate": 4.822105258882007e-05, + "loss": 4.3334, + "step": 108 + }, + { + "epoch": 0.12192393736017897, + "grad_norm": 2.949611186981201, + "learning_rate": 4.818836218338373e-05, + "loss": 4.2813, + "step": 109 + }, + { + "epoch": 0.12304250559284116, + "grad_norm": 3.2048230171203613, + "learning_rate": 4.81553854298275e-05, + "loss": 4.3333, + "step": 110 + }, + { + "epoch": 0.12416107382550336, + "grad_norm": 2.2151906490325928, + "learning_rate": 4.812212273537425e-05, + "loss": 4.0185, + "step": 111 + }, + { + "epoch": 0.12527964205816555, + "grad_norm": 3.4764933586120605, + "learning_rate": 4.808857451077788e-05, + "loss": 4.0535, + "step": 112 + }, + { + "epoch": 0.12639821029082773, + "grad_norm": 2.3762760162353516, + "learning_rate": 4.805474117031822e-05, + "loss": 4.3605, + "step": 113 + }, + { + "epoch": 0.12751677852348994, + "grad_norm": 2.3675270080566406, + "learning_rate": 4.802062313179595e-05, + "loss": 4.1901, + "step": 114 + }, + { + "epoch": 0.12863534675615212, + "grad_norm": 2.8210885524749756, + "learning_rate": 4.798622081652743e-05, + "loss": 4.5484, + "step": 115 + }, + { + "epoch": 0.1297539149888143, + "grad_norm": 2.597954511642456, + "learning_rate": 4.795153464933948e-05, + "loss": 4.3055, + "step": 116 + }, + { + "epoch": 0.13087248322147652, + "grad_norm": 2.793868064880371, + "learning_rate": 4.7916565058564155e-05, + "loss": 4.4277, + "step": 117 + }, + { + "epoch": 0.1319910514541387, + "grad_norm": 3.061800718307495, + "learning_rate": 4.788131247603345e-05, + "loss": 4.2027, + "step": 118 + }, + { + "epoch": 0.1331096196868009, + "grad_norm": 3.1223654747009277, + "learning_rate": 4.784577733707394e-05, + "loss": 4.3513, + "step": 119 + }, + { + "epoch": 0.1342281879194631, + "grad_norm": 2.714157819747925, + "learning_rate": 4.7809960080501464e-05, + "loss": 4.2208, + "step": 120 + }, + { + "epoch": 0.13534675615212527, + "grad_norm": 2.218320369720459, + "learning_rate": 4.777386114861565e-05, + "loss": 4.2255, + "step": 121 + }, + { + "epoch": 0.13646532438478748, + "grad_norm": 2.74094557762146, + "learning_rate": 4.7737480987194484e-05, + "loss": 4.4531, + "step": 122 + }, + { + "epoch": 0.13758389261744966, + "grad_norm": 2.0145950317382812, + "learning_rate": 4.7700820045488783e-05, + "loss": 4.2616, + "step": 123 + }, + { + "epoch": 0.13870246085011187, + "grad_norm": 3.6711690425872803, + "learning_rate": 4.766387877621667e-05, + "loss": 4.3626, + "step": 124 + }, + { + "epoch": 0.13982102908277405, + "grad_norm": 2.7168869972229004, + "learning_rate": 4.762665763555797e-05, + "loss": 4.1299, + "step": 125 + }, + { + "epoch": 0.14093959731543623, + "grad_norm": 2.299030065536499, + "learning_rate": 4.758915708314858e-05, + "loss": 4.378, + "step": 126 + }, + { + "epoch": 0.14205816554809844, + "grad_norm": 2.2006888389587402, + "learning_rate": 4.755137758207479e-05, + "loss": 4.5967, + "step": 127 + }, + { + "epoch": 0.14317673378076062, + "grad_norm": 2.05462908744812, + "learning_rate": 4.751331959886758e-05, + "loss": 4.235, + "step": 128 + }, + { + "epoch": 0.14429530201342283, + "grad_norm": 1.8818479776382446, + "learning_rate": 4.7474983603496815e-05, + "loss": 4.1797, + "step": 129 + }, + { + "epoch": 0.14541387024608501, + "grad_norm": 2.008810043334961, + "learning_rate": 4.7436370069365524e-05, + "loss": 4.26, + "step": 130 + }, + { + "epoch": 0.1465324384787472, + "grad_norm": 2.405534029006958, + "learning_rate": 4.739747947330394e-05, + "loss": 3.8538, + "step": 131 + }, + { + "epoch": 0.1476510067114094, + "grad_norm": 2.6731414794921875, + "learning_rate": 4.7358312295563734e-05, + "loss": 4.1691, + "step": 132 + }, + { + "epoch": 0.1487695749440716, + "grad_norm": 3.544353723526001, + "learning_rate": 4.7318869019811986e-05, + "loss": 4.0219, + "step": 133 + }, + { + "epoch": 0.14988814317673377, + "grad_norm": 2.278864622116089, + "learning_rate": 4.727915013312526e-05, + "loss": 4.3926, + "step": 134 + }, + { + "epoch": 0.15100671140939598, + "grad_norm": 2.240647554397583, + "learning_rate": 4.7239156125983594e-05, + "loss": 4.3835, + "step": 135 + }, + { + "epoch": 0.15212527964205816, + "grad_norm": 3.1606392860412598, + "learning_rate": 4.7198887492264416e-05, + "loss": 3.9971, + "step": 136 + }, + { + "epoch": 0.15324384787472037, + "grad_norm": 2.4671874046325684, + "learning_rate": 4.7158344729236454e-05, + "loss": 4.4052, + "step": 137 + }, + { + "epoch": 0.15436241610738255, + "grad_norm": 1.9113795757293701, + "learning_rate": 4.711752833755362e-05, + "loss": 4.1998, + "step": 138 + }, + { + "epoch": 0.15548098434004473, + "grad_norm": 2.9461162090301514, + "learning_rate": 4.707643882124878e-05, + "loss": 4.212, + "step": 139 + }, + { + "epoch": 0.15659955257270694, + "grad_norm": 3.278677463531494, + "learning_rate": 4.70350766877276e-05, + "loss": 4.4147, + "step": 140 + }, + { + "epoch": 0.15771812080536912, + "grad_norm": 3.2217676639556885, + "learning_rate": 4.699344244776218e-05, + "loss": 4.2919, + "step": 141 + }, + { + "epoch": 0.15883668903803133, + "grad_norm": 1.8135348558425903, + "learning_rate": 4.6951536615484854e-05, + "loss": 4.0982, + "step": 142 + }, + { + "epoch": 0.1599552572706935, + "grad_norm": 3.3648054599761963, + "learning_rate": 4.6909359708381775e-05, + "loss": 4.3256, + "step": 143 + }, + { + "epoch": 0.1610738255033557, + "grad_norm": 2.8032402992248535, + "learning_rate": 4.686691224728652e-05, + "loss": 4.3029, + "step": 144 + }, + { + "epoch": 0.1621923937360179, + "grad_norm": 2.585247039794922, + "learning_rate": 4.682419475637372e-05, + "loss": 4.1191, + "step": 145 + }, + { + "epoch": 0.16331096196868009, + "grad_norm": 2.0398333072662354, + "learning_rate": 4.678120776315251e-05, + "loss": 4.2553, + "step": 146 + }, + { + "epoch": 0.1644295302013423, + "grad_norm": 2.485276460647583, + "learning_rate": 4.673795179846007e-05, + "loss": 4.2602, + "step": 147 + }, + { + "epoch": 0.16554809843400448, + "grad_norm": 2.1474947929382324, + "learning_rate": 4.669442739645506e-05, + "loss": 4.2519, + "step": 148 + }, + { + "epoch": 0.16666666666666666, + "grad_norm": 2.4320313930511475, + "learning_rate": 4.665063509461097e-05, + "loss": 4.2698, + "step": 149 + }, + { + "epoch": 0.16778523489932887, + "grad_norm": 3.272096872329712, + "learning_rate": 4.660657543370958e-05, + "loss": 4.2166, + "step": 150 + }, + { + "epoch": 0.16890380313199105, + "grad_norm": 2.540706157684326, + "learning_rate": 4.656224895783421e-05, + "loss": 4.133, + "step": 151 + }, + { + "epoch": 0.17002237136465326, + "grad_norm": 2.285813808441162, + "learning_rate": 4.651765621436303e-05, + "loss": 4.4777, + "step": 152 + }, + { + "epoch": 0.17114093959731544, + "grad_norm": 3.893460273742676, + "learning_rate": 4.6472797753962246e-05, + "loss": 4.2023, + "step": 153 + }, + { + "epoch": 0.17225950782997762, + "grad_norm": 2.54040789604187, + "learning_rate": 4.6427674130579424e-05, + "loss": 4.2924, + "step": 154 + }, + { + "epoch": 0.17337807606263983, + "grad_norm": 2.005213975906372, + "learning_rate": 4.6382285901436495e-05, + "loss": 3.8809, + "step": 155 + }, + { + "epoch": 0.174496644295302, + "grad_norm": 3.2139828205108643, + "learning_rate": 4.633663362702299e-05, + "loss": 4.5209, + "step": 156 + }, + { + "epoch": 0.1756152125279642, + "grad_norm": 2.509523868560791, + "learning_rate": 4.62907178710891e-05, + "loss": 4.3067, + "step": 157 + }, + { + "epoch": 0.1767337807606264, + "grad_norm": 2.318631649017334, + "learning_rate": 4.6244539200638626e-05, + "loss": 4.0036, + "step": 158 + }, + { + "epoch": 0.17785234899328858, + "grad_norm": 2.146442413330078, + "learning_rate": 4.61980981859221e-05, + "loss": 4.3332, + "step": 159 + }, + { + "epoch": 0.1789709172259508, + "grad_norm": 2.574519634246826, + "learning_rate": 4.615139540042966e-05, + "loss": 4.0637, + "step": 160 + }, + { + "epoch": 0.18008948545861297, + "grad_norm": 2.722486972808838, + "learning_rate": 4.610443142088402e-05, + "loss": 4.1284, + "step": 161 + }, + { + "epoch": 0.18120805369127516, + "grad_norm": 2.3555033206939697, + "learning_rate": 4.60572068272333e-05, + "loss": 4.2058, + "step": 162 + }, + { + "epoch": 0.18232662192393737, + "grad_norm": 2.389615297317505, + "learning_rate": 4.60097222026439e-05, + "loss": 4.0544, + "step": 163 + }, + { + "epoch": 0.18344519015659955, + "grad_norm": 2.4774980545043945, + "learning_rate": 4.596197813349328e-05, + "loss": 4.0337, + "step": 164 + }, + { + "epoch": 0.18456375838926176, + "grad_norm": 2.1499996185302734, + "learning_rate": 4.591397520936271e-05, + "loss": 4.1015, + "step": 165 + }, + { + "epoch": 0.18568232662192394, + "grad_norm": 2.3946421146392822, + "learning_rate": 4.586571402303006e-05, + "loss": 3.8225, + "step": 166 + }, + { + "epoch": 0.18680089485458612, + "grad_norm": 2.3064377307891846, + "learning_rate": 4.581719517046236e-05, + "loss": 4.0789, + "step": 167 + }, + { + "epoch": 0.18791946308724833, + "grad_norm": 2.358966827392578, + "learning_rate": 4.576841925080853e-05, + "loss": 3.9453, + "step": 168 + }, + { + "epoch": 0.1890380313199105, + "grad_norm": 1.9163504838943481, + "learning_rate": 4.5719386866391976e-05, + "loss": 3.9758, + "step": 169 + }, + { + "epoch": 0.19015659955257272, + "grad_norm": 2.147257089614868, + "learning_rate": 4.56700986227031e-05, + "loss": 4.0456, + "step": 170 + }, + { + "epoch": 0.1912751677852349, + "grad_norm": 2.340684413909912, + "learning_rate": 4.5620555128391884e-05, + "loss": 3.9842, + "step": 171 + }, + { + "epoch": 0.19239373601789708, + "grad_norm": 1.9256740808486938, + "learning_rate": 4.557075699526032e-05, + "loss": 3.9314, + "step": 172 + }, + { + "epoch": 0.1935123042505593, + "grad_norm": 2.1142566204071045, + "learning_rate": 4.552070483825489e-05, + "loss": 4.216, + "step": 173 + }, + { + "epoch": 0.19463087248322147, + "grad_norm": 2.934136152267456, + "learning_rate": 4.5470399275458985e-05, + "loss": 4.3844, + "step": 174 + }, + { + "epoch": 0.19574944071588368, + "grad_norm": 2.802530527114868, + "learning_rate": 4.541984092808521e-05, + "loss": 4.1614, + "step": 175 + }, + { + "epoch": 0.19686800894854586, + "grad_norm": 3.0535061359405518, + "learning_rate": 4.536903042046777e-05, + "loss": 4.0901, + "step": 176 + }, + { + "epoch": 0.19798657718120805, + "grad_norm": 3.6601295471191406, + "learning_rate": 4.531796838005477e-05, + "loss": 4.1447, + "step": 177 + }, + { + "epoch": 0.19910514541387025, + "grad_norm": 2.675509214401245, + "learning_rate": 4.526665543740038e-05, + "loss": 4.0328, + "step": 178 + }, + { + "epoch": 0.20022371364653244, + "grad_norm": 3.0935540199279785, + "learning_rate": 4.5215092226157165e-05, + "loss": 4.0755, + "step": 179 + }, + { + "epoch": 0.20134228187919462, + "grad_norm": 2.0524206161499023, + "learning_rate": 4.516327938306818e-05, + "loss": 4.2309, + "step": 180 + }, + { + "epoch": 0.20246085011185683, + "grad_norm": 2.2571775913238525, + "learning_rate": 4.5111217547959114e-05, + "loss": 4.1893, + "step": 181 + }, + { + "epoch": 0.203579418344519, + "grad_norm": 3.665895462036133, + "learning_rate": 4.505890736373045e-05, + "loss": 4.2883, + "step": 182 + }, + { + "epoch": 0.20469798657718122, + "grad_norm": 2.241145372390747, + "learning_rate": 4.500634947634943e-05, + "loss": 4.1803, + "step": 183 + }, + { + "epoch": 0.2058165548098434, + "grad_norm": 2.3424227237701416, + "learning_rate": 4.495354453484216e-05, + "loss": 3.9584, + "step": 184 + }, + { + "epoch": 0.20693512304250558, + "grad_norm": 2.7996134757995605, + "learning_rate": 4.4900493191285554e-05, + "loss": 3.977, + "step": 185 + }, + { + "epoch": 0.2080536912751678, + "grad_norm": 2.122816324234009, + "learning_rate": 4.48471961007993e-05, + "loss": 4.1925, + "step": 186 + }, + { + "epoch": 0.20917225950782997, + "grad_norm": 2.3053228855133057, + "learning_rate": 4.479365392153776e-05, + "loss": 4.3152, + "step": 187 + }, + { + "epoch": 0.21029082774049218, + "grad_norm": 2.4579198360443115, + "learning_rate": 4.473986731468183e-05, + "loss": 3.8768, + "step": 188 + }, + { + "epoch": 0.21140939597315436, + "grad_norm": 4.130041599273682, + "learning_rate": 4.4685836944430816e-05, + "loss": 3.9151, + "step": 189 + }, + { + "epoch": 0.21252796420581654, + "grad_norm": 3.137632369995117, + "learning_rate": 4.4631563477994184e-05, + "loss": 4.2563, + "step": 190 + }, + { + "epoch": 0.21364653243847875, + "grad_norm": 2.419900894165039, + "learning_rate": 4.457704758558335e-05, + "loss": 4.0084, + "step": 191 + }, + { + "epoch": 0.21476510067114093, + "grad_norm": 2.5032031536102295, + "learning_rate": 4.4522289940403404e-05, + "loss": 4.2155, + "step": 192 + }, + { + "epoch": 0.21588366890380314, + "grad_norm": 2.497797727584839, + "learning_rate": 4.446729121864478e-05, + "loss": 3.8909, + "step": 193 + }, + { + "epoch": 0.21700223713646533, + "grad_norm": 3.136139154434204, + "learning_rate": 4.441205209947491e-05, + "loss": 4.45, + "step": 194 + }, + { + "epoch": 0.2181208053691275, + "grad_norm": 2.3615260124206543, + "learning_rate": 4.435657326502986e-05, + "loss": 4.3952, + "step": 195 + }, + { + "epoch": 0.21923937360178972, + "grad_norm": 2.7231903076171875, + "learning_rate": 4.430085540040587e-05, + "loss": 4.0338, + "step": 196 + }, + { + "epoch": 0.2203579418344519, + "grad_norm": 2.807499408721924, + "learning_rate": 4.4244899193650933e-05, + "loss": 4.4239, + "step": 197 + }, + { + "epoch": 0.2214765100671141, + "grad_norm": 2.6443018913269043, + "learning_rate": 4.418870533575625e-05, + "loss": 4.1129, + "step": 198 + }, + { + "epoch": 0.2225950782997763, + "grad_norm": 2.2121267318725586, + "learning_rate": 4.4132274520647754e-05, + "loss": 4.2689, + "step": 199 + }, + { + "epoch": 0.22371364653243847, + "grad_norm": 2.53283953666687, + "learning_rate": 4.407560744517749e-05, + "loss": 4.1621, + "step": 200 + }, + { + "epoch": 0.22483221476510068, + "grad_norm": 2.109092950820923, + "learning_rate": 4.401870480911504e-05, + "loss": 4.2749, + "step": 201 + }, + { + "epoch": 0.22595078299776286, + "grad_norm": 2.3910768032073975, + "learning_rate": 4.3961567315138885e-05, + "loss": 4.2594, + "step": 202 + }, + { + "epoch": 0.22706935123042504, + "grad_norm": 3.1483852863311768, + "learning_rate": 4.3904195668827697e-05, + "loss": 3.8371, + "step": 203 + }, + { + "epoch": 0.22818791946308725, + "grad_norm": 3.3514859676361084, + "learning_rate": 4.384659057865165e-05, + "loss": 4.1402, + "step": 204 + }, + { + "epoch": 0.22930648769574943, + "grad_norm": 4.020583629608154, + "learning_rate": 4.378875275596367e-05, + "loss": 3.8622, + "step": 205 + }, + { + "epoch": 0.23042505592841164, + "grad_norm": 2.541602611541748, + "learning_rate": 4.3730682914990653e-05, + "loss": 4.3944, + "step": 206 + }, + { + "epoch": 0.23154362416107382, + "grad_norm": 2.3997037410736084, + "learning_rate": 4.3672381772824615e-05, + "loss": 4.1781, + "step": 207 + }, + { + "epoch": 0.232662192393736, + "grad_norm": 3.2126035690307617, + "learning_rate": 4.36138500494139e-05, + "loss": 4.115, + "step": 208 + }, + { + "epoch": 0.23378076062639822, + "grad_norm": 2.839704990386963, + "learning_rate": 4.355508846755422e-05, + "loss": 3.9238, + "step": 209 + }, + { + "epoch": 0.2348993288590604, + "grad_norm": 3.05448317527771, + "learning_rate": 4.3496097752879764e-05, + "loss": 4.11, + "step": 210 + }, + { + "epoch": 0.2360178970917226, + "grad_norm": 2.8709630966186523, + "learning_rate": 4.343687863385425e-05, + "loss": 3.9743, + "step": 211 + }, + { + "epoch": 0.2371364653243848, + "grad_norm": 2.701570749282837, + "learning_rate": 4.3377431841761875e-05, + "loss": 3.8432, + "step": 212 + }, + { + "epoch": 0.23825503355704697, + "grad_norm": 2.5893900394439697, + "learning_rate": 4.331775811069837e-05, + "loss": 4.0372, + "step": 213 + }, + { + "epoch": 0.23937360178970918, + "grad_norm": 2.2405996322631836, + "learning_rate": 4.325785817756186e-05, + "loss": 3.8983, + "step": 214 + }, + { + "epoch": 0.24049217002237136, + "grad_norm": 2.6390066146850586, + "learning_rate": 4.3197732782043784e-05, + "loss": 4.2161, + "step": 215 + }, + { + "epoch": 0.24161073825503357, + "grad_norm": 2.5067667961120605, + "learning_rate": 4.3137382666619783e-05, + "loss": 4.3013, + "step": 216 + }, + { + "epoch": 0.24272930648769575, + "grad_norm": 2.50183367729187, + "learning_rate": 4.307680857654052e-05, + "loss": 4.1306, + "step": 217 + }, + { + "epoch": 0.24384787472035793, + "grad_norm": 2.526111364364624, + "learning_rate": 4.301601125982245e-05, + "loss": 3.9925, + "step": 218 + }, + { + "epoch": 0.24496644295302014, + "grad_norm": 2.189241886138916, + "learning_rate": 4.2954991467238634e-05, + "loss": 4.1998, + "step": 219 + }, + { + "epoch": 0.24608501118568232, + "grad_norm": 2.3332066535949707, + "learning_rate": 4.289374995230942e-05, + "loss": 4.1606, + "step": 220 + }, + { + "epoch": 0.24720357941834453, + "grad_norm": 3.0104753971099854, + "learning_rate": 4.2832287471293155e-05, + "loss": 4.3294, + "step": 221 + }, + { + "epoch": 0.2483221476510067, + "grad_norm": 3.3456063270568848, + "learning_rate": 4.277060478317687e-05, + "loss": 4.3516, + "step": 222 + }, + { + "epoch": 0.2494407158836689, + "grad_norm": 1.9993607997894287, + "learning_rate": 4.270870264966687e-05, + "loss": 4.2042, + "step": 223 + }, + { + "epoch": 0.2505592841163311, + "grad_norm": 3.882202625274658, + "learning_rate": 4.264658183517935e-05, + "loss": 4.1279, + "step": 224 + }, + { + "epoch": 0.2516778523489933, + "grad_norm": 2.162135124206543, + "learning_rate": 4.258424310683094e-05, + "loss": 4.0197, + "step": 225 + }, + { + "epoch": 0.25279642058165547, + "grad_norm": 1.9756934642791748, + "learning_rate": 4.2521687234429264e-05, + "loss": 4.1365, + "step": 226 + }, + { + "epoch": 0.2539149888143177, + "grad_norm": 2.928903579711914, + "learning_rate": 4.245891499046338e-05, + "loss": 4.1895, + "step": 227 + }, + { + "epoch": 0.2550335570469799, + "grad_norm": 2.3427515029907227, + "learning_rate": 4.239592715009429e-05, + "loss": 4.3197, + "step": 228 + }, + { + "epoch": 0.25615212527964204, + "grad_norm": 2.0358877182006836, + "learning_rate": 4.2332724491145374e-05, + "loss": 4.0241, + "step": 229 + }, + { + "epoch": 0.25727069351230425, + "grad_norm": 3.108318567276001, + "learning_rate": 4.226930779409271e-05, + "loss": 3.8592, + "step": 230 + }, + { + "epoch": 0.25838926174496646, + "grad_norm": 2.0987706184387207, + "learning_rate": 4.2205677842055516e-05, + "loss": 4.1741, + "step": 231 + }, + { + "epoch": 0.2595078299776286, + "grad_norm": 2.86515474319458, + "learning_rate": 4.214183542078646e-05, + "loss": 4.0576, + "step": 232 + }, + { + "epoch": 0.2606263982102908, + "grad_norm": 3.104260206222534, + "learning_rate": 4.207778131866191e-05, + "loss": 3.9961, + "step": 233 + }, + { + "epoch": 0.26174496644295303, + "grad_norm": 2.444136142730713, + "learning_rate": 4.2013516326672273e-05, + "loss": 4.0128, + "step": 234 + }, + { + "epoch": 0.26286353467561524, + "grad_norm": 2.4717466831207275, + "learning_rate": 4.194904123841218e-05, + "loss": 4.4341, + "step": 235 + }, + { + "epoch": 0.2639821029082774, + "grad_norm": 2.2069778442382812, + "learning_rate": 4.188435685007069e-05, + "loss": 3.9367, + "step": 236 + }, + { + "epoch": 0.2651006711409396, + "grad_norm": 2.673962354660034, + "learning_rate": 4.1819463960421454e-05, + "loss": 3.9492, + "step": 237 + }, + { + "epoch": 0.2662192393736018, + "grad_norm": 2.869675397872925, + "learning_rate": 4.175436337081289e-05, + "loss": 4.2262, + "step": 238 + }, + { + "epoch": 0.26733780760626397, + "grad_norm": 2.8593602180480957, + "learning_rate": 4.168905588515822e-05, + "loss": 3.9327, + "step": 239 + }, + { + "epoch": 0.2684563758389262, + "grad_norm": 3.113313674926758, + "learning_rate": 4.162354230992562e-05, + "loss": 4.3994, + "step": 240 + }, + { + "epoch": 0.2695749440715884, + "grad_norm": 3.0361926555633545, + "learning_rate": 4.155782345412819e-05, + "loss": 4.2951, + "step": 241 + }, + { + "epoch": 0.27069351230425054, + "grad_norm": 2.609422206878662, + "learning_rate": 4.149190012931402e-05, + "loss": 4.1109, + "step": 242 + }, + { + "epoch": 0.27181208053691275, + "grad_norm": 2.932523250579834, + "learning_rate": 4.1425773149556134e-05, + "loss": 4.1605, + "step": 243 + }, + { + "epoch": 0.27293064876957496, + "grad_norm": 2.0494723320007324, + "learning_rate": 4.135944333144244e-05, + "loss": 4.2708, + "step": 244 + }, + { + "epoch": 0.2740492170022371, + "grad_norm": 2.5390541553497314, + "learning_rate": 4.129291149406567e-05, + "loss": 3.9803, + "step": 245 + }, + { + "epoch": 0.2751677852348993, + "grad_norm": 2.236856460571289, + "learning_rate": 4.122617845901322e-05, + "loss": 4.1316, + "step": 246 + }, + { + "epoch": 0.27628635346756153, + "grad_norm": 2.3391566276550293, + "learning_rate": 4.1159245050357065e-05, + "loss": 4.2234, + "step": 247 + }, + { + "epoch": 0.27740492170022374, + "grad_norm": 2.809293031692505, + "learning_rate": 4.1092112094643543e-05, + "loss": 4.0531, + "step": 248 + }, + { + "epoch": 0.2785234899328859, + "grad_norm": 2.943033456802368, + "learning_rate": 4.102478042088315e-05, + "loss": 3.9868, + "step": 249 + }, + { + "epoch": 0.2796420581655481, + "grad_norm": 1.9977362155914307, + "learning_rate": 4.095725086054029e-05, + "loss": 3.7822, + "step": 250 + }, + { + "epoch": 0.2807606263982103, + "grad_norm": 2.405503988265991, + "learning_rate": 4.088952424752307e-05, + "loss": 4.1802, + "step": 251 + }, + { + "epoch": 0.28187919463087246, + "grad_norm": 2.7163617610931396, + "learning_rate": 4.082160141817293e-05, + "loss": 4.1736, + "step": 252 + }, + { + "epoch": 0.2829977628635347, + "grad_norm": 2.2036311626434326, + "learning_rate": 4.075348321125433e-05, + "loss": 3.9607, + "step": 253 + }, + { + "epoch": 0.2841163310961969, + "grad_norm": 3.2602269649505615, + "learning_rate": 4.068517046794443e-05, + "loss": 3.984, + "step": 254 + }, + { + "epoch": 0.28523489932885904, + "grad_norm": 3.4510889053344727, + "learning_rate": 4.0616664031822684e-05, + "loss": 4.1701, + "step": 255 + }, + { + "epoch": 0.28635346756152125, + "grad_norm": 2.2935943603515625, + "learning_rate": 4.0547964748860386e-05, + "loss": 4.1199, + "step": 256 + }, + { + "epoch": 0.28747203579418346, + "grad_norm": 3.6087756156921387, + "learning_rate": 4.0479073467410286e-05, + "loss": 4.0622, + "step": 257 + }, + { + "epoch": 0.28859060402684567, + "grad_norm": 3.2882027626037598, + "learning_rate": 4.040999103819606e-05, + "loss": 4.2735, + "step": 258 + }, + { + "epoch": 0.2897091722595078, + "grad_norm": 2.9699313640594482, + "learning_rate": 4.034071831430184e-05, + "loss": 3.9871, + "step": 259 + }, + { + "epoch": 0.29082774049217003, + "grad_norm": 2.889160394668579, + "learning_rate": 4.0271256151161664e-05, + "loss": 4.0421, + "step": 260 + }, + { + "epoch": 0.29194630872483224, + "grad_norm": 2.320803165435791, + "learning_rate": 4.020160540654891e-05, + "loss": 4.4205, + "step": 261 + }, + { + "epoch": 0.2930648769574944, + "grad_norm": 3.7101683616638184, + "learning_rate": 4.0131766940565715e-05, + "loss": 3.9198, + "step": 262 + }, + { + "epoch": 0.2941834451901566, + "grad_norm": 2.4428420066833496, + "learning_rate": 4.006174161563233e-05, + "loss": 4.1809, + "step": 263 + }, + { + "epoch": 0.2953020134228188, + "grad_norm": 3.8638551235198975, + "learning_rate": 3.999153029647651e-05, + "loss": 3.9782, + "step": 264 + }, + { + "epoch": 0.29642058165548096, + "grad_norm": 1.7899916172027588, + "learning_rate": 3.9921133850122805e-05, + "loss": 4.3221, + "step": 265 + }, + { + "epoch": 0.2975391498881432, + "grad_norm": 3.8810582160949707, + "learning_rate": 3.9850553145881854e-05, + "loss": 3.8899, + "step": 266 + }, + { + "epoch": 0.2986577181208054, + "grad_norm": 2.2642791271209717, + "learning_rate": 3.9779789055339656e-05, + "loss": 4.0838, + "step": 267 + }, + { + "epoch": 0.29977628635346754, + "grad_norm": 2.4321177005767822, + "learning_rate": 3.9708842452346836e-05, + "loss": 4.2623, + "step": 268 + }, + { + "epoch": 0.30089485458612975, + "grad_norm": 2.04886794090271, + "learning_rate": 3.963771421300777e-05, + "loss": 4.1379, + "step": 269 + }, + { + "epoch": 0.30201342281879195, + "grad_norm": 1.9282593727111816, + "learning_rate": 3.956640521566989e-05, + "loss": 4.2569, + "step": 270 + }, + { + "epoch": 0.30313199105145416, + "grad_norm": 2.2058980464935303, + "learning_rate": 3.949491634091272e-05, + "loss": 4.7175, + "step": 271 + }, + { + "epoch": 0.3042505592841163, + "grad_norm": 2.039492130279541, + "learning_rate": 3.9423248471537065e-05, + "loss": 4.234, + "step": 272 + }, + { + "epoch": 0.3053691275167785, + "grad_norm": 1.9331101179122925, + "learning_rate": 3.935140249255412e-05, + "loss": 3.9734, + "step": 273 + }, + { + "epoch": 0.30648769574944074, + "grad_norm": 2.2190299034118652, + "learning_rate": 3.9279379291174465e-05, + "loss": 4.1752, + "step": 274 + }, + { + "epoch": 0.3076062639821029, + "grad_norm": 2.8630807399749756, + "learning_rate": 3.920717975679723e-05, + "loss": 4.0549, + "step": 275 + }, + { + "epoch": 0.3087248322147651, + "grad_norm": 2.7145142555236816, + "learning_rate": 3.913480478099897e-05, + "loss": 4.1861, + "step": 276 + }, + { + "epoch": 0.3098434004474273, + "grad_norm": 3.505417823791504, + "learning_rate": 3.9062255257522794e-05, + "loss": 4.0303, + "step": 277 + }, + { + "epoch": 0.31096196868008946, + "grad_norm": 3.250004291534424, + "learning_rate": 3.8989532082267225e-05, + "loss": 4.2068, + "step": 278 + }, + { + "epoch": 0.31208053691275167, + "grad_norm": 2.1549458503723145, + "learning_rate": 3.891663615327518e-05, + "loss": 3.921, + "step": 279 + }, + { + "epoch": 0.3131991051454139, + "grad_norm": 2.008209705352783, + "learning_rate": 3.884356837072288e-05, + "loss": 3.8706, + "step": 280 + }, + { + "epoch": 0.3143176733780761, + "grad_norm": 3.1446456909179688, + "learning_rate": 3.877032963690873e-05, + "loss": 4.0936, + "step": 281 + }, + { + "epoch": 0.31543624161073824, + "grad_norm": 2.9265387058258057, + "learning_rate": 3.8696920856242174e-05, + "loss": 4.2987, + "step": 282 + }, + { + "epoch": 0.31655480984340045, + "grad_norm": 2.246251344680786, + "learning_rate": 3.8623342935232525e-05, + "loss": 4.1966, + "step": 283 + }, + { + "epoch": 0.31767337807606266, + "grad_norm": 2.046858549118042, + "learning_rate": 3.854959678247778e-05, + "loss": 4.3037, + "step": 284 + }, + { + "epoch": 0.3187919463087248, + "grad_norm": 2.070281744003296, + "learning_rate": 3.847568330865338e-05, + "loss": 3.8673, + "step": 285 + }, + { + "epoch": 0.319910514541387, + "grad_norm": 2.2606728076934814, + "learning_rate": 3.8401603426501e-05, + "loss": 4.3719, + "step": 286 + }, + { + "epoch": 0.32102908277404923, + "grad_norm": 2.7562570571899414, + "learning_rate": 3.8327358050817234e-05, + "loss": 3.8372, + "step": 287 + }, + { + "epoch": 0.3221476510067114, + "grad_norm": 2.6188101768493652, + "learning_rate": 3.8252948098442344e-05, + "loss": 3.9952, + "step": 288 + }, + { + "epoch": 0.3232662192393736, + "grad_norm": 3.580763101577759, + "learning_rate": 3.817837448824888e-05, + "loss": 4.0735, + "step": 289 + }, + { + "epoch": 0.3243847874720358, + "grad_norm": 2.4125702381134033, + "learning_rate": 3.81036381411304e-05, + "loss": 4.0671, + "step": 290 + }, + { + "epoch": 0.32550335570469796, + "grad_norm": 2.0095221996307373, + "learning_rate": 3.8028739979990066e-05, + "loss": 4.1353, + "step": 291 + }, + { + "epoch": 0.32662192393736017, + "grad_norm": 2.5612494945526123, + "learning_rate": 3.7953680929729215e-05, + "loss": 3.9072, + "step": 292 + }, + { + "epoch": 0.3277404921700224, + "grad_norm": 2.8321776390075684, + "learning_rate": 3.787846191723599e-05, + "loss": 4.1067, + "step": 293 + }, + { + "epoch": 0.3288590604026846, + "grad_norm": 2.400554895401001, + "learning_rate": 3.780308387137387e-05, + "loss": 4.1338, + "step": 294 + }, + { + "epoch": 0.32997762863534674, + "grad_norm": 2.4202167987823486, + "learning_rate": 3.772754772297022e-05, + "loss": 4.139, + "step": 295 + }, + { + "epoch": 0.33109619686800895, + "grad_norm": 2.7618842124938965, + "learning_rate": 3.7651854404804755e-05, + "loss": 3.7908, + "step": 296 + }, + { + "epoch": 0.33221476510067116, + "grad_norm": 3.0693230628967285, + "learning_rate": 3.757600485159805e-05, + "loss": 3.9235, + "step": 297 + }, + { + "epoch": 0.3333333333333333, + "grad_norm": 1.86801016330719, + "learning_rate": 3.7500000000000003e-05, + "loss": 4.1265, + "step": 298 + }, + { + "epoch": 0.3344519015659955, + "grad_norm": 2.0775158405303955, + "learning_rate": 3.742384078857824e-05, + "loss": 4.0009, + "step": 299 + }, + { + "epoch": 0.33557046979865773, + "grad_norm": 2.6333956718444824, + "learning_rate": 3.7347528157806586e-05, + "loss": 4.161, + "step": 300 + }, + { + "epoch": 0.3366890380313199, + "grad_norm": 2.197185754776001, + "learning_rate": 3.727106305005336e-05, + "loss": 3.9653, + "step": 301 + }, + { + "epoch": 0.3378076062639821, + "grad_norm": 2.323561191558838, + "learning_rate": 3.719444640956981e-05, + "loss": 4.0114, + "step": 302 + }, + { + "epoch": 0.3389261744966443, + "grad_norm": 2.5178816318511963, + "learning_rate": 3.7117679182478417e-05, + "loss": 3.8501, + "step": 303 + }, + { + "epoch": 0.3400447427293065, + "grad_norm": 2.5967633724212646, + "learning_rate": 3.704076231676125e-05, + "loss": 3.7106, + "step": 304 + }, + { + "epoch": 0.34116331096196867, + "grad_norm": 2.9387967586517334, + "learning_rate": 3.696369676224819e-05, + "loss": 3.9001, + "step": 305 + }, + { + "epoch": 0.3422818791946309, + "grad_norm": 2.7642838954925537, + "learning_rate": 3.688648347060529e-05, + "loss": 4.0384, + "step": 306 + }, + { + "epoch": 0.3434004474272931, + "grad_norm": 2.5087080001831055, + "learning_rate": 3.680912339532296e-05, + "loss": 4.1227, + "step": 307 + }, + { + "epoch": 0.34451901565995524, + "grad_norm": 2.084219217300415, + "learning_rate": 3.67316174917042e-05, + "loss": 4.0968, + "step": 308 + }, + { + "epoch": 0.34563758389261745, + "grad_norm": 3.1434361934661865, + "learning_rate": 3.66539667168528e-05, + "loss": 4.3648, + "step": 309 + }, + { + "epoch": 0.34675615212527966, + "grad_norm": 2.9753198623657227, + "learning_rate": 3.657617202966158e-05, + "loss": 3.8644, + "step": 310 + }, + { + "epoch": 0.3478747203579418, + "grad_norm": 1.990665316581726, + "learning_rate": 3.649823439080047e-05, + "loss": 4.0785, + "step": 311 + }, + { + "epoch": 0.348993288590604, + "grad_norm": 3.1740851402282715, + "learning_rate": 3.6420154762704686e-05, + "loss": 4.1794, + "step": 312 + }, + { + "epoch": 0.35011185682326623, + "grad_norm": 2.5907697677612305, + "learning_rate": 3.634193410956282e-05, + "loss": 3.887, + "step": 313 + }, + { + "epoch": 0.3512304250559284, + "grad_norm": 2.6295742988586426, + "learning_rate": 3.6263573397305e-05, + "loss": 4.0013, + "step": 314 + }, + { + "epoch": 0.3523489932885906, + "grad_norm": 3.465045213699341, + "learning_rate": 3.618507359359087e-05, + "loss": 3.9178, + "step": 315 + }, + { + "epoch": 0.3534675615212528, + "grad_norm": 2.5845601558685303, + "learning_rate": 3.6106435667797685e-05, + "loss": 3.9463, + "step": 316 + }, + { + "epoch": 0.354586129753915, + "grad_norm": 2.198699712753296, + "learning_rate": 3.602766059100838e-05, + "loss": 3.6981, + "step": 317 + }, + { + "epoch": 0.35570469798657717, + "grad_norm": 2.5952913761138916, + "learning_rate": 3.5948749335999496e-05, + "loss": 4.3058, + "step": 318 + }, + { + "epoch": 0.3568232662192394, + "grad_norm": 2.6661579608917236, + "learning_rate": 3.586970287722923e-05, + "loss": 3.9699, + "step": 319 + }, + { + "epoch": 0.3579418344519016, + "grad_norm": 2.2353670597076416, + "learning_rate": 3.5790522190825365e-05, + "loss": 3.9394, + "step": 320 + }, + { + "epoch": 0.35906040268456374, + "grad_norm": 1.800933599472046, + "learning_rate": 3.571120825457327e-05, + "loss": 3.9037, + "step": 321 + }, + { + "epoch": 0.36017897091722595, + "grad_norm": 2.2671263217926025, + "learning_rate": 3.563176204790374e-05, + "loss": 4.2906, + "step": 322 + }, + { + "epoch": 0.36129753914988816, + "grad_norm": 2.249171733856201, + "learning_rate": 3.555218455188099e-05, + "loss": 4.0766, + "step": 323 + }, + { + "epoch": 0.3624161073825503, + "grad_norm": 2.222735643386841, + "learning_rate": 3.547247674919046e-05, + "loss": 4.0171, + "step": 324 + }, + { + "epoch": 0.3635346756152125, + "grad_norm": 2.537754774093628, + "learning_rate": 3.539263962412676e-05, + "loss": 3.9333, + "step": 325 + }, + { + "epoch": 0.36465324384787473, + "grad_norm": 3.2057714462280273, + "learning_rate": 3.5312674162581436e-05, + "loss": 4.1573, + "step": 326 + }, + { + "epoch": 0.36577181208053694, + "grad_norm": 3.4859671592712402, + "learning_rate": 3.523258135203087e-05, + "loss": 3.8762, + "step": 327 + }, + { + "epoch": 0.3668903803131991, + "grad_norm": 3.155507802963257, + "learning_rate": 3.5152362181524014e-05, + "loss": 3.9051, + "step": 328 + }, + { + "epoch": 0.3680089485458613, + "grad_norm": 2.1552867889404297, + "learning_rate": 3.507201764167024e-05, + "loss": 3.8245, + "step": 329 + }, + { + "epoch": 0.3691275167785235, + "grad_norm": 2.492128372192383, + "learning_rate": 3.4991548724627054e-05, + "loss": 3.9806, + "step": 330 + }, + { + "epoch": 0.37024608501118567, + "grad_norm": 2.913975715637207, + "learning_rate": 3.491095642408791e-05, + "loss": 4.0708, + "step": 331 + }, + { + "epoch": 0.3713646532438479, + "grad_norm": 1.982664942741394, + "learning_rate": 3.483024173526985e-05, + "loss": 3.7897, + "step": 332 + }, + { + "epoch": 0.3724832214765101, + "grad_norm": 2.7133326530456543, + "learning_rate": 3.4749405654901294e-05, + "loss": 4.0167, + "step": 333 + }, + { + "epoch": 0.37360178970917224, + "grad_norm": 2.769604444503784, + "learning_rate": 3.46684491812097e-05, + "loss": 4.2449, + "step": 334 + }, + { + "epoch": 0.37472035794183445, + "grad_norm": 2.3599345684051514, + "learning_rate": 3.45873733139092e-05, + "loss": 4.0715, + "step": 335 + }, + { + "epoch": 0.37583892617449666, + "grad_norm": 3.5112671852111816, + "learning_rate": 3.4506179054188344e-05, + "loss": 3.896, + "step": 336 + }, + { + "epoch": 0.3769574944071588, + "grad_norm": 2.5493431091308594, + "learning_rate": 3.442486740469766e-05, + "loss": 3.8645, + "step": 337 + }, + { + "epoch": 0.378076062639821, + "grad_norm": 2.665292501449585, + "learning_rate": 3.434343936953729e-05, + "loss": 3.9132, + "step": 338 + }, + { + "epoch": 0.37919463087248323, + "grad_norm": 2.7781031131744385, + "learning_rate": 3.426189595424461e-05, + "loss": 4.0039, + "step": 339 + }, + { + "epoch": 0.38031319910514544, + "grad_norm": 3.5199902057647705, + "learning_rate": 3.41802381657818e-05, + "loss": 4.0221, + "step": 340 + }, + { + "epoch": 0.3814317673378076, + "grad_norm": 2.529172897338867, + "learning_rate": 3.4098467012523404e-05, + "loss": 3.9565, + "step": 341 + }, + { + "epoch": 0.3825503355704698, + "grad_norm": 3.5351314544677734, + "learning_rate": 3.401658350424389e-05, + "loss": 4.2232, + "step": 342 + }, + { + "epoch": 0.383668903803132, + "grad_norm": 2.705423593521118, + "learning_rate": 3.393458865210516e-05, + "loss": 3.8003, + "step": 343 + }, + { + "epoch": 0.38478747203579416, + "grad_norm": 2.4735591411590576, + "learning_rate": 3.38524834686441e-05, + "loss": 3.9762, + "step": 344 + }, + { + "epoch": 0.3859060402684564, + "grad_norm": 3.4464943408966064, + "learning_rate": 3.377026896776003e-05, + "loss": 4.2064, + "step": 345 + }, + { + "epoch": 0.3870246085011186, + "grad_norm": 2.1392390727996826, + "learning_rate": 3.368794616470222e-05, + "loss": 3.9257, + "step": 346 + }, + { + "epoch": 0.38814317673378074, + "grad_norm": 3.5217792987823486, + "learning_rate": 3.360551607605735e-05, + "loss": 3.7378, + "step": 347 + }, + { + "epoch": 0.38926174496644295, + "grad_norm": 2.3355872631073, + "learning_rate": 3.3522979719736926e-05, + "loss": 4.3103, + "step": 348 + }, + { + "epoch": 0.39038031319910516, + "grad_norm": 1.909463882446289, + "learning_rate": 3.344033811496475e-05, + "loss": 3.9676, + "step": 349 + }, + { + "epoch": 0.39149888143176736, + "grad_norm": 2.3093831539154053, + "learning_rate": 3.3357592282264296e-05, + "loss": 3.915, + "step": 350 + }, + { + "epoch": 0.3926174496644295, + "grad_norm": 2.086834192276001, + "learning_rate": 3.327474324344614e-05, + "loss": 3.9272, + "step": 351 + }, + { + "epoch": 0.39373601789709173, + "grad_norm": 2.3872201442718506, + "learning_rate": 3.3191792021595316e-05, + "loss": 3.8731, + "step": 352 + }, + { + "epoch": 0.39485458612975394, + "grad_norm": 2.4958016872406006, + "learning_rate": 3.310873964105872e-05, + "loss": 3.9696, + "step": 353 + }, + { + "epoch": 0.3959731543624161, + "grad_norm": 2.3622846603393555, + "learning_rate": 3.302558712743241e-05, + "loss": 3.6666, + "step": 354 + }, + { + "epoch": 0.3970917225950783, + "grad_norm": 3.684812068939209, + "learning_rate": 3.2942335507548966e-05, + "loss": 3.9342, + "step": 355 + }, + { + "epoch": 0.3982102908277405, + "grad_norm": 2.6267335414886475, + "learning_rate": 3.285898580946482e-05, + "loss": 4.0503, + "step": 356 + }, + { + "epoch": 0.39932885906040266, + "grad_norm": 2.092944860458374, + "learning_rate": 3.277553906244756e-05, + "loss": 4.1497, + "step": 357 + }, + { + "epoch": 0.4004474272930649, + "grad_norm": 2.3737025260925293, + "learning_rate": 3.2691996296963186e-05, + "loss": 3.8479, + "step": 358 + }, + { + "epoch": 0.4015659955257271, + "grad_norm": 2.4298672676086426, + "learning_rate": 3.260835854466342e-05, + "loss": 4.0085, + "step": 359 + }, + { + "epoch": 0.40268456375838924, + "grad_norm": 2.19952130317688, + "learning_rate": 3.252462683837297e-05, + "loss": 4.4371, + "step": 360 + }, + { + "epoch": 0.40380313199105144, + "grad_norm": 2.4244048595428467, + "learning_rate": 3.244080221207674e-05, + "loss": 4.2144, + "step": 361 + }, + { + "epoch": 0.40492170022371365, + "grad_norm": 2.4055869579315186, + "learning_rate": 3.23568857009071e-05, + "loss": 3.9106, + "step": 362 + }, + { + "epoch": 0.40604026845637586, + "grad_norm": 2.3508224487304688, + "learning_rate": 3.2272878341131075e-05, + "loss": 4.237, + "step": 363 + }, + { + "epoch": 0.407158836689038, + "grad_norm": 2.4762392044067383, + "learning_rate": 3.218878117013756e-05, + "loss": 4.0447, + "step": 364 + }, + { + "epoch": 0.4082774049217002, + "grad_norm": 2.0015032291412354, + "learning_rate": 3.210459522642452e-05, + "loss": 4.0781, + "step": 365 + }, + { + "epoch": 0.40939597315436244, + "grad_norm": 2.745539665222168, + "learning_rate": 3.2020321549586154e-05, + "loss": 3.8968, + "step": 366 + }, + { + "epoch": 0.4105145413870246, + "grad_norm": 2.3843181133270264, + "learning_rate": 3.193596118030005e-05, + "loss": 4.0991, + "step": 367 + }, + { + "epoch": 0.4116331096196868, + "grad_norm": 2.248856782913208, + "learning_rate": 3.185151516031434e-05, + "loss": 4.034, + "step": 368 + }, + { + "epoch": 0.412751677852349, + "grad_norm": 2.3670918941497803, + "learning_rate": 3.1766984532434853e-05, + "loss": 4.0161, + "step": 369 + }, + { + "epoch": 0.41387024608501116, + "grad_norm": 2.298100709915161, + "learning_rate": 3.1682370340512217e-05, + "loss": 4.1546, + "step": 370 + }, + { + "epoch": 0.41498881431767337, + "grad_norm": 1.788293719291687, + "learning_rate": 3.159767362942896e-05, + "loss": 4.1429, + "step": 371 + }, + { + "epoch": 0.4161073825503356, + "grad_norm": 2.500016212463379, + "learning_rate": 3.1512895445086636e-05, + "loss": 3.7988, + "step": 372 + }, + { + "epoch": 0.4172259507829978, + "grad_norm": 3.096670627593994, + "learning_rate": 3.14280368343929e-05, + "loss": 4.406, + "step": 373 + }, + { + "epoch": 0.41834451901565994, + "grad_norm": 2.44936466217041, + "learning_rate": 3.134309884524856e-05, + "loss": 4.1445, + "step": 374 + }, + { + "epoch": 0.41946308724832215, + "grad_norm": 2.5851917266845703, + "learning_rate": 3.125808252653466e-05, + "loss": 4.6243, + "step": 375 + }, + { + "epoch": 0.42058165548098436, + "grad_norm": 2.472642183303833, + "learning_rate": 3.1172988928099525e-05, + "loss": 4.2828, + "step": 376 + }, + { + "epoch": 0.4217002237136465, + "grad_norm": 2.0510940551757812, + "learning_rate": 3.108781910074578e-05, + "loss": 3.7741, + "step": 377 + }, + { + "epoch": 0.4228187919463087, + "grad_norm": 2.4596452713012695, + "learning_rate": 3.100257409621738e-05, + "loss": 3.974, + "step": 378 + }, + { + "epoch": 0.42393736017897093, + "grad_norm": 2.0561556816101074, + "learning_rate": 3.0917254967186635e-05, + "loss": 3.8241, + "step": 379 + }, + { + "epoch": 0.4250559284116331, + "grad_norm": 1.8558775186538696, + "learning_rate": 3.0831862767241205e-05, + "loss": 4.1329, + "step": 380 + }, + { + "epoch": 0.4261744966442953, + "grad_norm": 2.460034132003784, + "learning_rate": 3.074639855087109e-05, + "loss": 4.0921, + "step": 381 + }, + { + "epoch": 0.4272930648769575, + "grad_norm": 2.0531368255615234, + "learning_rate": 3.0660863373455595e-05, + "loss": 3.8236, + "step": 382 + }, + { + "epoch": 0.42841163310961966, + "grad_norm": 2.9021363258361816, + "learning_rate": 3.057525829125032e-05, + "loss": 3.8496, + "step": 383 + }, + { + "epoch": 0.42953020134228187, + "grad_norm": 2.5614798069000244, + "learning_rate": 3.0489584361374074e-05, + "loss": 4.1064, + "step": 384 + }, + { + "epoch": 0.4306487695749441, + "grad_norm": 2.2442526817321777, + "learning_rate": 3.04038426417959e-05, + "loss": 4.1327, + "step": 385 + }, + { + "epoch": 0.4317673378076063, + "grad_norm": 2.7794744968414307, + "learning_rate": 3.031803419132192e-05, + "loss": 4.1708, + "step": 386 + }, + { + "epoch": 0.43288590604026844, + "grad_norm": 2.464383363723755, + "learning_rate": 3.0232160069582332e-05, + "loss": 4.1961, + "step": 387 + }, + { + "epoch": 0.43400447427293065, + "grad_norm": 1.990610122680664, + "learning_rate": 3.014622133701826e-05, + "loss": 3.8704, + "step": 388 + }, + { + "epoch": 0.43512304250559286, + "grad_norm": 2.630527973175049, + "learning_rate": 3.0060219054868727e-05, + "loss": 4.0868, + "step": 389 + }, + { + "epoch": 0.436241610738255, + "grad_norm": 1.9832707643508911, + "learning_rate": 2.9974154285157497e-05, + "loss": 4.0184, + "step": 390 + }, + { + "epoch": 0.4373601789709172, + "grad_norm": 2.482058048248291, + "learning_rate": 2.9888028090679982e-05, + "loss": 4.0066, + "step": 391 + }, + { + "epoch": 0.43847874720357943, + "grad_norm": 2.484102487564087, + "learning_rate": 2.9801841534990115e-05, + "loss": 4.1228, + "step": 392 + }, + { + "epoch": 0.4395973154362416, + "grad_norm": 2.7631897926330566, + "learning_rate": 2.9715595682387242e-05, + "loss": 3.863, + "step": 393 + }, + { + "epoch": 0.4407158836689038, + "grad_norm": 2.0470075607299805, + "learning_rate": 2.9629291597902898e-05, + "loss": 3.8224, + "step": 394 + }, + { + "epoch": 0.441834451901566, + "grad_norm": 1.8407243490219116, + "learning_rate": 2.954293034728776e-05, + "loss": 3.8581, + "step": 395 + }, + { + "epoch": 0.4429530201342282, + "grad_norm": 1.9803217649459839, + "learning_rate": 2.9456512996998424e-05, + "loss": 3.914, + "step": 396 + }, + { + "epoch": 0.44407158836689037, + "grad_norm": 2.7368860244750977, + "learning_rate": 2.9370040614184245e-05, + "loss": 4.0718, + "step": 397 + }, + { + "epoch": 0.4451901565995526, + "grad_norm": 5.358017444610596, + "learning_rate": 2.9283514266674168e-05, + "loss": 3.7782, + "step": 398 + }, + { + "epoch": 0.4463087248322148, + "grad_norm": 2.5471818447113037, + "learning_rate": 2.9196935022963525e-05, + "loss": 3.7342, + "step": 399 + }, + { + "epoch": 0.44742729306487694, + "grad_norm": 2.3943824768066406, + "learning_rate": 2.9110303952200863e-05, + "loss": 4.0393, + "step": 400 + }, + { + "epoch": 0.44854586129753915, + "grad_norm": 2.130612373352051, + "learning_rate": 2.902362212417472e-05, + "loss": 3.7082, + "step": 401 + }, + { + "epoch": 0.44966442953020136, + "grad_norm": 2.9800944328308105, + "learning_rate": 2.893689060930045e-05, + "loss": 3.8505, + "step": 402 + }, + { + "epoch": 0.4507829977628635, + "grad_norm": 2.35976243019104, + "learning_rate": 2.8850110478606938e-05, + "loss": 4.1292, + "step": 403 + }, + { + "epoch": 0.4519015659955257, + "grad_norm": 2.627814292907715, + "learning_rate": 2.876328280372346e-05, + "loss": 3.6394, + "step": 404 + }, + { + "epoch": 0.45302013422818793, + "grad_norm": 2.3630151748657227, + "learning_rate": 2.8676408656866353e-05, + "loss": 3.7281, + "step": 405 + }, + { + "epoch": 0.4541387024608501, + "grad_norm": 2.5293986797332764, + "learning_rate": 2.8589489110825897e-05, + "loss": 3.7605, + "step": 406 + }, + { + "epoch": 0.4552572706935123, + "grad_norm": 2.503148078918457, + "learning_rate": 2.8502525238952916e-05, + "loss": 4.1672, + "step": 407 + }, + { + "epoch": 0.4563758389261745, + "grad_norm": 2.9510624408721924, + "learning_rate": 2.8415518115145674e-05, + "loss": 3.9833, + "step": 408 + }, + { + "epoch": 0.4574944071588367, + "grad_norm": 2.3186118602752686, + "learning_rate": 2.8328468813836493e-05, + "loss": 4.1216, + "step": 409 + }, + { + "epoch": 0.45861297539149887, + "grad_norm": 3.173356294631958, + "learning_rate": 2.824137840997858e-05, + "loss": 3.9926, + "step": 410 + }, + { + "epoch": 0.4597315436241611, + "grad_norm": 1.9115232229232788, + "learning_rate": 2.8154247979032665e-05, + "loss": 4.0426, + "step": 411 + }, + { + "epoch": 0.4608501118568233, + "grad_norm": 2.5544466972351074, + "learning_rate": 2.8067078596953796e-05, + "loss": 3.9857, + "step": 412 + }, + { + "epoch": 0.46196868008948544, + "grad_norm": 1.8539265394210815, + "learning_rate": 2.7979871340178003e-05, + "loss": 3.894, + "step": 413 + }, + { + "epoch": 0.46308724832214765, + "grad_norm": 2.6608541011810303, + "learning_rate": 2.7892627285609035e-05, + "loss": 4.2197, + "step": 414 + }, + { + "epoch": 0.46420581655480986, + "grad_norm": 2.6225533485412598, + "learning_rate": 2.780534751060504e-05, + "loss": 3.8299, + "step": 415 + }, + { + "epoch": 0.465324384787472, + "grad_norm": 3.5249505043029785, + "learning_rate": 2.771803309296527e-05, + "loss": 3.4724, + "step": 416 + }, + { + "epoch": 0.4664429530201342, + "grad_norm": 2.1255197525024414, + "learning_rate": 2.7630685110916778e-05, + "loss": 4.12, + "step": 417 + }, + { + "epoch": 0.46756152125279643, + "grad_norm": 2.8229820728302, + "learning_rate": 2.754330464310108e-05, + "loss": 4.3171, + "step": 418 + }, + { + "epoch": 0.46868008948545864, + "grad_norm": 2.2736423015594482, + "learning_rate": 2.7455892768560888e-05, + "loss": 3.9784, + "step": 419 + }, + { + "epoch": 0.4697986577181208, + "grad_norm": 1.8942090272903442, + "learning_rate": 2.736845056672671e-05, + "loss": 3.7484, + "step": 420 + }, + { + "epoch": 0.470917225950783, + "grad_norm": 3.5483438968658447, + "learning_rate": 2.7280979117403575e-05, + "loss": 4.1573, + "step": 421 + }, + { + "epoch": 0.4720357941834452, + "grad_norm": 2.8570756912231445, + "learning_rate": 2.7193479500757685e-05, + "loss": 3.8323, + "step": 422 + }, + { + "epoch": 0.47315436241610737, + "grad_norm": 2.425438404083252, + "learning_rate": 2.710595279730308e-05, + "loss": 4.0871, + "step": 423 + }, + { + "epoch": 0.4742729306487696, + "grad_norm": 2.321178674697876, + "learning_rate": 2.7018400087888263e-05, + "loss": 4.1138, + "step": 424 + }, + { + "epoch": 0.4753914988814318, + "grad_norm": 2.9220635890960693, + "learning_rate": 2.6930822453682915e-05, + "loss": 3.8734, + "step": 425 + }, + { + "epoch": 0.47651006711409394, + "grad_norm": 3.5046780109405518, + "learning_rate": 2.684322097616448e-05, + "loss": 4.3458, + "step": 426 + }, + { + "epoch": 0.47762863534675615, + "grad_norm": 3.008169651031494, + "learning_rate": 2.675559673710485e-05, + "loss": 3.7159, + "step": 427 + }, + { + "epoch": 0.47874720357941836, + "grad_norm": 2.289055109024048, + "learning_rate": 2.6667950818556993e-05, + "loss": 4.2936, + "step": 428 + }, + { + "epoch": 0.4798657718120805, + "grad_norm": 1.8744932413101196, + "learning_rate": 2.658028430284159e-05, + "loss": 3.9657, + "step": 429 + }, + { + "epoch": 0.4809843400447427, + "grad_norm": 2.0840201377868652, + "learning_rate": 2.649259827253368e-05, + "loss": 4.265, + "step": 430 + }, + { + "epoch": 0.48210290827740493, + "grad_norm": 2.352440118789673, + "learning_rate": 2.6404893810449272e-05, + "loss": 3.7011, + "step": 431 + }, + { + "epoch": 0.48322147651006714, + "grad_norm": 2.6247146129608154, + "learning_rate": 2.631717199963199e-05, + "loss": 3.5364, + "step": 432 + }, + { + "epoch": 0.4843400447427293, + "grad_norm": 3.915102243423462, + "learning_rate": 2.6229433923339696e-05, + "loss": 4.2285, + "step": 433 + }, + { + "epoch": 0.4854586129753915, + "grad_norm": 2.190736770629883, + "learning_rate": 2.6141680665031116e-05, + "loss": 3.9863, + "step": 434 + }, + { + "epoch": 0.4865771812080537, + "grad_norm": 2.417031764984131, + "learning_rate": 2.6053913308352428e-05, + "loss": 3.9451, + "step": 435 + }, + { + "epoch": 0.48769574944071586, + "grad_norm": 2.4824278354644775, + "learning_rate": 2.596613293712396e-05, + "loss": 3.9539, + "step": 436 + }, + { + "epoch": 0.4888143176733781, + "grad_norm": 2.9509193897247314, + "learning_rate": 2.5878340635326686e-05, + "loss": 3.8236, + "step": 437 + }, + { + "epoch": 0.4899328859060403, + "grad_norm": 2.627767562866211, + "learning_rate": 2.5790537487088974e-05, + "loss": 4.0342, + "step": 438 + }, + { + "epoch": 0.49105145413870244, + "grad_norm": 2.270059108734131, + "learning_rate": 2.5702724576673088e-05, + "loss": 4.3256, + "step": 439 + }, + { + "epoch": 0.49217002237136465, + "grad_norm": 2.08921217918396, + "learning_rate": 2.561490298846186e-05, + "loss": 3.8144, + "step": 440 + }, + { + "epoch": 0.49328859060402686, + "grad_norm": 2.151421308517456, + "learning_rate": 2.5527073806945278e-05, + "loss": 3.881, + "step": 441 + }, + { + "epoch": 0.49440715883668906, + "grad_norm": 3.5347208976745605, + "learning_rate": 2.5439238116707102e-05, + "loss": 4.1678, + "step": 442 + }, + { + "epoch": 0.4955257270693512, + "grad_norm": 1.9540199041366577, + "learning_rate": 2.5351397002411477e-05, + "loss": 4.028, + "step": 443 + }, + { + "epoch": 0.4966442953020134, + "grad_norm": 2.151306629180908, + "learning_rate": 2.5263551548789495e-05, + "loss": 3.6328, + "step": 444 + }, + { + "epoch": 0.49776286353467564, + "grad_norm": 3.1425187587738037, + "learning_rate": 2.517570284062586e-05, + "loss": 3.9881, + "step": 445 + }, + { + "epoch": 0.4988814317673378, + "grad_norm": 2.3265583515167236, + "learning_rate": 2.5087851962745468e-05, + "loss": 3.9283, + "step": 446 + }, + { + "epoch": 0.5, + "grad_norm": 2.091590166091919, + "learning_rate": 2.5e-05, + "loss": 3.6892, + "step": 447 + }, + { + "epoch": 0.5011185682326622, + "grad_norm": 2.2790603637695312, + "learning_rate": 2.4912148037254535e-05, + "loss": 3.7534, + "step": 448 + }, + { + "epoch": 0.5022371364653244, + "grad_norm": 2.3313305377960205, + "learning_rate": 2.4824297159374142e-05, + "loss": 3.8916, + "step": 449 + }, + { + "epoch": 0.5033557046979866, + "grad_norm": 2.9091544151306152, + "learning_rate": 2.473644845121051e-05, + "loss": 4.1545, + "step": 450 + }, + { + "epoch": 0.5044742729306487, + "grad_norm": 1.9648659229278564, + "learning_rate": 2.464860299758854e-05, + "loss": 3.9658, + "step": 451 + }, + { + "epoch": 0.5055928411633109, + "grad_norm": 2.6131553649902344, + "learning_rate": 2.45607618832929e-05, + "loss": 3.9665, + "step": 452 + }, + { + "epoch": 0.5067114093959731, + "grad_norm": 2.341771364212036, + "learning_rate": 2.4472926193054728e-05, + "loss": 3.9947, + "step": 453 + }, + { + "epoch": 0.5078299776286354, + "grad_norm": 2.526048183441162, + "learning_rate": 2.4385097011538144e-05, + "loss": 4.0324, + "step": 454 + }, + { + "epoch": 0.5089485458612976, + "grad_norm": 3.4752910137176514, + "learning_rate": 2.4297275423326918e-05, + "loss": 4.0246, + "step": 455 + }, + { + "epoch": 0.5100671140939598, + "grad_norm": 2.2056233882904053, + "learning_rate": 2.420946251291103e-05, + "loss": 3.8516, + "step": 456 + }, + { + "epoch": 0.5111856823266219, + "grad_norm": 2.7161202430725098, + "learning_rate": 2.412165936467332e-05, + "loss": 4.2025, + "step": 457 + }, + { + "epoch": 0.5123042505592841, + "grad_norm": 2.7404301166534424, + "learning_rate": 2.4033867062876052e-05, + "loss": 3.7109, + "step": 458 + }, + { + "epoch": 0.5134228187919463, + "grad_norm": 3.4687747955322266, + "learning_rate": 2.3946086691647575e-05, + "loss": 3.8982, + "step": 459 + }, + { + "epoch": 0.5145413870246085, + "grad_norm": 2.116055488586426, + "learning_rate": 2.3858319334968893e-05, + "loss": 3.9515, + "step": 460 + }, + { + "epoch": 0.5156599552572707, + "grad_norm": 2.8268518447875977, + "learning_rate": 2.377056607666031e-05, + "loss": 4.4184, + "step": 461 + }, + { + "epoch": 0.5167785234899329, + "grad_norm": 2.4674103260040283, + "learning_rate": 2.368282800036801e-05, + "loss": 4.0095, + "step": 462 + }, + { + "epoch": 0.5178970917225951, + "grad_norm": 2.0366992950439453, + "learning_rate": 2.359510618955073e-05, + "loss": 4.4928, + "step": 463 + }, + { + "epoch": 0.5190156599552572, + "grad_norm": 2.1218111515045166, + "learning_rate": 2.350740172746633e-05, + "loss": 4.149, + "step": 464 + }, + { + "epoch": 0.5201342281879194, + "grad_norm": 3.7365167140960693, + "learning_rate": 2.3419715697158416e-05, + "loss": 3.6304, + "step": 465 + }, + { + "epoch": 0.5212527964205816, + "grad_norm": 2.4339170455932617, + "learning_rate": 2.3332049181443016e-05, + "loss": 3.9618, + "step": 466 + }, + { + "epoch": 0.5223713646532439, + "grad_norm": 2.091418981552124, + "learning_rate": 2.3244403262895153e-05, + "loss": 4.3328, + "step": 467 + }, + { + "epoch": 0.5234899328859061, + "grad_norm": 2.5647706985473633, + "learning_rate": 2.3156779023835525e-05, + "loss": 3.9168, + "step": 468 + }, + { + "epoch": 0.5246085011185683, + "grad_norm": 3.0460050106048584, + "learning_rate": 2.3069177546317087e-05, + "loss": 3.6819, + "step": 469 + }, + { + "epoch": 0.5257270693512305, + "grad_norm": 1.7986963987350464, + "learning_rate": 2.2981599912111736e-05, + "loss": 4.0007, + "step": 470 + }, + { + "epoch": 0.5268456375838926, + "grad_norm": 2.6966090202331543, + "learning_rate": 2.289404720269693e-05, + "loss": 4.2807, + "step": 471 + }, + { + "epoch": 0.5279642058165548, + "grad_norm": 2.5532920360565186, + "learning_rate": 2.280652049924232e-05, + "loss": 4.1665, + "step": 472 + }, + { + "epoch": 0.529082774049217, + "grad_norm": 2.1815285682678223, + "learning_rate": 2.2719020882596427e-05, + "loss": 3.6647, + "step": 473 + }, + { + "epoch": 0.5302013422818792, + "grad_norm": 2.97127103805542, + "learning_rate": 2.2631549433273293e-05, + "loss": 3.8236, + "step": 474 + }, + { + "epoch": 0.5313199105145414, + "grad_norm": 2.294546365737915, + "learning_rate": 2.2544107231439114e-05, + "loss": 3.6576, + "step": 475 + }, + { + "epoch": 0.5324384787472036, + "grad_norm": 2.5872814655303955, + "learning_rate": 2.2456695356898916e-05, + "loss": 4.077, + "step": 476 + }, + { + "epoch": 0.5335570469798657, + "grad_norm": 2.548797369003296, + "learning_rate": 2.2369314889083235e-05, + "loss": 3.7512, + "step": 477 + }, + { + "epoch": 0.5346756152125279, + "grad_norm": 2.8900959491729736, + "learning_rate": 2.2281966907034733e-05, + "loss": 4.3499, + "step": 478 + }, + { + "epoch": 0.5357941834451901, + "grad_norm": 3.7104897499084473, + "learning_rate": 2.2194652489394967e-05, + "loss": 3.5961, + "step": 479 + }, + { + "epoch": 0.5369127516778524, + "grad_norm": 2.2205984592437744, + "learning_rate": 2.2107372714390974e-05, + "loss": 3.6201, + "step": 480 + }, + { + "epoch": 0.5380313199105146, + "grad_norm": 1.851414442062378, + "learning_rate": 2.2020128659822e-05, + "loss": 4.0772, + "step": 481 + }, + { + "epoch": 0.5391498881431768, + "grad_norm": 2.894435167312622, + "learning_rate": 2.1932921403046207e-05, + "loss": 3.6691, + "step": 482 + }, + { + "epoch": 0.540268456375839, + "grad_norm": 2.332059621810913, + "learning_rate": 2.1845752020967337e-05, + "loss": 3.7673, + "step": 483 + }, + { + "epoch": 0.5413870246085011, + "grad_norm": 2.175400972366333, + "learning_rate": 2.1758621590021426e-05, + "loss": 3.8803, + "step": 484 + }, + { + "epoch": 0.5425055928411633, + "grad_norm": 2.4179916381835938, + "learning_rate": 2.1671531186163512e-05, + "loss": 3.6459, + "step": 485 + }, + { + "epoch": 0.5436241610738255, + "grad_norm": 3.060016393661499, + "learning_rate": 2.158448188485433e-05, + "loss": 3.8407, + "step": 486 + }, + { + "epoch": 0.5447427293064877, + "grad_norm": 2.318657875061035, + "learning_rate": 2.1497474761047086e-05, + "loss": 3.7452, + "step": 487 + }, + { + "epoch": 0.5458612975391499, + "grad_norm": 2.2398786544799805, + "learning_rate": 2.141051088917411e-05, + "loss": 4.078, + "step": 488 + }, + { + "epoch": 0.5469798657718121, + "grad_norm": 2.4985172748565674, + "learning_rate": 2.1323591343133646e-05, + "loss": 3.8253, + "step": 489 + }, + { + "epoch": 0.5480984340044742, + "grad_norm": 2.5000696182250977, + "learning_rate": 2.1236717196276558e-05, + "loss": 3.753, + "step": 490 + }, + { + "epoch": 0.5492170022371364, + "grad_norm": 2.586850643157959, + "learning_rate": 2.114988952139307e-05, + "loss": 4.1803, + "step": 491 + }, + { + "epoch": 0.5503355704697986, + "grad_norm": 2.7555367946624756, + "learning_rate": 2.106310939069956e-05, + "loss": 3.8567, + "step": 492 + }, + { + "epoch": 0.5514541387024608, + "grad_norm": 2.633790969848633, + "learning_rate": 2.0976377875825283e-05, + "loss": 4.1461, + "step": 493 + }, + { + "epoch": 0.5525727069351231, + "grad_norm": 3.8133838176727295, + "learning_rate": 2.0889696047799143e-05, + "loss": 4.0689, + "step": 494 + }, + { + "epoch": 0.5536912751677853, + "grad_norm": 1.6977475881576538, + "learning_rate": 2.0803064977036478e-05, + "loss": 3.9074, + "step": 495 + }, + { + "epoch": 0.5548098434004475, + "grad_norm": 3.245305299758911, + "learning_rate": 2.071648573332583e-05, + "loss": 3.7151, + "step": 496 + }, + { + "epoch": 0.5559284116331096, + "grad_norm": 2.754373550415039, + "learning_rate": 2.0629959385815757e-05, + "loss": 4.0098, + "step": 497 + }, + { + "epoch": 0.5570469798657718, + "grad_norm": 2.561891794204712, + "learning_rate": 2.054348700300158e-05, + "loss": 3.8599, + "step": 498 + }, + { + "epoch": 0.558165548098434, + "grad_norm": 2.573894500732422, + "learning_rate": 2.0457069652712242e-05, + "loss": 4.1772, + "step": 499 + }, + { + "epoch": 0.5592841163310962, + "grad_norm": 2.2721850872039795, + "learning_rate": 2.037070840209711e-05, + "loss": 4.2124, + "step": 500 + }, + { + "epoch": 0.5604026845637584, + "grad_norm": 2.0952377319335938, + "learning_rate": 2.0284404317612764e-05, + "loss": 3.9423, + "step": 501 + }, + { + "epoch": 0.5615212527964206, + "grad_norm": 3.6689555644989014, + "learning_rate": 2.019815846500988e-05, + "loss": 3.8274, + "step": 502 + }, + { + "epoch": 0.5626398210290827, + "grad_norm": 2.6612648963928223, + "learning_rate": 2.0111971909320027e-05, + "loss": 3.9185, + "step": 503 + }, + { + "epoch": 0.5637583892617449, + "grad_norm": 2.889272928237915, + "learning_rate": 2.0025845714842516e-05, + "loss": 3.7501, + "step": 504 + }, + { + "epoch": 0.5648769574944071, + "grad_norm": 3.470381736755371, + "learning_rate": 1.993978094513128e-05, + "loss": 3.6945, + "step": 505 + }, + { + "epoch": 0.5659955257270693, + "grad_norm": 2.4324817657470703, + "learning_rate": 1.9853778662981744e-05, + "loss": 4.0761, + "step": 506 + }, + { + "epoch": 0.5671140939597316, + "grad_norm": 2.1613693237304688, + "learning_rate": 1.9767839930417673e-05, + "loss": 4.3073, + "step": 507 + }, + { + "epoch": 0.5682326621923938, + "grad_norm": 2.2693519592285156, + "learning_rate": 1.968196580867808e-05, + "loss": 4.0687, + "step": 508 + }, + { + "epoch": 0.569351230425056, + "grad_norm": 1.9982450008392334, + "learning_rate": 1.9596157358204097e-05, + "loss": 3.6727, + "step": 509 + }, + { + "epoch": 0.5704697986577181, + "grad_norm": 3.018151044845581, + "learning_rate": 1.9510415638625932e-05, + "loss": 3.6253, + "step": 510 + }, + { + "epoch": 0.5715883668903803, + "grad_norm": 1.8852089643478394, + "learning_rate": 1.9424741708749695e-05, + "loss": 4.0768, + "step": 511 + }, + { + "epoch": 0.5727069351230425, + "grad_norm": 2.7679364681243896, + "learning_rate": 1.9339136626544407e-05, + "loss": 3.6646, + "step": 512 + }, + { + "epoch": 0.5738255033557047, + "grad_norm": 2.5092828273773193, + "learning_rate": 1.9253601449128914e-05, + "loss": 4.0424, + "step": 513 + }, + { + "epoch": 0.5749440715883669, + "grad_norm": 1.9775032997131348, + "learning_rate": 1.9168137232758797e-05, + "loss": 3.8255, + "step": 514 + }, + { + "epoch": 0.5760626398210291, + "grad_norm": 2.3674395084381104, + "learning_rate": 1.9082745032813368e-05, + "loss": 3.9264, + "step": 515 + }, + { + "epoch": 0.5771812080536913, + "grad_norm": 1.9700473546981812, + "learning_rate": 1.8997425903782627e-05, + "loss": 3.9236, + "step": 516 + }, + { + "epoch": 0.5782997762863534, + "grad_norm": 2.295315742492676, + "learning_rate": 1.891218089925423e-05, + "loss": 3.773, + "step": 517 + }, + { + "epoch": 0.5794183445190156, + "grad_norm": 1.7736364603042603, + "learning_rate": 1.8827011071900474e-05, + "loss": 3.8288, + "step": 518 + }, + { + "epoch": 0.5805369127516778, + "grad_norm": 2.1893913745880127, + "learning_rate": 1.874191747346534e-05, + "loss": 3.8251, + "step": 519 + }, + { + "epoch": 0.5816554809843401, + "grad_norm": 2.590500593185425, + "learning_rate": 1.865690115475144e-05, + "loss": 4.1849, + "step": 520 + }, + { + "epoch": 0.5827740492170023, + "grad_norm": 2.320377826690674, + "learning_rate": 1.8571963165607104e-05, + "loss": 3.8381, + "step": 521 + }, + { + "epoch": 0.5838926174496645, + "grad_norm": 2.334003210067749, + "learning_rate": 1.8487104554913363e-05, + "loss": 3.9728, + "step": 522 + }, + { + "epoch": 0.5850111856823266, + "grad_norm": 2.939434051513672, + "learning_rate": 1.8402326370571056e-05, + "loss": 3.9855, + "step": 523 + }, + { + "epoch": 0.5861297539149888, + "grad_norm": 2.41349720954895, + "learning_rate": 1.8317629659487796e-05, + "loss": 4.0445, + "step": 524 + }, + { + "epoch": 0.587248322147651, + "grad_norm": 2.5825881958007812, + "learning_rate": 1.8233015467565152e-05, + "loss": 3.852, + "step": 525 + }, + { + "epoch": 0.5883668903803132, + "grad_norm": 2.247124671936035, + "learning_rate": 1.8148484839685662e-05, + "loss": 4.341, + "step": 526 + }, + { + "epoch": 0.5894854586129754, + "grad_norm": 2.6871159076690674, + "learning_rate": 1.806403881969996e-05, + "loss": 3.9645, + "step": 527 + }, + { + "epoch": 0.5906040268456376, + "grad_norm": 3.4999427795410156, + "learning_rate": 1.7979678450413845e-05, + "loss": 4.1412, + "step": 528 + }, + { + "epoch": 0.5917225950782998, + "grad_norm": 3.4024899005889893, + "learning_rate": 1.789540477357548e-05, + "loss": 4.191, + "step": 529 + }, + { + "epoch": 0.5928411633109619, + "grad_norm": 2.1801340579986572, + "learning_rate": 1.781121882986245e-05, + "loss": 3.6334, + "step": 530 + }, + { + "epoch": 0.5939597315436241, + "grad_norm": 2.63584303855896, + "learning_rate": 1.7727121658868934e-05, + "loss": 3.7061, + "step": 531 + }, + { + "epoch": 0.5950782997762863, + "grad_norm": 2.664759635925293, + "learning_rate": 1.764311429909291e-05, + "loss": 4.0002, + "step": 532 + }, + { + "epoch": 0.5961968680089486, + "grad_norm": 3.4378206729888916, + "learning_rate": 1.7559197787923263e-05, + "loss": 3.8005, + "step": 533 + }, + { + "epoch": 0.5973154362416108, + "grad_norm": 2.902148485183716, + "learning_rate": 1.7475373161627034e-05, + "loss": 4.086, + "step": 534 + }, + { + "epoch": 0.598434004474273, + "grad_norm": 1.8565155267715454, + "learning_rate": 1.739164145533658e-05, + "loss": 3.6199, + "step": 535 + }, + { + "epoch": 0.5995525727069351, + "grad_norm": 2.2975833415985107, + "learning_rate": 1.730800370303683e-05, + "loss": 4.0409, + "step": 536 + }, + { + "epoch": 0.6006711409395973, + "grad_norm": 2.3823060989379883, + "learning_rate": 1.7224460937552446e-05, + "loss": 3.9938, + "step": 537 + }, + { + "epoch": 0.6017897091722595, + "grad_norm": 1.8713372945785522, + "learning_rate": 1.714101419053518e-05, + "loss": 3.9404, + "step": 538 + }, + { + "epoch": 0.6029082774049217, + "grad_norm": 2.7200701236724854, + "learning_rate": 1.7057664492451036e-05, + "loss": 3.6198, + "step": 539 + }, + { + "epoch": 0.6040268456375839, + "grad_norm": 2.899989604949951, + "learning_rate": 1.6974412872567597e-05, + "loss": 3.8505, + "step": 540 + }, + { + "epoch": 0.6051454138702461, + "grad_norm": 3.0191915035247803, + "learning_rate": 1.6891260358941276e-05, + "loss": 3.8781, + "step": 541 + }, + { + "epoch": 0.6062639821029083, + "grad_norm": 3.2697198390960693, + "learning_rate": 1.6808207978404683e-05, + "loss": 3.7293, + "step": 542 + }, + { + "epoch": 0.6073825503355704, + "grad_norm": 4.28400182723999, + "learning_rate": 1.672525675655387e-05, + "loss": 4.0706, + "step": 543 + }, + { + "epoch": 0.6085011185682326, + "grad_norm": 2.090829849243164, + "learning_rate": 1.664240771773571e-05, + "loss": 4.0724, + "step": 544 + }, + { + "epoch": 0.6096196868008948, + "grad_norm": 2.9155173301696777, + "learning_rate": 1.655966188503526e-05, + "loss": 3.7549, + "step": 545 + }, + { + "epoch": 0.610738255033557, + "grad_norm": 2.9214680194854736, + "learning_rate": 1.647702028026308e-05, + "loss": 3.9448, + "step": 546 + }, + { + "epoch": 0.6118568232662193, + "grad_norm": 1.5951290130615234, + "learning_rate": 1.6394483923942655e-05, + "loss": 4.1215, + "step": 547 + }, + { + "epoch": 0.6129753914988815, + "grad_norm": 2.0903103351593018, + "learning_rate": 1.6312053835297784e-05, + "loss": 3.8019, + "step": 548 + }, + { + "epoch": 0.6140939597315436, + "grad_norm": 2.552449941635132, + "learning_rate": 1.6229731032239988e-05, + "loss": 4.0168, + "step": 549 + }, + { + "epoch": 0.6152125279642058, + "grad_norm": 2.0024003982543945, + "learning_rate": 1.6147516531355912e-05, + "loss": 3.8326, + "step": 550 + }, + { + "epoch": 0.616331096196868, + "grad_norm": 2.165592908859253, + "learning_rate": 1.606541134789485e-05, + "loss": 3.798, + "step": 551 + }, + { + "epoch": 0.6174496644295302, + "grad_norm": 2.631091356277466, + "learning_rate": 1.5983416495756116e-05, + "loss": 3.8511, + "step": 552 + }, + { + "epoch": 0.6185682326621924, + "grad_norm": 2.507314682006836, + "learning_rate": 1.5901532987476598e-05, + "loss": 4.1623, + "step": 553 + }, + { + "epoch": 0.6196868008948546, + "grad_norm": 2.264253854751587, + "learning_rate": 1.58197618342182e-05, + "loss": 3.8904, + "step": 554 + }, + { + "epoch": 0.6208053691275168, + "grad_norm": 2.0525124073028564, + "learning_rate": 1.573810404575539e-05, + "loss": 3.8115, + "step": 555 + }, + { + "epoch": 0.6219239373601789, + "grad_norm": 2.31563401222229, + "learning_rate": 1.565656063046272e-05, + "loss": 3.9583, + "step": 556 + }, + { + "epoch": 0.6230425055928411, + "grad_norm": 2.9089858531951904, + "learning_rate": 1.5575132595302352e-05, + "loss": 3.9061, + "step": 557 + }, + { + "epoch": 0.6241610738255033, + "grad_norm": 2.3452372550964355, + "learning_rate": 1.549382094581166e-05, + "loss": 3.8342, + "step": 558 + }, + { + "epoch": 0.6252796420581656, + "grad_norm": 2.0573930740356445, + "learning_rate": 1.541262668609081e-05, + "loss": 3.7931, + "step": 559 + }, + { + "epoch": 0.6263982102908278, + "grad_norm": 2.165867328643799, + "learning_rate": 1.5331550818790315e-05, + "loss": 3.8872, + "step": 560 + }, + { + "epoch": 0.62751677852349, + "grad_norm": 2.885354518890381, + "learning_rate": 1.525059434509871e-05, + "loss": 3.9573, + "step": 561 + }, + { + "epoch": 0.6286353467561522, + "grad_norm": 2.424731731414795, + "learning_rate": 1.516975826473015e-05, + "loss": 3.7392, + "step": 562 + }, + { + "epoch": 0.6297539149888143, + "grad_norm": 1.858024001121521, + "learning_rate": 1.5089043575912099e-05, + "loss": 3.8223, + "step": 563 + }, + { + "epoch": 0.6308724832214765, + "grad_norm": 2.57297420501709, + "learning_rate": 1.5008451275372949e-05, + "loss": 3.4912, + "step": 564 + }, + { + "epoch": 0.6319910514541387, + "grad_norm": 2.9374403953552246, + "learning_rate": 1.4927982358329768e-05, + "loss": 4.142, + "step": 565 + }, + { + "epoch": 0.6331096196868009, + "grad_norm": 2.8601393699645996, + "learning_rate": 1.484763781847599e-05, + "loss": 4.0499, + "step": 566 + }, + { + "epoch": 0.6342281879194631, + "grad_norm": 2.611351251602173, + "learning_rate": 1.4767418647969133e-05, + "loss": 4.1721, + "step": 567 + }, + { + "epoch": 0.6353467561521253, + "grad_norm": 2.2778544425964355, + "learning_rate": 1.4687325837418563e-05, + "loss": 4.0221, + "step": 568 + }, + { + "epoch": 0.6364653243847874, + "grad_norm": 2.24289608001709, + "learning_rate": 1.4607360375873246e-05, + "loss": 4.0047, + "step": 569 + }, + { + "epoch": 0.6375838926174496, + "grad_norm": 2.1393930912017822, + "learning_rate": 1.4527523250809545e-05, + "loss": 3.7402, + "step": 570 + }, + { + "epoch": 0.6387024608501118, + "grad_norm": 2.4611830711364746, + "learning_rate": 1.4447815448119017e-05, + "loss": 4.1909, + "step": 571 + }, + { + "epoch": 0.639821029082774, + "grad_norm": 2.261955976486206, + "learning_rate": 1.4368237952096258e-05, + "loss": 3.6671, + "step": 572 + }, + { + "epoch": 0.6409395973154363, + "grad_norm": 2.2543327808380127, + "learning_rate": 1.4288791745426739e-05, + "loss": 3.719, + "step": 573 + }, + { + "epoch": 0.6420581655480985, + "grad_norm": 3.0173962116241455, + "learning_rate": 1.4209477809174634e-05, + "loss": 3.9079, + "step": 574 + }, + { + "epoch": 0.6431767337807607, + "grad_norm": 2.1777987480163574, + "learning_rate": 1.4130297122770774e-05, + "loss": 3.6232, + "step": 575 + }, + { + "epoch": 0.6442953020134228, + "grad_norm": 2.0531842708587646, + "learning_rate": 1.4051250664000515e-05, + "loss": 3.9533, + "step": 576 + }, + { + "epoch": 0.645413870246085, + "grad_norm": 2.7439255714416504, + "learning_rate": 1.3972339408991626e-05, + "loss": 3.8229, + "step": 577 + }, + { + "epoch": 0.6465324384787472, + "grad_norm": 2.413195848464966, + "learning_rate": 1.3893564332202319e-05, + "loss": 4.346, + "step": 578 + }, + { + "epoch": 0.6476510067114094, + "grad_norm": 2.382014274597168, + "learning_rate": 1.381492640640914e-05, + "loss": 3.771, + "step": 579 + }, + { + "epoch": 0.6487695749440716, + "grad_norm": 1.874355673789978, + "learning_rate": 1.3736426602694998e-05, + "loss": 3.7268, + "step": 580 + }, + { + "epoch": 0.6498881431767338, + "grad_norm": 1.958233118057251, + "learning_rate": 1.365806589043718e-05, + "loss": 3.9595, + "step": 581 + }, + { + "epoch": 0.6510067114093959, + "grad_norm": 2.6319327354431152, + "learning_rate": 1.357984523729533e-05, + "loss": 4.2425, + "step": 582 + }, + { + "epoch": 0.6521252796420581, + "grad_norm": 3.006190061569214, + "learning_rate": 1.3501765609199534e-05, + "loss": 3.8662, + "step": 583 + }, + { + "epoch": 0.6532438478747203, + "grad_norm": 2.89542818069458, + "learning_rate": 1.342382797033842e-05, + "loss": 3.8635, + "step": 584 + }, + { + "epoch": 0.6543624161073825, + "grad_norm": 2.149411916732788, + "learning_rate": 1.3346033283147196e-05, + "loss": 3.9787, + "step": 585 + }, + { + "epoch": 0.6554809843400448, + "grad_norm": 2.2859368324279785, + "learning_rate": 1.3268382508295812e-05, + "loss": 3.735, + "step": 586 + }, + { + "epoch": 0.656599552572707, + "grad_norm": 3.07099986076355, + "learning_rate": 1.3190876604677043e-05, + "loss": 3.9014, + "step": 587 + }, + { + "epoch": 0.6577181208053692, + "grad_norm": 2.119368553161621, + "learning_rate": 1.3113516529394704e-05, + "loss": 4.0083, + "step": 588 + }, + { + "epoch": 0.6588366890380313, + "grad_norm": 2.585334300994873, + "learning_rate": 1.3036303237751812e-05, + "loss": 3.7846, + "step": 589 + }, + { + "epoch": 0.6599552572706935, + "grad_norm": 2.7365801334381104, + "learning_rate": 1.2959237683238767e-05, + "loss": 3.6363, + "step": 590 + }, + { + "epoch": 0.6610738255033557, + "grad_norm": 2.1108875274658203, + "learning_rate": 1.2882320817521588e-05, + "loss": 3.8154, + "step": 591 + }, + { + "epoch": 0.6621923937360179, + "grad_norm": 2.750692129135132, + "learning_rate": 1.2805553590430197e-05, + "loss": 3.7864, + "step": 592 + }, + { + "epoch": 0.6633109619686801, + "grad_norm": 1.9360623359680176, + "learning_rate": 1.2728936949946638e-05, + "loss": 3.5591, + "step": 593 + }, + { + "epoch": 0.6644295302013423, + "grad_norm": 2.5163891315460205, + "learning_rate": 1.2652471842193415e-05, + "loss": 3.8985, + "step": 594 + }, + { + "epoch": 0.6655480984340044, + "grad_norm": 2.350154399871826, + "learning_rate": 1.2576159211421763e-05, + "loss": 3.8853, + "step": 595 + }, + { + "epoch": 0.6666666666666666, + "grad_norm": 2.915135622024536, + "learning_rate": 1.2500000000000006e-05, + "loss": 3.6496, + "step": 596 + }, + { + "epoch": 0.6677852348993288, + "grad_norm": 3.425151824951172, + "learning_rate": 1.2423995148401954e-05, + "loss": 4.1841, + "step": 597 + }, + { + "epoch": 0.668903803131991, + "grad_norm": 2.1344897747039795, + "learning_rate": 1.2348145595195245e-05, + "loss": 3.6813, + "step": 598 + }, + { + "epoch": 0.6700223713646533, + "grad_norm": 3.043466806411743, + "learning_rate": 1.2272452277029783e-05, + "loss": 3.717, + "step": 599 + }, + { + "epoch": 0.6711409395973155, + "grad_norm": 2.1800496578216553, + "learning_rate": 1.2196916128626126e-05, + "loss": 3.7066, + "step": 600 + }, + { + "epoch": 0.6722595078299777, + "grad_norm": 2.1941945552825928, + "learning_rate": 1.2121538082764009e-05, + "loss": 3.8536, + "step": 601 + }, + { + "epoch": 0.6733780760626398, + "grad_norm": 1.9257776737213135, + "learning_rate": 1.2046319070270792e-05, + "loss": 3.8676, + "step": 602 + }, + { + "epoch": 0.674496644295302, + "grad_norm": 1.9963043928146362, + "learning_rate": 1.1971260020009944e-05, + "loss": 4.0604, + "step": 603 + }, + { + "epoch": 0.6756152125279642, + "grad_norm": 4.963601112365723, + "learning_rate": 1.1896361858869598e-05, + "loss": 3.8178, + "step": 604 + }, + { + "epoch": 0.6767337807606264, + "grad_norm": 2.1134705543518066, + "learning_rate": 1.1821625511751122e-05, + "loss": 4.3728, + "step": 605 + }, + { + "epoch": 0.6778523489932886, + "grad_norm": 1.895908236503601, + "learning_rate": 1.174705190155766e-05, + "loss": 4.0292, + "step": 606 + }, + { + "epoch": 0.6789709172259508, + "grad_norm": 2.871042490005493, + "learning_rate": 1.1672641949182769e-05, + "loss": 3.6701, + "step": 607 + }, + { + "epoch": 0.680089485458613, + "grad_norm": 3.3493497371673584, + "learning_rate": 1.1598396573499005e-05, + "loss": 4.0769, + "step": 608 + }, + { + "epoch": 0.6812080536912751, + "grad_norm": 2.001706838607788, + "learning_rate": 1.152431669134663e-05, + "loss": 3.9975, + "step": 609 + }, + { + "epoch": 0.6823266219239373, + "grad_norm": 2.3988122940063477, + "learning_rate": 1.145040321752223e-05, + "loss": 4.1463, + "step": 610 + }, + { + "epoch": 0.6834451901565995, + "grad_norm": 2.896697759628296, + "learning_rate": 1.1376657064767485e-05, + "loss": 3.6005, + "step": 611 + }, + { + "epoch": 0.6845637583892618, + "grad_norm": 2.0190067291259766, + "learning_rate": 1.130307914375783e-05, + "loss": 3.8035, + "step": 612 + }, + { + "epoch": 0.685682326621924, + "grad_norm": 1.734345555305481, + "learning_rate": 1.122967036309127e-05, + "loss": 3.9921, + "step": 613 + }, + { + "epoch": 0.6868008948545862, + "grad_norm": 2.797583818435669, + "learning_rate": 1.1156431629277118e-05, + "loss": 3.7049, + "step": 614 + }, + { + "epoch": 0.6879194630872483, + "grad_norm": 2.020956039428711, + "learning_rate": 1.1083363846724822e-05, + "loss": 3.9454, + "step": 615 + }, + { + "epoch": 0.6890380313199105, + "grad_norm": 2.833671808242798, + "learning_rate": 1.1010467917732784e-05, + "loss": 3.9514, + "step": 616 + }, + { + "epoch": 0.6901565995525727, + "grad_norm": 1.816365122795105, + "learning_rate": 1.093774474247721e-05, + "loss": 4.1637, + "step": 617 + }, + { + "epoch": 0.6912751677852349, + "grad_norm": 2.2042744159698486, + "learning_rate": 1.086519521900103e-05, + "loss": 3.8951, + "step": 618 + }, + { + "epoch": 0.6923937360178971, + "grad_norm": 3.488060712814331, + "learning_rate": 1.0792820243202772e-05, + "loss": 3.942, + "step": 619 + }, + { + "epoch": 0.6935123042505593, + "grad_norm": 2.248445510864258, + "learning_rate": 1.0720620708825536e-05, + "loss": 3.8709, + "step": 620 + }, + { + "epoch": 0.6946308724832215, + "grad_norm": 2.1262619495391846, + "learning_rate": 1.0648597507445884e-05, + "loss": 3.9484, + "step": 621 + }, + { + "epoch": 0.6957494407158836, + "grad_norm": 2.552934169769287, + "learning_rate": 1.0576751528462935e-05, + "loss": 3.508, + "step": 622 + }, + { + "epoch": 0.6968680089485458, + "grad_norm": 2.420132875442505, + "learning_rate": 1.0505083659087284e-05, + "loss": 3.6366, + "step": 623 + }, + { + "epoch": 0.697986577181208, + "grad_norm": 2.1934938430786133, + "learning_rate": 1.043359478433012e-05, + "loss": 4.1458, + "step": 624 + }, + { + "epoch": 0.6991051454138703, + "grad_norm": 2.8820436000823975, + "learning_rate": 1.0362285786992231e-05, + "loss": 4.145, + "step": 625 + }, + { + "epoch": 0.7002237136465325, + "grad_norm": 2.036956548690796, + "learning_rate": 1.0291157547653172e-05, + "loss": 4.0621, + "step": 626 + }, + { + "epoch": 0.7013422818791947, + "grad_norm": 2.075648784637451, + "learning_rate": 1.0220210944660338e-05, + "loss": 4.1723, + "step": 627 + }, + { + "epoch": 0.7024608501118568, + "grad_norm": 2.007321357727051, + "learning_rate": 1.0149446854118153e-05, + "loss": 3.9694, + "step": 628 + }, + { + "epoch": 0.703579418344519, + "grad_norm": 2.436429977416992, + "learning_rate": 1.0078866149877206e-05, + "loss": 4.1056, + "step": 629 + }, + { + "epoch": 0.7046979865771812, + "grad_norm": 1.8887969255447388, + "learning_rate": 1.0008469703523493e-05, + "loss": 3.848, + "step": 630 + }, + { + "epoch": 0.7058165548098434, + "grad_norm": 2.362123489379883, + "learning_rate": 9.93825838436767e-06, + "loss": 4.1663, + "step": 631 + }, + { + "epoch": 0.7069351230425056, + "grad_norm": 2.105339765548706, + "learning_rate": 9.868233059434288e-06, + "loss": 3.6117, + "step": 632 + }, + { + "epoch": 0.7080536912751678, + "grad_norm": 2.77400541305542, + "learning_rate": 9.798394593451091e-06, + "loss": 4.035, + "step": 633 + }, + { + "epoch": 0.70917225950783, + "grad_norm": 3.4026005268096924, + "learning_rate": 9.728743848838333e-06, + "loss": 3.9344, + "step": 634 + }, + { + "epoch": 0.7102908277404921, + "grad_norm": 3.4259018898010254, + "learning_rate": 9.659281685698162e-06, + "loss": 4.2997, + "step": 635 + }, + { + "epoch": 0.7114093959731543, + "grad_norm": 2.261091470718384, + "learning_rate": 9.590008961803943e-06, + "loss": 3.7143, + "step": 636 + }, + { + "epoch": 0.7125279642058165, + "grad_norm": 3.5165352821350098, + "learning_rate": 9.520926532589725e-06, + "loss": 3.8508, + "step": 637 + }, + { + "epoch": 0.7136465324384788, + "grad_norm": 2.4055497646331787, + "learning_rate": 9.45203525113962e-06, + "loss": 3.6208, + "step": 638 + }, + { + "epoch": 0.714765100671141, + "grad_norm": 3.532019853591919, + "learning_rate": 9.383335968177323e-06, + "loss": 3.8882, + "step": 639 + }, + { + "epoch": 0.7158836689038032, + "grad_norm": 2.0139145851135254, + "learning_rate": 9.314829532055569e-06, + "loss": 3.9722, + "step": 640 + }, + { + "epoch": 0.7170022371364653, + "grad_norm": 3.198631525039673, + "learning_rate": 9.24651678874568e-06, + "loss": 3.7721, + "step": 641 + }, + { + "epoch": 0.7181208053691275, + "grad_norm": 2.2410922050476074, + "learning_rate": 9.178398581827085e-06, + "loss": 3.978, + "step": 642 + }, + { + "epoch": 0.7192393736017897, + "grad_norm": 2.0665810108184814, + "learning_rate": 9.110475752476935e-06, + "loss": 4.0286, + "step": 643 + }, + { + "epoch": 0.7203579418344519, + "grad_norm": 2.212315559387207, + "learning_rate": 9.04274913945971e-06, + "loss": 3.92, + "step": 644 + }, + { + "epoch": 0.7214765100671141, + "grad_norm": 2.330747127532959, + "learning_rate": 8.975219579116863e-06, + "loss": 4.0911, + "step": 645 + }, + { + "epoch": 0.7225950782997763, + "grad_norm": 2.7871272563934326, + "learning_rate": 8.90788790535646e-06, + "loss": 3.8816, + "step": 646 + }, + { + "epoch": 0.7237136465324385, + "grad_norm": 2.5167412757873535, + "learning_rate": 8.840754949642935e-06, + "loss": 3.8268, + "step": 647 + }, + { + "epoch": 0.7248322147651006, + "grad_norm": 1.9781793355941772, + "learning_rate": 8.77382154098679e-06, + "loss": 3.7334, + "step": 648 + }, + { + "epoch": 0.7259507829977628, + "grad_norm": 2.6232032775878906, + "learning_rate": 8.70708850593434e-06, + "loss": 3.7084, + "step": 649 + }, + { + "epoch": 0.727069351230425, + "grad_norm": 2.177091121673584, + "learning_rate": 8.64055666855757e-06, + "loss": 3.9966, + "step": 650 + }, + { + "epoch": 0.7281879194630873, + "grad_norm": 2.444685220718384, + "learning_rate": 8.574226850443872e-06, + "loss": 3.7267, + "step": 651 + }, + { + "epoch": 0.7293064876957495, + "grad_norm": 3.574202299118042, + "learning_rate": 8.50809987068598e-06, + "loss": 3.8466, + "step": 652 + }, + { + "epoch": 0.7304250559284117, + "grad_norm": 3.1503660678863525, + "learning_rate": 8.442176545871805e-06, + "loss": 3.7964, + "step": 653 + }, + { + "epoch": 0.7315436241610739, + "grad_norm": 1.677579641342163, + "learning_rate": 8.376457690074385e-06, + "loss": 3.8638, + "step": 654 + }, + { + "epoch": 0.732662192393736, + "grad_norm": 2.3635876178741455, + "learning_rate": 8.310944114841786e-06, + "loss": 3.8608, + "step": 655 + }, + { + "epoch": 0.7337807606263982, + "grad_norm": 3.185817241668701, + "learning_rate": 8.24563662918712e-06, + "loss": 4.1601, + "step": 656 + }, + { + "epoch": 0.7348993288590604, + "grad_norm": 3.7147934436798096, + "learning_rate": 8.180536039578545e-06, + "loss": 3.9529, + "step": 657 + }, + { + "epoch": 0.7360178970917226, + "grad_norm": 3.5860955715179443, + "learning_rate": 8.115643149929318e-06, + "loss": 4.2173, + "step": 658 + }, + { + "epoch": 0.7371364653243848, + "grad_norm": 2.398571252822876, + "learning_rate": 8.05095876158782e-06, + "loss": 3.7547, + "step": 659 + }, + { + "epoch": 0.738255033557047, + "grad_norm": 2.6575210094451904, + "learning_rate": 7.986483673327724e-06, + "loss": 3.9889, + "step": 660 + }, + { + "epoch": 0.7393736017897091, + "grad_norm": 2.9067556858062744, + "learning_rate": 7.922218681338095e-06, + "loss": 3.6771, + "step": 661 + }, + { + "epoch": 0.7404921700223713, + "grad_norm": 2.622535467147827, + "learning_rate": 7.858164579213547e-06, + "loss": 3.8512, + "step": 662 + }, + { + "epoch": 0.7416107382550335, + "grad_norm": 2.1957478523254395, + "learning_rate": 7.79432215794449e-06, + "loss": 3.9811, + "step": 663 + }, + { + "epoch": 0.7427293064876958, + "grad_norm": 2.333820343017578, + "learning_rate": 7.730692205907294e-06, + "loss": 3.5702, + "step": 664 + }, + { + "epoch": 0.743847874720358, + "grad_norm": 2.3756656646728516, + "learning_rate": 7.667275508854627e-06, + "loss": 4.0504, + "step": 665 + }, + { + "epoch": 0.7449664429530202, + "grad_norm": 2.2094883918762207, + "learning_rate": 7.604072849905708e-06, + "loss": 3.773, + "step": 666 + }, + { + "epoch": 0.7460850111856824, + "grad_norm": 2.3588247299194336, + "learning_rate": 7.541085009536625e-06, + "loss": 3.7015, + "step": 667 + }, + { + "epoch": 0.7472035794183445, + "grad_norm": 2.534830093383789, + "learning_rate": 7.4783127655707465e-06, + "loss": 3.8194, + "step": 668 + }, + { + "epoch": 0.7483221476510067, + "grad_norm": 2.694711208343506, + "learning_rate": 7.415756893169062e-06, + "loss": 3.9235, + "step": 669 + }, + { + "epoch": 0.7494407158836689, + "grad_norm": 3.596769332885742, + "learning_rate": 7.3534181648206555e-06, + "loss": 3.8338, + "step": 670 + }, + { + "epoch": 0.7505592841163311, + "grad_norm": 3.1873180866241455, + "learning_rate": 7.291297350333137e-06, + "loss": 3.6435, + "step": 671 + }, + { + "epoch": 0.7516778523489933, + "grad_norm": 1.9209433794021606, + "learning_rate": 7.2293952168231316e-06, + "loss": 3.7388, + "step": 672 + }, + { + "epoch": 0.7527964205816555, + "grad_norm": 2.5653648376464844, + "learning_rate": 7.167712528706844e-06, + "loss": 3.8961, + "step": 673 + }, + { + "epoch": 0.7539149888143176, + "grad_norm": 1.9980870485305786, + "learning_rate": 7.106250047690588e-06, + "loss": 3.9753, + "step": 674 + }, + { + "epoch": 0.7550335570469798, + "grad_norm": 2.534268379211426, + "learning_rate": 7.045008532761366e-06, + "loss": 3.7062, + "step": 675 + }, + { + "epoch": 0.756152125279642, + "grad_norm": 1.909271001815796, + "learning_rate": 6.983988740177552e-06, + "loss": 3.9121, + "step": 676 + }, + { + "epoch": 0.7572706935123042, + "grad_norm": 1.9393848180770874, + "learning_rate": 6.923191423459482e-06, + "loss": 3.9378, + "step": 677 + }, + { + "epoch": 0.7583892617449665, + "grad_norm": 3.104290723800659, + "learning_rate": 6.862617333380214e-06, + "loss": 3.6173, + "step": 678 + }, + { + "epoch": 0.7595078299776287, + "grad_norm": 1.9617093801498413, + "learning_rate": 6.802267217956221e-06, + "loss": 3.8215, + "step": 679 + }, + { + "epoch": 0.7606263982102909, + "grad_norm": 2.2684578895568848, + "learning_rate": 6.742141822438144e-06, + "loss": 3.8073, + "step": 680 + }, + { + "epoch": 0.761744966442953, + "grad_norm": 4.458706855773926, + "learning_rate": 6.682241889301635e-06, + "loss": 3.5522, + "step": 681 + }, + { + "epoch": 0.7628635346756152, + "grad_norm": 1.9600862264633179, + "learning_rate": 6.622568158238126e-06, + "loss": 3.7883, + "step": 682 + }, + { + "epoch": 0.7639821029082774, + "grad_norm": 2.454009771347046, + "learning_rate": 6.563121366145758e-06, + "loss": 3.723, + "step": 683 + }, + { + "epoch": 0.7651006711409396, + "grad_norm": 1.8544883728027344, + "learning_rate": 6.503902247120239e-06, + "loss": 3.894, + "step": 684 + }, + { + "epoch": 0.7662192393736018, + "grad_norm": 2.5093746185302734, + "learning_rate": 6.444911532445782e-06, + "loss": 3.9987, + "step": 685 + }, + { + "epoch": 0.767337807606264, + "grad_norm": 3.8050105571746826, + "learning_rate": 6.386149950586098e-06, + "loss": 3.7126, + "step": 686 + }, + { + "epoch": 0.7684563758389261, + "grad_norm": 2.824566125869751, + "learning_rate": 6.327618227175389e-06, + "loss": 3.7332, + "step": 687 + }, + { + "epoch": 0.7695749440715883, + "grad_norm": 3.873337984085083, + "learning_rate": 6.269317085009363e-06, + "loss": 3.6191, + "step": 688 + }, + { + "epoch": 0.7706935123042505, + "grad_norm": 2.275029182434082, + "learning_rate": 6.211247244036339e-06, + "loss": 3.6649, + "step": 689 + }, + { + "epoch": 0.7718120805369127, + "grad_norm": 3.252362012863159, + "learning_rate": 6.153409421348358e-06, + "loss": 3.8387, + "step": 690 + }, + { + "epoch": 0.772930648769575, + "grad_norm": 5.253241539001465, + "learning_rate": 6.095804331172308e-06, + "loss": 3.503, + "step": 691 + }, + { + "epoch": 0.7740492170022372, + "grad_norm": 1.9518181085586548, + "learning_rate": 6.0384326848611225e-06, + "loss": 4.2226, + "step": 692 + }, + { + "epoch": 0.7751677852348994, + "grad_norm": 2.372791051864624, + "learning_rate": 5.981295190884961e-06, + "loss": 4.0139, + "step": 693 + }, + { + "epoch": 0.7762863534675615, + "grad_norm": 2.35370135307312, + "learning_rate": 5.924392554822519e-06, + "loss": 3.7916, + "step": 694 + }, + { + "epoch": 0.7774049217002237, + "grad_norm": 2.5694122314453125, + "learning_rate": 5.867725479352251e-06, + "loss": 3.9369, + "step": 695 + }, + { + "epoch": 0.7785234899328859, + "grad_norm": 2.143845558166504, + "learning_rate": 5.811294664243752e-06, + "loss": 4.0455, + "step": 696 + }, + { + "epoch": 0.7796420581655481, + "grad_norm": 1.9714043140411377, + "learning_rate": 5.755100806349076e-06, + "loss": 3.9658, + "step": 697 + }, + { + "epoch": 0.7807606263982103, + "grad_norm": 3.7511520385742188, + "learning_rate": 5.699144599594131e-06, + "loss": 3.4662, + "step": 698 + }, + { + "epoch": 0.7818791946308725, + "grad_norm": 2.6238203048706055, + "learning_rate": 5.643426734970139e-06, + "loss": 3.7648, + "step": 699 + }, + { + "epoch": 0.7829977628635347, + "grad_norm": 2.54085111618042, + "learning_rate": 5.587947900525093e-06, + "loss": 3.9747, + "step": 700 + }, + { + "epoch": 0.7841163310961968, + "grad_norm": 2.562391519546509, + "learning_rate": 5.532708781355231e-06, + "loss": 3.7304, + "step": 701 + }, + { + "epoch": 0.785234899328859, + "grad_norm": 1.9780319929122925, + "learning_rate": 5.4777100595965994e-06, + "loss": 3.7713, + "step": 702 + }, + { + "epoch": 0.7863534675615212, + "grad_norm": 2.6395468711853027, + "learning_rate": 5.422952414416649e-06, + "loss": 3.6651, + "step": 703 + }, + { + "epoch": 0.7874720357941835, + "grad_norm": 2.2408347129821777, + "learning_rate": 5.368436522005815e-06, + "loss": 3.8067, + "step": 704 + }, + { + "epoch": 0.7885906040268457, + "grad_norm": 2.3998045921325684, + "learning_rate": 5.314163055569188e-06, + "loss": 3.9425, + "step": 705 + }, + { + "epoch": 0.7897091722595079, + "grad_norm": 1.973858118057251, + "learning_rate": 5.26013268531817e-06, + "loss": 4.0397, + "step": 706 + }, + { + "epoch": 0.79082774049217, + "grad_norm": 2.511659860610962, + "learning_rate": 5.206346078462249e-06, + "loss": 3.661, + "step": 707 + }, + { + "epoch": 0.7919463087248322, + "grad_norm": 2.8732688426971436, + "learning_rate": 5.1528038992007e-06, + "loss": 3.8967, + "step": 708 + }, + { + "epoch": 0.7930648769574944, + "grad_norm": 2.265322685241699, + "learning_rate": 5.099506808714452e-06, + "loss": 3.8421, + "step": 709 + }, + { + "epoch": 0.7941834451901566, + "grad_norm": 3.5834262371063232, + "learning_rate": 5.046455465157848e-06, + "loss": 4.1527, + "step": 710 + }, + { + "epoch": 0.7953020134228188, + "grad_norm": 3.056978225708008, + "learning_rate": 4.993650523650575e-06, + "loss": 3.4209, + "step": 711 + }, + { + "epoch": 0.796420581655481, + "grad_norm": 2.173541784286499, + "learning_rate": 4.941092636269554e-06, + "loss": 3.6813, + "step": 712 + }, + { + "epoch": 0.7975391498881432, + "grad_norm": 2.102139949798584, + "learning_rate": 4.888782452040885e-06, + "loss": 3.7805, + "step": 713 + }, + { + "epoch": 0.7986577181208053, + "grad_norm": 2.5020039081573486, + "learning_rate": 4.836720616931831e-06, + "loss": 3.7669, + "step": 714 + }, + { + "epoch": 0.7997762863534675, + "grad_norm": 2.2521355152130127, + "learning_rate": 4.784907773842839e-06, + "loss": 4.0918, + "step": 715 + }, + { + "epoch": 0.8008948545861297, + "grad_norm": 2.366150379180908, + "learning_rate": 4.733344562599623e-06, + "loss": 4.0847, + "step": 716 + }, + { + "epoch": 0.802013422818792, + "grad_norm": 3.198983907699585, + "learning_rate": 4.682031619945238e-06, + "loss": 3.5284, + "step": 717 + }, + { + "epoch": 0.8031319910514542, + "grad_norm": 1.931961178779602, + "learning_rate": 4.630969579532232e-06, + "loss": 3.8843, + "step": 718 + }, + { + "epoch": 0.8042505592841164, + "grad_norm": 2.4697344303131104, + "learning_rate": 4.580159071914794e-06, + "loss": 3.8388, + "step": 719 + }, + { + "epoch": 0.8053691275167785, + "grad_norm": 2.202491283416748, + "learning_rate": 4.529600724541022e-06, + "loss": 3.8412, + "step": 720 + }, + { + "epoch": 0.8064876957494407, + "grad_norm": 3.0058343410491943, + "learning_rate": 4.479295161745109e-06, + "loss": 4.1481, + "step": 721 + }, + { + "epoch": 0.8076062639821029, + "grad_norm": 2.1180286407470703, + "learning_rate": 4.4292430047396914e-06, + "loss": 3.8576, + "step": 722 + }, + { + "epoch": 0.8087248322147651, + "grad_norm": 2.6063430309295654, + "learning_rate": 4.379444871608124e-06, + "loss": 3.7986, + "step": 723 + }, + { + "epoch": 0.8098434004474273, + "grad_norm": 2.227067470550537, + "learning_rate": 4.329901377296902e-06, + "loss": 3.5418, + "step": 724 + }, + { + "epoch": 0.8109619686800895, + "grad_norm": 2.3542473316192627, + "learning_rate": 4.280613133608028e-06, + "loss": 3.7732, + "step": 725 + }, + { + "epoch": 0.8120805369127517, + "grad_norm": 3.143573045730591, + "learning_rate": 4.231580749191474e-06, + "loss": 4.1625, + "step": 726 + }, + { + "epoch": 0.8131991051454138, + "grad_norm": 2.3651700019836426, + "learning_rate": 4.182804829537654e-06, + "loss": 4.3148, + "step": 727 + }, + { + "epoch": 0.814317673378076, + "grad_norm": 2.07985258102417, + "learning_rate": 4.134285976969948e-06, + "loss": 3.8726, + "step": 728 + }, + { + "epoch": 0.8154362416107382, + "grad_norm": 3.1206445693969727, + "learning_rate": 4.086024790637285e-06, + "loss": 3.6906, + "step": 729 + }, + { + "epoch": 0.8165548098434005, + "grad_norm": 2.517094850540161, + "learning_rate": 4.038021866506725e-06, + "loss": 4.0283, + "step": 730 + }, + { + "epoch": 0.8176733780760627, + "grad_norm": 2.244453191757202, + "learning_rate": 3.990277797356104e-06, + "loss": 3.959, + "step": 731 + }, + { + "epoch": 0.8187919463087249, + "grad_norm": 1.9631199836730957, + "learning_rate": 3.942793172766699e-06, + "loss": 4.2073, + "step": 732 + }, + { + "epoch": 0.819910514541387, + "grad_norm": 2.9783220291137695, + "learning_rate": 3.895568579115983e-06, + "loss": 3.7062, + "step": 733 + }, + { + "epoch": 0.8210290827740492, + "grad_norm": 2.1372721195220947, + "learning_rate": 3.848604599570338e-06, + "loss": 3.7815, + "step": 734 + }, + { + "epoch": 0.8221476510067114, + "grad_norm": 2.8807451725006104, + "learning_rate": 3.80190181407791e-06, + "loss": 3.639, + "step": 735 + }, + { + "epoch": 0.8232662192393736, + "grad_norm": 1.9709060192108154, + "learning_rate": 3.755460799361382e-06, + "loss": 3.8587, + "step": 736 + }, + { + "epoch": 0.8243847874720358, + "grad_norm": 2.44718861579895, + "learning_rate": 3.709282128910907e-06, + "loss": 3.9482, + "step": 737 + }, + { + "epoch": 0.825503355704698, + "grad_norm": 2.036914110183716, + "learning_rate": 3.6633663729770008e-06, + "loss": 3.9106, + "step": 738 + }, + { + "epoch": 0.8266219239373602, + "grad_norm": 2.3819079399108887, + "learning_rate": 3.61771409856351e-06, + "loss": 4.0078, + "step": 739 + }, + { + "epoch": 0.8277404921700223, + "grad_norm": 2.206578493118286, + "learning_rate": 3.572325869420587e-06, + "loss": 3.839, + "step": 740 + }, + { + "epoch": 0.8288590604026845, + "grad_norm": 2.039262294769287, + "learning_rate": 3.527202246037756e-06, + "loss": 3.8825, + "step": 741 + }, + { + "epoch": 0.8299776286353467, + "grad_norm": 2.297976493835449, + "learning_rate": 3.4823437856369794e-06, + "loss": 3.9343, + "step": 742 + }, + { + "epoch": 0.831096196868009, + "grad_norm": 2.6157867908477783, + "learning_rate": 3.4377510421657906e-06, + "loss": 3.7783, + "step": 743 + }, + { + "epoch": 0.8322147651006712, + "grad_norm": 2.9260687828063965, + "learning_rate": 3.393424566290421e-06, + "loss": 3.8069, + "step": 744 + }, + { + "epoch": 0.8333333333333334, + "grad_norm": 4.025168418884277, + "learning_rate": 3.3493649053890326e-06, + "loss": 3.743, + "step": 745 + }, + { + "epoch": 0.8344519015659956, + "grad_norm": 2.532195806503296, + "learning_rate": 3.3055726035449484e-06, + "loss": 3.6006, + "step": 746 + }, + { + "epoch": 0.8355704697986577, + "grad_norm": 3.2442009449005127, + "learning_rate": 3.2620482015399302e-06, + "loss": 3.7358, + "step": 747 + }, + { + "epoch": 0.8366890380313199, + "grad_norm": 4.921333312988281, + "learning_rate": 3.2187922368474952e-06, + "loss": 3.8525, + "step": 748 + }, + { + "epoch": 0.8378076062639821, + "grad_norm": 2.6702213287353516, + "learning_rate": 3.175805243626284e-06, + "loss": 3.8665, + "step": 749 + }, + { + "epoch": 0.8389261744966443, + "grad_norm": 4.156053066253662, + "learning_rate": 3.133087752713479e-06, + "loss": 4.2799, + "step": 750 + }, + { + "epoch": 0.8400447427293065, + "grad_norm": 2.0885608196258545, + "learning_rate": 3.0906402916182297e-06, + "loss": 3.8346, + "step": 751 + }, + { + "epoch": 0.8411633109619687, + "grad_norm": 2.58349871635437, + "learning_rate": 3.0484633845151488e-06, + "loss": 3.7313, + "step": 752 + }, + { + "epoch": 0.8422818791946308, + "grad_norm": 3.2410836219787598, + "learning_rate": 3.0065575522378264e-06, + "loss": 3.7262, + "step": 753 + }, + { + "epoch": 0.843400447427293, + "grad_norm": 1.6323789358139038, + "learning_rate": 2.9649233122724105e-06, + "loss": 3.9603, + "step": 754 + }, + { + "epoch": 0.8445190156599552, + "grad_norm": 1.915100336074829, + "learning_rate": 2.923561178751219e-06, + "loss": 4.0198, + "step": 755 + }, + { + "epoch": 0.8456375838926175, + "grad_norm": 2.1250874996185303, + "learning_rate": 2.88247166244639e-06, + "loss": 3.9345, + "step": 756 + }, + { + "epoch": 0.8467561521252797, + "grad_norm": 3.2116942405700684, + "learning_rate": 2.841655270763549e-06, + "loss": 3.6633, + "step": 757 + }, + { + "epoch": 0.8478747203579419, + "grad_norm": 2.246950626373291, + "learning_rate": 2.801112507735587e-06, + "loss": 3.9842, + "step": 758 + }, + { + "epoch": 0.8489932885906041, + "grad_norm": 1.9269198179244995, + "learning_rate": 2.760843874016403e-06, + "loss": 4.043, + "step": 759 + }, + { + "epoch": 0.8501118568232662, + "grad_norm": 3.6027183532714844, + "learning_rate": 2.720849866874739e-06, + "loss": 3.693, + "step": 760 + }, + { + "epoch": 0.8512304250559284, + "grad_norm": 2.76350474357605, + "learning_rate": 2.6811309801880208e-06, + "loss": 4.1343, + "step": 761 + }, + { + "epoch": 0.8523489932885906, + "grad_norm": 2.693434238433838, + "learning_rate": 2.6416877044362685e-06, + "loss": 4.3856, + "step": 762 + }, + { + "epoch": 0.8534675615212528, + "grad_norm": 2.068485736846924, + "learning_rate": 2.602520526696059e-06, + "loss": 4.0551, + "step": 763 + }, + { + "epoch": 0.854586129753915, + "grad_norm": 2.524249792098999, + "learning_rate": 2.563629930634487e-06, + "loss": 3.776, + "step": 764 + }, + { + "epoch": 0.8557046979865772, + "grad_norm": 2.40083909034729, + "learning_rate": 2.525016396503185e-06, + "loss": 3.8733, + "step": 765 + }, + { + "epoch": 0.8568232662192393, + "grad_norm": 2.9691314697265625, + "learning_rate": 2.4866804011324296e-06, + "loss": 3.7485, + "step": 766 + }, + { + "epoch": 0.8579418344519015, + "grad_norm": 3.3835642337799072, + "learning_rate": 2.448622417925214e-06, + "loss": 3.7207, + "step": 767 + }, + { + "epoch": 0.8590604026845637, + "grad_norm": 2.4943041801452637, + "learning_rate": 2.4108429168514245e-06, + "loss": 3.4865, + "step": 768 + }, + { + "epoch": 0.860178970917226, + "grad_norm": 2.2213940620422363, + "learning_rate": 2.3733423644420353e-06, + "loss": 3.9459, + "step": 769 + }, + { + "epoch": 0.8612975391498882, + "grad_norm": 2.157849073410034, + "learning_rate": 2.3361212237833356e-06, + "loss": 3.8979, + "step": 770 + }, + { + "epoch": 0.8624161073825504, + "grad_norm": 2.031329870223999, + "learning_rate": 2.2991799545112215e-06, + "loss": 3.9194, + "step": 771 + }, + { + "epoch": 0.8635346756152126, + "grad_norm": 3.01249098777771, + "learning_rate": 2.2625190128055168e-06, + "loss": 3.7412, + "step": 772 + }, + { + "epoch": 0.8646532438478747, + "grad_norm": 2.1558024883270264, + "learning_rate": 2.2261388513843515e-06, + "loss": 4.12, + "step": 773 + }, + { + "epoch": 0.8657718120805369, + "grad_norm": 3.150437593460083, + "learning_rate": 2.190039919498543e-06, + "loss": 3.8121, + "step": 774 + }, + { + "epoch": 0.8668903803131991, + "grad_norm": 2.319565534591675, + "learning_rate": 2.154222662926067e-06, + "loss": 3.8679, + "step": 775 + }, + { + "epoch": 0.8680089485458613, + "grad_norm": 2.686678171157837, + "learning_rate": 2.118687523966559e-06, + "loss": 4.2429, + "step": 776 + }, + { + "epoch": 0.8691275167785235, + "grad_norm": 2.2821829319000244, + "learning_rate": 2.0834349414358495e-06, + "loss": 3.9263, + "step": 777 + }, + { + "epoch": 0.8702460850111857, + "grad_norm": 2.5908377170562744, + "learning_rate": 2.048465350660522e-06, + "loss": 3.4955, + "step": 778 + }, + { + "epoch": 0.8713646532438478, + "grad_norm": 3.3101370334625244, + "learning_rate": 2.013779183472575e-06, + "loss": 4.0192, + "step": 779 + }, + { + "epoch": 0.87248322147651, + "grad_norm": 2.681217908859253, + "learning_rate": 1.9793768682040524e-06, + "loss": 3.6952, + "step": 780 + }, + { + "epoch": 0.8736017897091722, + "grad_norm": 2.5683112144470215, + "learning_rate": 1.9452588296817843e-06, + "loss": 3.9709, + "step": 781 + }, + { + "epoch": 0.8747203579418344, + "grad_norm": 3.1038596630096436, + "learning_rate": 1.911425489222127e-06, + "loss": 4.0497, + "step": 782 + }, + { + "epoch": 0.8758389261744967, + "grad_norm": 2.36666202545166, + "learning_rate": 1.8778772646257491e-06, + "loss": 4.0068, + "step": 783 + }, + { + "epoch": 0.8769574944071589, + "grad_norm": 2.2028393745422363, + "learning_rate": 1.8446145701724983e-06, + "loss": 3.6578, + "step": 784 + }, + { + "epoch": 0.8780760626398211, + "grad_norm": 4.168335437774658, + "learning_rate": 1.8116378166162685e-06, + "loss": 4.1479, + "step": 785 + }, + { + "epoch": 0.8791946308724832, + "grad_norm": 2.6044557094573975, + "learning_rate": 1.778947411179932e-06, + "loss": 3.573, + "step": 786 + }, + { + "epoch": 0.8803131991051454, + "grad_norm": 1.6643822193145752, + "learning_rate": 1.7465437575502952e-06, + "loss": 3.8277, + "step": 787 + }, + { + "epoch": 0.8814317673378076, + "grad_norm": 2.9622278213500977, + "learning_rate": 1.7144272558731467e-06, + "loss": 3.6871, + "step": 788 + }, + { + "epoch": 0.8825503355704698, + "grad_norm": 2.0143864154815674, + "learning_rate": 1.6825983027482867e-06, + "loss": 3.9285, + "step": 789 + }, + { + "epoch": 0.883668903803132, + "grad_norm": 2.577881097793579, + "learning_rate": 1.6510572912246475e-06, + "loss": 3.6799, + "step": 790 + }, + { + "epoch": 0.8847874720357942, + "grad_norm": 2.5292065143585205, + "learning_rate": 1.6198046107954223e-06, + "loss": 3.9658, + "step": 791 + }, + { + "epoch": 0.8859060402684564, + "grad_norm": 2.8256359100341797, + "learning_rate": 1.5888406473932692e-06, + "loss": 3.7874, + "step": 792 + }, + { + "epoch": 0.8870246085011185, + "grad_norm": 4.170554161071777, + "learning_rate": 1.5581657833855512e-06, + "loss": 3.6907, + "step": 793 + }, + { + "epoch": 0.8881431767337807, + "grad_norm": 1.9856733083724976, + "learning_rate": 1.5277803975695864e-06, + "loss": 3.8945, + "step": 794 + }, + { + "epoch": 0.889261744966443, + "grad_norm": 2.3850162029266357, + "learning_rate": 1.497684865168006e-06, + "loss": 3.7603, + "step": 795 + }, + { + "epoch": 0.8903803131991052, + "grad_norm": 2.514587879180908, + "learning_rate": 1.4678795578240894e-06, + "loss": 3.9449, + "step": 796 + }, + { + "epoch": 0.8914988814317674, + "grad_norm": 3.2767693996429443, + "learning_rate": 1.4383648435972007e-06, + "loss": 3.9914, + "step": 797 + }, + { + "epoch": 0.8926174496644296, + "grad_norm": 2.3066697120666504, + "learning_rate": 1.4091410869582267e-06, + "loss": 3.7617, + "step": 798 + }, + { + "epoch": 0.8937360178970917, + "grad_norm": 2.5513951778411865, + "learning_rate": 1.3802086487850807e-06, + "loss": 3.7506, + "step": 799 + }, + { + "epoch": 0.8948545861297539, + "grad_norm": 3.390204429626465, + "learning_rate": 1.3515678863582431e-06, + "loss": 3.6553, + "step": 800 + }, + { + "epoch": 0.8959731543624161, + "grad_norm": 1.921186923980713, + "learning_rate": 1.3232191533563586e-06, + "loss": 3.7809, + "step": 801 + }, + { + "epoch": 0.8970917225950783, + "grad_norm": 2.1258394718170166, + "learning_rate": 1.2951627998518623e-06, + "loss": 3.8879, + "step": 802 + }, + { + "epoch": 0.8982102908277405, + "grad_norm": 3.5940604209899902, + "learning_rate": 1.267399172306663e-06, + "loss": 3.5702, + "step": 803 + }, + { + "epoch": 0.8993288590604027, + "grad_norm": 3.0253639221191406, + "learning_rate": 1.2399286135678423e-06, + "loss": 3.9817, + "step": 804 + }, + { + "epoch": 0.9004474272930649, + "grad_norm": 3.926387071609497, + "learning_rate": 1.212751462863454e-06, + "loss": 3.449, + "step": 805 + }, + { + "epoch": 0.901565995525727, + "grad_norm": 3.5187435150146484, + "learning_rate": 1.1858680557983171e-06, + "loss": 3.8943, + "step": 806 + }, + { + "epoch": 0.9026845637583892, + "grad_norm": 2.5346837043762207, + "learning_rate": 1.1592787243498631e-06, + "loss": 3.98, + "step": 807 + }, + { + "epoch": 0.9038031319910514, + "grad_norm": 2.5961804389953613, + "learning_rate": 1.132983796864054e-06, + "loss": 4.0311, + "step": 808 + }, + { + "epoch": 0.9049217002237137, + "grad_norm": 2.107836961746216, + "learning_rate": 1.10698359805132e-06, + "loss": 3.6457, + "step": 809 + }, + { + "epoch": 0.9060402684563759, + "grad_norm": 2.8412914276123047, + "learning_rate": 1.0812784489825507e-06, + "loss": 4.1803, + "step": 810 + }, + { + "epoch": 0.9071588366890381, + "grad_norm": 2.366673469543457, + "learning_rate": 1.0558686670851303e-06, + "loss": 3.9735, + "step": 811 + }, + { + "epoch": 0.9082774049217002, + "grad_norm": 2.198279857635498, + "learning_rate": 1.0307545661390139e-06, + "loss": 3.7553, + "step": 812 + }, + { + "epoch": 0.9093959731543624, + "grad_norm": 2.628816604614258, + "learning_rate": 1.0059364562728518e-06, + "loss": 3.7233, + "step": 813 + }, + { + "epoch": 0.9105145413870246, + "grad_norm": 2.737968921661377, + "learning_rate": 9.81414643960174e-07, + "loss": 3.8348, + "step": 814 + }, + { + "epoch": 0.9116331096196868, + "grad_norm": 1.8730359077453613, + "learning_rate": 9.5718943201559e-07, + "loss": 3.946, + "step": 815 + }, + { + "epoch": 0.912751677852349, + "grad_norm": 2.630864143371582, + "learning_rate": 9.332611195910584e-07, + "loss": 3.9293, + "step": 816 + }, + { + "epoch": 0.9138702460850112, + "grad_norm": 2.5964674949645996, + "learning_rate": 9.09630002172182e-07, + "loss": 3.9155, + "step": 817 + }, + { + "epoch": 0.9149888143176734, + "grad_norm": 1.8273096084594727, + "learning_rate": 8.862963715745687e-07, + "loss": 3.9176, + "step": 818 + }, + { + "epoch": 0.9161073825503355, + "grad_norm": 2.2656948566436768, + "learning_rate": 8.632605159402341e-07, + "loss": 4.1211, + "step": 819 + }, + { + "epoch": 0.9172259507829977, + "grad_norm": 2.1946887969970703, + "learning_rate": 8.405227197340216e-07, + "loss": 4.3939, + "step": 820 + }, + { + "epoch": 0.9183445190156599, + "grad_norm": 2.742736577987671, + "learning_rate": 8.180832637401075e-07, + "loss": 3.5518, + "step": 821 + }, + { + "epoch": 0.9194630872483222, + "grad_norm": 2.608285665512085, + "learning_rate": 7.959424250585323e-07, + "loss": 3.6809, + "step": 822 + }, + { + "epoch": 0.9205816554809844, + "grad_norm": 2.192291498184204, + "learning_rate": 7.741004771017691e-07, + "loss": 4.1284, + "step": 823 + }, + { + "epoch": 0.9217002237136466, + "grad_norm": 2.3379416465759277, + "learning_rate": 7.525576895913655e-07, + "loss": 3.8505, + "step": 824 + }, + { + "epoch": 0.9228187919463087, + "grad_norm": 2.133988618850708, + "learning_rate": 7.313143285545832e-07, + "loss": 3.9045, + "step": 825 + }, + { + "epoch": 0.9239373601789709, + "grad_norm": 2.171081781387329, + "learning_rate": 7.10370656321141e-07, + "loss": 3.7649, + "step": 826 + }, + { + "epoch": 0.9250559284116331, + "grad_norm": 1.7111164331436157, + "learning_rate": 6.897269315199628e-07, + "loss": 3.6116, + "step": 827 + }, + { + "epoch": 0.9261744966442953, + "grad_norm": 2.5523173809051514, + "learning_rate": 6.69383409075991e-07, + "loss": 4.0026, + "step": 828 + }, + { + "epoch": 0.9272930648769575, + "grad_norm": 2.1570398807525635, + "learning_rate": 6.493403402070247e-07, + "loss": 3.8648, + "step": 829 + }, + { + "epoch": 0.9284116331096197, + "grad_norm": 1.9127180576324463, + "learning_rate": 6.295979724206313e-07, + "loss": 3.9427, + "step": 830 + }, + { + "epoch": 0.9295302013422819, + "grad_norm": 2.699845790863037, + "learning_rate": 6.101565495110817e-07, + "loss": 3.5363, + "step": 831 + }, + { + "epoch": 0.930648769574944, + "grad_norm": 3.6851859092712402, + "learning_rate": 5.910163115563471e-07, + "loss": 3.7037, + "step": 832 + }, + { + "epoch": 0.9317673378076062, + "grad_norm": 2.0086777210235596, + "learning_rate": 5.721774949151298e-07, + "loss": 3.8431, + "step": 833 + }, + { + "epoch": 0.9328859060402684, + "grad_norm": 1.9770376682281494, + "learning_rate": 5.536403322239369e-07, + "loss": 3.8387, + "step": 834 + }, + { + "epoch": 0.9340044742729307, + "grad_norm": 2.4417667388916016, + "learning_rate": 5.35405052394225e-07, + "loss": 3.8362, + "step": 835 + }, + { + "epoch": 0.9351230425055929, + "grad_norm": 2.8682196140289307, + "learning_rate": 5.17471880609563e-07, + "loss": 3.8365, + "step": 836 + }, + { + "epoch": 0.9362416107382551, + "grad_norm": 2.904750108718872, + "learning_rate": 4.998410383228458e-07, + "loss": 3.7828, + "step": 837 + }, + { + "epoch": 0.9373601789709173, + "grad_norm": 2.7965922355651855, + "learning_rate": 4.825127432535714e-07, + "loss": 3.5812, + "step": 838 + }, + { + "epoch": 0.9384787472035794, + "grad_norm": 1.820050835609436, + "learning_rate": 4.654872093851487e-07, + "loss": 4.0985, + "step": 839 + }, + { + "epoch": 0.9395973154362416, + "grad_norm": 2.777374029159546, + "learning_rate": 4.487646469622464e-07, + "loss": 3.5268, + "step": 840 + }, + { + "epoch": 0.9407158836689038, + "grad_norm": 2.9808647632598877, + "learning_rate": 4.3234526248820686e-07, + "loss": 3.5723, + "step": 841 + }, + { + "epoch": 0.941834451901566, + "grad_norm": 3.188647985458374, + "learning_rate": 4.162292587224947e-07, + "loss": 3.3647, + "step": 842 + }, + { + "epoch": 0.9429530201342282, + "grad_norm": 2.2936878204345703, + "learning_rate": 4.0041683467819115e-07, + "loss": 3.887, + "step": 843 + }, + { + "epoch": 0.9440715883668904, + "grad_norm": 2.2591426372528076, + "learning_rate": 3.8490818561953414e-07, + "loss": 3.8009, + "step": 844 + }, + { + "epoch": 0.9451901565995525, + "grad_norm": 2.7897520065307617, + "learning_rate": 3.697035030595125e-07, + "loss": 3.556, + "step": 845 + }, + { + "epoch": 0.9463087248322147, + "grad_norm": 2.164191246032715, + "learning_rate": 3.548029747574927e-07, + "loss": 3.7152, + "step": 846 + }, + { + "epoch": 0.9474272930648769, + "grad_norm": 2.1192331314086914, + "learning_rate": 3.4020678471690934e-07, + "loss": 4.0369, + "step": 847 + }, + { + "epoch": 0.9485458612975392, + "grad_norm": 2.3934147357940674, + "learning_rate": 3.2591511318298686e-07, + "loss": 4.0403, + "step": 848 + }, + { + "epoch": 0.9496644295302014, + "grad_norm": 2.3561530113220215, + "learning_rate": 3.119281366405213e-07, + "loss": 3.7943, + "step": 849 + }, + { + "epoch": 0.9507829977628636, + "grad_norm": 2.1101136207580566, + "learning_rate": 2.9824602781168833e-07, + "loss": 3.9526, + "step": 850 + }, + { + "epoch": 0.9519015659955258, + "grad_norm": 2.3856959342956543, + "learning_rate": 2.8486895565391913e-07, + "loss": 3.7386, + "step": 851 + }, + { + "epoch": 0.9530201342281879, + "grad_norm": 2.2807648181915283, + "learning_rate": 2.7179708535781943e-07, + "loss": 3.6456, + "step": 852 + }, + { + "epoch": 0.9541387024608501, + "grad_norm": 2.5656046867370605, + "learning_rate": 2.590305783451097e-07, + "loss": 3.8164, + "step": 853 + }, + { + "epoch": 0.9552572706935123, + "grad_norm": 2.58715558052063, + "learning_rate": 2.4656959226665446e-07, + "loss": 4.0307, + "step": 854 + }, + { + "epoch": 0.9563758389261745, + "grad_norm": 2.5126988887786865, + "learning_rate": 2.3441428100050566e-07, + "loss": 3.7051, + "step": 855 + }, + { + "epoch": 0.9574944071588367, + "grad_norm": 2.549105644226074, + "learning_rate": 2.2256479464999313e-07, + "loss": 3.8441, + "step": 856 + }, + { + "epoch": 0.9586129753914989, + "grad_norm": 2.6297028064727783, + "learning_rate": 2.110212795418953e-07, + "loss": 3.4869, + "step": 857 + }, + { + "epoch": 0.959731543624161, + "grad_norm": 1.7067815065383911, + "learning_rate": 1.9978387822460197e-07, + "loss": 3.8855, + "step": 858 + }, + { + "epoch": 0.9608501118568232, + "grad_norm": 2.1829380989074707, + "learning_rate": 1.8885272946637944e-07, + "loss": 3.7166, + "step": 859 + }, + { + "epoch": 0.9619686800894854, + "grad_norm": 2.6756093502044678, + "learning_rate": 1.782279682536414e-07, + "loss": 3.706, + "step": 860 + }, + { + "epoch": 0.9630872483221476, + "grad_norm": 2.74208664894104, + "learning_rate": 1.679097257892892e-07, + "loss": 4.2839, + "step": 861 + }, + { + "epoch": 0.9642058165548099, + "grad_norm": 2.6451363563537598, + "learning_rate": 1.578981294910936e-07, + "loss": 3.8752, + "step": 862 + }, + { + "epoch": 0.9653243847874721, + "grad_norm": 2.704803705215454, + "learning_rate": 1.4819330299011004e-07, + "loss": 3.9167, + "step": 863 + }, + { + "epoch": 0.9664429530201343, + "grad_norm": 1.9471423625946045, + "learning_rate": 1.38795366129163e-07, + "loss": 3.9097, + "step": 864 + }, + { + "epoch": 0.9675615212527964, + "grad_norm": 2.0253708362579346, + "learning_rate": 1.2970443496136407e-07, + "loss": 3.8803, + "step": 865 + }, + { + "epoch": 0.9686800894854586, + "grad_norm": 1.8609451055526733, + "learning_rate": 1.2092062174867413e-07, + "loss": 3.8417, + "step": 866 + }, + { + "epoch": 0.9697986577181208, + "grad_norm": 2.4085533618927, + "learning_rate": 1.1244403496052658e-07, + "loss": 3.7451, + "step": 867 + }, + { + "epoch": 0.970917225950783, + "grad_norm": 3.3671045303344727, + "learning_rate": 1.042747792724702e-07, + "loss": 3.7212, + "step": 868 + }, + { + "epoch": 0.9720357941834452, + "grad_norm": 1.9560606479644775, + "learning_rate": 9.641295556489793e-08, + "loss": 3.9571, + "step": 869 + }, + { + "epoch": 0.9731543624161074, + "grad_norm": 1.984646201133728, + "learning_rate": 8.885866092178952e-08, + "loss": 3.9128, + "step": 870 + }, + { + "epoch": 0.9742729306487695, + "grad_norm": 2.1129088401794434, + "learning_rate": 8.161198862950692e-08, + "loss": 3.8933, + "step": 871 + }, + { + "epoch": 0.9753914988814317, + "grad_norm": 2.0977556705474854, + "learning_rate": 7.46730281756619e-08, + "loss": 3.9295, + "step": 872 + }, + { + "epoch": 0.9765100671140939, + "grad_norm": 2.7585601806640625, + "learning_rate": 6.804186524798362e-08, + "loss": 3.8923, + "step": 873 + }, + { + "epoch": 0.9776286353467561, + "grad_norm": 3.502300977706909, + "learning_rate": 6.171858173328604e-08, + "loss": 3.9105, + "step": 874 + }, + { + "epoch": 0.9787472035794184, + "grad_norm": 3.1645312309265137, + "learning_rate": 5.5703255716446637e-08, + "loss": 3.912, + "step": 875 + }, + { + "epoch": 0.9798657718120806, + "grad_norm": 3.4221417903900146, + "learning_rate": 4.999596147943486e-08, + "loss": 3.8685, + "step": 876 + }, + { + "epoch": 0.9809843400447428, + "grad_norm": 3.9628350734710693, + "learning_rate": 4.4596769500407366e-08, + "loss": 3.8902, + "step": 877 + }, + { + "epoch": 0.9821029082774049, + "grad_norm": 2.23947811126709, + "learning_rate": 3.9505746452830896e-08, + "loss": 3.7735, + "step": 878 + }, + { + "epoch": 0.9832214765100671, + "grad_norm": 3.1684134006500244, + "learning_rate": 3.4722955204652406e-08, + "loss": 3.7775, + "step": 879 + }, + { + "epoch": 0.9843400447427293, + "grad_norm": 2.4224395751953125, + "learning_rate": 3.02484548175469e-08, + "loss": 3.8073, + "step": 880 + }, + { + "epoch": 0.9854586129753915, + "grad_norm": 2.877962589263916, + "learning_rate": 2.6082300546154127e-08, + "loss": 3.9051, + "step": 881 + }, + { + "epoch": 0.9865771812080537, + "grad_norm": 1.9865648746490479, + "learning_rate": 2.2224543837423562e-08, + "loss": 3.8927, + "step": 882 + }, + { + "epoch": 0.9876957494407159, + "grad_norm": 2.3272900581359863, + "learning_rate": 1.8675232329967708e-08, + "loss": 3.9353, + "step": 883 + }, + { + "epoch": 0.9888143176733781, + "grad_norm": 1.9652239084243774, + "learning_rate": 1.5434409853473662e-08, + "loss": 3.8841, + "step": 884 + }, + { + "epoch": 0.9899328859060402, + "grad_norm": 3.6001975536346436, + "learning_rate": 1.2502116428161902e-08, + "loss": 4.0209, + "step": 885 + }, + { + "epoch": 0.9910514541387024, + "grad_norm": 2.0110864639282227, + "learning_rate": 9.878388264300541e-09, + "loss": 3.9221, + "step": 886 + }, + { + "epoch": 0.9921700223713646, + "grad_norm": 2.5183699131011963, + "learning_rate": 7.563257761744601e-09, + "loss": 3.6199, + "step": 887 + }, + { + "epoch": 0.9932885906040269, + "grad_norm": 2.3704769611358643, + "learning_rate": 5.55675350954743e-09, + "loss": 3.603, + "step": 888 + }, + { + "epoch": 0.9944071588366891, + "grad_norm": 2.055521249771118, + "learning_rate": 3.8589002855971e-09, + "loss": 3.906, + "step": 889 + }, + { + "epoch": 0.9955257270693513, + "grad_norm": 2.1967926025390625, + "learning_rate": 2.4697190563194305e-09, + "loss": 3.9161, + "step": 890 + }, + { + "epoch": 0.9966442953020134, + "grad_norm": 3.107943534851074, + "learning_rate": 1.3892269764198551e-09, + "loss": 3.9214, + "step": 891 + }, + { + "epoch": 0.9977628635346756, + "grad_norm": 2.653813123703003, + "learning_rate": 6.174373886586037e-10, + "loss": 3.7566, + "step": 892 + }, + { + "epoch": 0.9988814317673378, + "grad_norm": 2.262310743331909, + "learning_rate": 1.5435982370359813e-10, + "loss": 3.6915, + "step": 893 + }, + { + "epoch": 1.0, + "grad_norm": 2.972426414489746, + "learning_rate": 0.0, + "loss": 4.5787, + "step": 894 + } + ], + "logging_steps": 1, + "max_steps": 894, + "num_input_tokens_seen": 0, + "num_train_epochs": 1, + "save_steps": 5000, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 1.4928497725151232e+16, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}