| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.9998234442495793, | |
| "eval_steps": 1000, | |
| "global_step": 5309, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0009416306689108865, | |
| "grad_norm": 13.400167465209961, | |
| "learning_rate": 1e-05, | |
| "loss": 4.0168, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.001883261337821773, | |
| "grad_norm": 12.05469036102295, | |
| "learning_rate": 1e-05, | |
| "loss": 3.4972, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.0028248920067326594, | |
| "grad_norm": 11.709904670715332, | |
| "learning_rate": 1e-05, | |
| "loss": 3.5447, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.003766522675643546, | |
| "grad_norm": 10.864925384521484, | |
| "learning_rate": 1e-05, | |
| "loss": 3.4348, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.004708153344554432, | |
| "grad_norm": 12.948342323303223, | |
| "learning_rate": 1e-05, | |
| "loss": 3.4288, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.005649784013465319, | |
| "grad_norm": 10.90282917022705, | |
| "learning_rate": 1e-05, | |
| "loss": 3.4199, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.006591414682376205, | |
| "grad_norm": 10.574660301208496, | |
| "learning_rate": 1e-05, | |
| "loss": 3.4564, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.007533045351287092, | |
| "grad_norm": 12.737661361694336, | |
| "learning_rate": 1e-05, | |
| "loss": 3.4071, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.008474676020197977, | |
| "grad_norm": 11.787413597106934, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3169, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.009416306689108865, | |
| "grad_norm": 11.487077713012695, | |
| "learning_rate": 1e-05, | |
| "loss": 3.4083, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.01035793735801975, | |
| "grad_norm": 10.931989669799805, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3384, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.011299568026930638, | |
| "grad_norm": 11.073975563049316, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2959, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.012241198695841523, | |
| "grad_norm": 10.520795822143555, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3513, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.01318282936475241, | |
| "grad_norm": 10.708564758300781, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1843, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.014124460033663296, | |
| "grad_norm": 9.278353691101074, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2907, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.015066090702574184, | |
| "grad_norm": 9.356634140014648, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3033, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.01600772137148507, | |
| "grad_norm": 10.317462921142578, | |
| "learning_rate": 1e-05, | |
| "loss": 3.402, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.016949352040395955, | |
| "grad_norm": 9.032744407653809, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3137, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.017890982709306842, | |
| "grad_norm": 9.985954284667969, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3074, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.01883261337821773, | |
| "grad_norm": 9.465397834777832, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3329, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.019774244047128613, | |
| "grad_norm": 9.615800857543945, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3839, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.0207158747160395, | |
| "grad_norm": 9.558786392211914, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3383, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.021657505384950388, | |
| "grad_norm": 10.330078125, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2845, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.022599136053861275, | |
| "grad_norm": 9.462120056152344, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2918, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.02354076672277216, | |
| "grad_norm": 10.415687561035156, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3509, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.024482397391683047, | |
| "grad_norm": 9.891234397888184, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3196, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.025424028060593934, | |
| "grad_norm": 10.6549072265625, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2702, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.02636565872950482, | |
| "grad_norm": 10.514317512512207, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3639, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.027307289398415705, | |
| "grad_norm": 10.233137130737305, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3693, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.028248920067326592, | |
| "grad_norm": 9.689544677734375, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2754, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.02919055073623748, | |
| "grad_norm": 10.0300931930542, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2058, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.030132181405148367, | |
| "grad_norm": 9.759441375732422, | |
| "learning_rate": 1e-05, | |
| "loss": 3.216, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.03107381207405925, | |
| "grad_norm": 9.909936904907227, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3169, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.03201544274297014, | |
| "grad_norm": 10.085428237915039, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3579, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.032957073411881026, | |
| "grad_norm": 11.790485382080078, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3376, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.03389870408079191, | |
| "grad_norm": 9.588286399841309, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2441, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.0348403347497028, | |
| "grad_norm": 9.073866844177246, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3118, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.035781965418613684, | |
| "grad_norm": 11.346445083618164, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2795, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.03672359608752457, | |
| "grad_norm": 8.872209548950195, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1868, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.03766522675643546, | |
| "grad_norm": 10.55789566040039, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2716, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.03860685742534634, | |
| "grad_norm": 9.99712085723877, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2368, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.03954848809425723, | |
| "grad_norm": 10.883671760559082, | |
| "learning_rate": 1e-05, | |
| "loss": 3.237, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.04049011876316812, | |
| "grad_norm": 10.810287475585938, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2724, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.041431749432079, | |
| "grad_norm": 8.746783256530762, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2741, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.04237338010098989, | |
| "grad_norm": 10.720795631408691, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2725, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.043315010769900776, | |
| "grad_norm": 10.715919494628906, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2604, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.04425664143881166, | |
| "grad_norm": 9.999095916748047, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3685, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.04519827210772255, | |
| "grad_norm": 9.06040096282959, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3883, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.046139902776633435, | |
| "grad_norm": 9.397573471069336, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2377, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.04708153344554432, | |
| "grad_norm": 11.011545181274414, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1805, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.04802316411445521, | |
| "grad_norm": 8.947606086730957, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1641, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.04896479478336609, | |
| "grad_norm": 9.644259452819824, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2127, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.049906425452276984, | |
| "grad_norm": 9.601791381835938, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2503, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.05084805612118787, | |
| "grad_norm": 9.706978797912598, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2276, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.05178968679009875, | |
| "grad_norm": 10.092705726623535, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2853, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.05273131745900964, | |
| "grad_norm": 10.306342124938965, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3613, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.053672948127920526, | |
| "grad_norm": 9.216752052307129, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1688, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.05461457879683141, | |
| "grad_norm": 9.452301979064941, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2417, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.0555562094657423, | |
| "grad_norm": 9.074745178222656, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3112, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.056497840134653185, | |
| "grad_norm": 8.492777824401855, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2811, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.05743947080356407, | |
| "grad_norm": 9.102340698242188, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2543, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.05838110147247496, | |
| "grad_norm": 9.913708686828613, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3047, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.059322732141385844, | |
| "grad_norm": 8.82358455657959, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2744, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.060264362810296734, | |
| "grad_norm": 10.01116943359375, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3184, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.06120599347920762, | |
| "grad_norm": 8.861343383789062, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3169, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.0621476241481185, | |
| "grad_norm": 8.552498817443848, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3267, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.06308925481702939, | |
| "grad_norm": 9.180069923400879, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1593, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.06403088548594028, | |
| "grad_norm": 8.194535255432129, | |
| "learning_rate": 1e-05, | |
| "loss": 3.252, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.06497251615485117, | |
| "grad_norm": 9.234159469604492, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3233, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.06591414682376205, | |
| "grad_norm": 9.615317344665527, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1951, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.06685577749267294, | |
| "grad_norm": 9.555344581604004, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3024, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.06779740816158382, | |
| "grad_norm": 9.033773422241211, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2833, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.0687390388304947, | |
| "grad_norm": 10.110182762145996, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3022, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.0696806694994056, | |
| "grad_norm": 8.350312232971191, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1566, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.07062230016831648, | |
| "grad_norm": 9.642204284667969, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2279, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.07156393083722737, | |
| "grad_norm": 8.922101020812988, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1227, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.07250556150613825, | |
| "grad_norm": 8.6968412399292, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2537, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.07344719217504914, | |
| "grad_norm": 9.628043174743652, | |
| "learning_rate": 1e-05, | |
| "loss": 3.335, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.07438882284396003, | |
| "grad_norm": 9.418974876403809, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2609, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.07533045351287092, | |
| "grad_norm": 9.786445617675781, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1916, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.0762720841817818, | |
| "grad_norm": 9.372411727905273, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2404, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.07721371485069269, | |
| "grad_norm": 9.255708694458008, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1984, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.07815534551960357, | |
| "grad_norm": 8.31712818145752, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2435, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.07909697618851445, | |
| "grad_norm": 9.45557975769043, | |
| "learning_rate": 1e-05, | |
| "loss": 3.271, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.08003860685742535, | |
| "grad_norm": 9.2068510055542, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2629, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.08098023752633624, | |
| "grad_norm": 9.854654312133789, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2466, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.08192186819524712, | |
| "grad_norm": 9.899443626403809, | |
| "learning_rate": 1e-05, | |
| "loss": 3.303, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.082863498864158, | |
| "grad_norm": 8.167075157165527, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1332, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.08380512953306889, | |
| "grad_norm": 9.357007026672363, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2161, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.08474676020197978, | |
| "grad_norm": 8.328939437866211, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2558, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.08568839087089067, | |
| "grad_norm": 10.344287872314453, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1148, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.08663002153980155, | |
| "grad_norm": 8.459121704101562, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2366, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.08757165220871244, | |
| "grad_norm": 9.339311599731445, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3596, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.08851328287762332, | |
| "grad_norm": 9.49207592010498, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2823, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.0894549135465342, | |
| "grad_norm": 9.24524974822998, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2593, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.0903965442154451, | |
| "grad_norm": 9.082175254821777, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1761, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.09133817488435599, | |
| "grad_norm": 8.596846580505371, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2005, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.09227980555326687, | |
| "grad_norm": 9.297995567321777, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2414, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.09322143622217775, | |
| "grad_norm": 9.254470825195312, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2309, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.09416306689108864, | |
| "grad_norm": 9.504804611206055, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2978, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.09510469755999953, | |
| "grad_norm": 9.160736083984375, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2522, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 0.09604632822891042, | |
| "grad_norm": 8.80601692199707, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1128, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.0969879588978213, | |
| "grad_norm": 9.246788024902344, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2469, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 0.09792958956673219, | |
| "grad_norm": 8.799399375915527, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1663, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.09887122023564307, | |
| "grad_norm": 8.722356796264648, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2616, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.09981285090455397, | |
| "grad_norm": 9.724103927612305, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2367, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.10075448157346485, | |
| "grad_norm": 9.14875602722168, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1932, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 0.10169611224237574, | |
| "grad_norm": 8.882226943969727, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1863, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.10263774291128662, | |
| "grad_norm": 8.96247673034668, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3411, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 0.1035793735801975, | |
| "grad_norm": 7.7259202003479, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2209, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.10452100424910839, | |
| "grad_norm": 8.8307466506958, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1653, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 0.10546263491801929, | |
| "grad_norm": 9.439279556274414, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2451, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.10640426558693017, | |
| "grad_norm": 8.951433181762695, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2327, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 0.10734589625584105, | |
| "grad_norm": 8.996528625488281, | |
| "learning_rate": 1e-05, | |
| "loss": 3.177, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.10828752692475194, | |
| "grad_norm": 8.661666870117188, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2389, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 0.10922915759366282, | |
| "grad_norm": 7.827404499053955, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2133, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.11017078826257372, | |
| "grad_norm": 8.948150634765625, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1059, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 0.1111124189314846, | |
| "grad_norm": 9.571684837341309, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2737, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.11205404960039549, | |
| "grad_norm": 8.415332794189453, | |
| "learning_rate": 1e-05, | |
| "loss": 3.181, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 0.11299568026930637, | |
| "grad_norm": 8.199889183044434, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1989, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.11393731093821725, | |
| "grad_norm": 8.935426712036133, | |
| "learning_rate": 1e-05, | |
| "loss": 3.165, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 0.11487894160712814, | |
| "grad_norm": 9.254772186279297, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2167, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.11582057227603904, | |
| "grad_norm": 8.660184860229492, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1573, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 0.11676220294494992, | |
| "grad_norm": 9.153916358947754, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2069, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.1177038336138608, | |
| "grad_norm": 10.136688232421875, | |
| "learning_rate": 1e-05, | |
| "loss": 3.184, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 0.11864546428277169, | |
| "grad_norm": 8.210442543029785, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2377, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.11958709495168257, | |
| "grad_norm": 10.24669361114502, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2338, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 0.12052872562059347, | |
| "grad_norm": 8.631328582763672, | |
| "learning_rate": 1e-05, | |
| "loss": 3.3123, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.12147035628950435, | |
| "grad_norm": 8.293212890625, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1506, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 0.12241198695841524, | |
| "grad_norm": 9.084228515625, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1423, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.12335361762732612, | |
| "grad_norm": 7.838293552398682, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0779, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 0.124295248296237, | |
| "grad_norm": 10.947461128234863, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1623, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.1252368789651479, | |
| "grad_norm": 9.293968200683594, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2047, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 0.12617850963405877, | |
| "grad_norm": 8.150403022766113, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1688, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.12712014030296967, | |
| "grad_norm": 8.54336929321289, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2182, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 0.12806177097188057, | |
| "grad_norm": 7.81638240814209, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2553, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.12900340164079144, | |
| "grad_norm": 8.324007034301758, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1702, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 0.12994503230970234, | |
| "grad_norm": 9.187409400939941, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1752, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.1308866629786132, | |
| "grad_norm": 8.79953670501709, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2211, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 0.1318282936475241, | |
| "grad_norm": 9.9981689453125, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1846, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.13276992431643497, | |
| "grad_norm": 8.74207592010498, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2351, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 0.13371155498534587, | |
| "grad_norm": 8.988236427307129, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2312, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.13465318565425677, | |
| "grad_norm": 8.005229949951172, | |
| "learning_rate": 1e-05, | |
| "loss": 3.218, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 0.13559481632316764, | |
| "grad_norm": 9.276175498962402, | |
| "learning_rate": 1e-05, | |
| "loss": 3.279, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.13653644699207854, | |
| "grad_norm": 8.09757137298584, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2876, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 0.1374780776609894, | |
| "grad_norm": 8.40072250366211, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2547, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.1384197083299003, | |
| "grad_norm": 8.614055633544922, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2795, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 0.1393613389988112, | |
| "grad_norm": 9.18989086151123, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0988, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.14030296966772207, | |
| "grad_norm": 8.401784896850586, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1612, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 0.14124460033663297, | |
| "grad_norm": 8.722879409790039, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1921, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.14218623100554384, | |
| "grad_norm": 8.14240550994873, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2371, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 0.14312786167445474, | |
| "grad_norm": 10.812965393066406, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2239, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.14406949234336563, | |
| "grad_norm": 8.65410327911377, | |
| "learning_rate": 1e-05, | |
| "loss": 3.267, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 0.1450111230122765, | |
| "grad_norm": 7.997138977050781, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2387, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.1459527536811874, | |
| "grad_norm": 8.726889610290527, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1542, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 0.14689438435009827, | |
| "grad_norm": 8.375578880310059, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1742, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.14783601501900917, | |
| "grad_norm": 8.577352523803711, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2441, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 0.14877764568792007, | |
| "grad_norm": 8.30477523803711, | |
| "learning_rate": 1e-05, | |
| "loss": 3.133, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.14971927635683094, | |
| "grad_norm": 8.190656661987305, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1822, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 0.15066090702574184, | |
| "grad_norm": 8.631675720214844, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2068, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.1516025376946527, | |
| "grad_norm": 9.275361061096191, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2194, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 0.1525441683635636, | |
| "grad_norm": 8.464102745056152, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1456, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.1534857990324745, | |
| "grad_norm": 8.44454288482666, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1675, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 0.15442742970138537, | |
| "grad_norm": 8.282485961914062, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1728, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.15536906037029627, | |
| "grad_norm": 8.142829895019531, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2012, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 0.15631069103920714, | |
| "grad_norm": 8.29863452911377, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1008, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.15725232170811804, | |
| "grad_norm": 8.145419120788574, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0711, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 0.1581939523770289, | |
| "grad_norm": 8.715362548828125, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1625, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.1591355830459398, | |
| "grad_norm": 8.283182144165039, | |
| "learning_rate": 1e-05, | |
| "loss": 3.066, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 0.1600772137148507, | |
| "grad_norm": 9.210651397705078, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2498, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.16101884438376157, | |
| "grad_norm": 8.161669731140137, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2414, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 0.16196047505267247, | |
| "grad_norm": 8.109793663024902, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1307, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.16290210572158334, | |
| "grad_norm": 8.473872184753418, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1473, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 0.16384373639049424, | |
| "grad_norm": 8.496044158935547, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1971, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.16478536705940514, | |
| "grad_norm": 7.473423004150391, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1617, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 0.165726997728316, | |
| "grad_norm": 7.770011901855469, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1806, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.1666686283972269, | |
| "grad_norm": 8.22207260131836, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2673, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 0.16761025906613777, | |
| "grad_norm": 9.465033531188965, | |
| "learning_rate": 1e-05, | |
| "loss": 3.273, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.16855188973504867, | |
| "grad_norm": 8.1371431350708, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1663, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 0.16949352040395957, | |
| "grad_norm": 8.64779281616211, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1979, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.17043515107287044, | |
| "grad_norm": 8.860164642333984, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2013, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 0.17137678174178134, | |
| "grad_norm": 8.785137176513672, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1721, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.1723184124106922, | |
| "grad_norm": 8.557008743286133, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1774, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 0.1732600430796031, | |
| "grad_norm": 8.913068771362305, | |
| "learning_rate": 1e-05, | |
| "loss": 3.095, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.174201673748514, | |
| "grad_norm": 7.609038352966309, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1806, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 0.17514330441742487, | |
| "grad_norm": 7.761782169342041, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2352, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.17608493508633577, | |
| "grad_norm": 8.180496215820312, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2167, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 0.17702656575524664, | |
| "grad_norm": 8.675491333007812, | |
| "learning_rate": 1e-05, | |
| "loss": 3.12, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.17796819642415754, | |
| "grad_norm": 8.088050842285156, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1027, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 0.1789098270930684, | |
| "grad_norm": 8.452052116394043, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9981, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.1798514577619793, | |
| "grad_norm": 8.1648588180542, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0539, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 0.1807930884308902, | |
| "grad_norm": 8.245767593383789, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1362, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.18173471909980107, | |
| "grad_norm": 8.551701545715332, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1821, | |
| "step": 965 | |
| }, | |
| { | |
| "epoch": 0.18267634976871197, | |
| "grad_norm": 8.178792953491211, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1523, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.18361798043762284, | |
| "grad_norm": 8.187594413757324, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2531, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 0.18455961110653374, | |
| "grad_norm": 7.719356060028076, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1006, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.18550124177544464, | |
| "grad_norm": 8.289334297180176, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1259, | |
| "step": 985 | |
| }, | |
| { | |
| "epoch": 0.1864428724443555, | |
| "grad_norm": 8.719657897949219, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1813, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.1873845031132664, | |
| "grad_norm": 8.192325592041016, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1389, | |
| "step": 995 | |
| }, | |
| { | |
| "epoch": 0.18832613378217727, | |
| "grad_norm": 7.759548664093018, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1203, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.18832613378217727, | |
| "eval_accuracy": 0.3959961907900079, | |
| "eval_loss": 3.179180860519409, | |
| "eval_runtime": 1109.8083, | |
| "eval_samples_per_second": 34.024, | |
| "eval_steps_per_second": 8.506, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.18926776445108817, | |
| "grad_norm": 8.558948516845703, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2909, | |
| "step": 1005 | |
| }, | |
| { | |
| "epoch": 0.19020939511999907, | |
| "grad_norm": 7.582333087921143, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1388, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.19115102578890994, | |
| "grad_norm": 8.941636085510254, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1176, | |
| "step": 1015 | |
| }, | |
| { | |
| "epoch": 0.19209265645782084, | |
| "grad_norm": 8.161066055297852, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1302, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.1930342871267317, | |
| "grad_norm": 8.573451042175293, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1978, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 0.1939759177956426, | |
| "grad_norm": 9.676179885864258, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2441, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.1949175484645535, | |
| "grad_norm": 9.518370628356934, | |
| "learning_rate": 1e-05, | |
| "loss": 3.209, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 0.19585917913346437, | |
| "grad_norm": 8.120454788208008, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0915, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.19680080980237527, | |
| "grad_norm": 8.362229347229004, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2262, | |
| "step": 1045 | |
| }, | |
| { | |
| "epoch": 0.19774244047128614, | |
| "grad_norm": 8.197416305541992, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1164, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.19868407114019704, | |
| "grad_norm": 8.58997631072998, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2323, | |
| "step": 1055 | |
| }, | |
| { | |
| "epoch": 0.19962570180910794, | |
| "grad_norm": 7.581511497497559, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1852, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.2005673324780188, | |
| "grad_norm": 7.385310649871826, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2516, | |
| "step": 1065 | |
| }, | |
| { | |
| "epoch": 0.2015089631469297, | |
| "grad_norm": 7.678237438201904, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1669, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.20245059381584057, | |
| "grad_norm": 8.25271224975586, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1575, | |
| "step": 1075 | |
| }, | |
| { | |
| "epoch": 0.20339222448475147, | |
| "grad_norm": 7.587473392486572, | |
| "learning_rate": 1e-05, | |
| "loss": 3.182, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.20433385515366234, | |
| "grad_norm": 7.0264153480529785, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1049, | |
| "step": 1085 | |
| }, | |
| { | |
| "epoch": 0.20527548582257324, | |
| "grad_norm": 7.00242280960083, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0769, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.20621711649148414, | |
| "grad_norm": 8.321081161499023, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0977, | |
| "step": 1095 | |
| }, | |
| { | |
| "epoch": 0.207158747160395, | |
| "grad_norm": 8.19550609588623, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1317, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.2081003778293059, | |
| "grad_norm": 7.681639671325684, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2471, | |
| "step": 1105 | |
| }, | |
| { | |
| "epoch": 0.20904200849821677, | |
| "grad_norm": 8.00836181640625, | |
| "learning_rate": 1e-05, | |
| "loss": 3.183, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.20998363916712767, | |
| "grad_norm": 7.7261576652526855, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2104, | |
| "step": 1115 | |
| }, | |
| { | |
| "epoch": 0.21092526983603857, | |
| "grad_norm": 7.512089729309082, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0842, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.21186690050494944, | |
| "grad_norm": 8.911776542663574, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2527, | |
| "step": 1125 | |
| }, | |
| { | |
| "epoch": 0.21280853117386034, | |
| "grad_norm": 7.831639766693115, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2335, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.2137501618427712, | |
| "grad_norm": 7.919592380523682, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1974, | |
| "step": 1135 | |
| }, | |
| { | |
| "epoch": 0.2146917925116821, | |
| "grad_norm": 7.262505531311035, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0964, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.215633423180593, | |
| "grad_norm": 7.787773132324219, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1731, | |
| "step": 1145 | |
| }, | |
| { | |
| "epoch": 0.21657505384950387, | |
| "grad_norm": 8.574644088745117, | |
| "learning_rate": 1e-05, | |
| "loss": 3.203, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.21751668451841477, | |
| "grad_norm": 8.63223648071289, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0797, | |
| "step": 1155 | |
| }, | |
| { | |
| "epoch": 0.21845831518732564, | |
| "grad_norm": 7.836942195892334, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1964, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.21939994585623654, | |
| "grad_norm": 8.048019409179688, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0157, | |
| "step": 1165 | |
| }, | |
| { | |
| "epoch": 0.22034157652514744, | |
| "grad_norm": 9.863574981689453, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1937, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.2212832071940583, | |
| "grad_norm": 7.458190441131592, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1185, | |
| "step": 1175 | |
| }, | |
| { | |
| "epoch": 0.2222248378629692, | |
| "grad_norm": 7.304734230041504, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0673, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.22316646853188007, | |
| "grad_norm": 8.21141529083252, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1101, | |
| "step": 1185 | |
| }, | |
| { | |
| "epoch": 0.22410809920079097, | |
| "grad_norm": 8.499639511108398, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2017, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.22504972986970187, | |
| "grad_norm": 8.153023719787598, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1853, | |
| "step": 1195 | |
| }, | |
| { | |
| "epoch": 0.22599136053861274, | |
| "grad_norm": 8.166133880615234, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0598, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.22693299120752364, | |
| "grad_norm": 8.11247730255127, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2231, | |
| "step": 1205 | |
| }, | |
| { | |
| "epoch": 0.2278746218764345, | |
| "grad_norm": 9.046285629272461, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0795, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 0.2288162525453454, | |
| "grad_norm": 8.127588272094727, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1155, | |
| "step": 1215 | |
| }, | |
| { | |
| "epoch": 0.22975788321425628, | |
| "grad_norm": 7.537458419799805, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1393, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.23069951388316717, | |
| "grad_norm": 7.949858665466309, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1817, | |
| "step": 1225 | |
| }, | |
| { | |
| "epoch": 0.23164114455207807, | |
| "grad_norm": 8.178318977355957, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1628, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.23258277522098894, | |
| "grad_norm": 7.652297019958496, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1317, | |
| "step": 1235 | |
| }, | |
| { | |
| "epoch": 0.23352440588989984, | |
| "grad_norm": 7.234852313995361, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1783, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.2344660365588107, | |
| "grad_norm": 7.958459854125977, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1752, | |
| "step": 1245 | |
| }, | |
| { | |
| "epoch": 0.2354076672277216, | |
| "grad_norm": 7.870058536529541, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2413, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.2363492978966325, | |
| "grad_norm": 8.846611976623535, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1678, | |
| "step": 1255 | |
| }, | |
| { | |
| "epoch": 0.23729092856554337, | |
| "grad_norm": 8.129776000976562, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1669, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.23823255923445427, | |
| "grad_norm": 6.927892208099365, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1201, | |
| "step": 1265 | |
| }, | |
| { | |
| "epoch": 0.23917418990336514, | |
| "grad_norm": 9.028277397155762, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1909, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 0.24011582057227604, | |
| "grad_norm": 8.353940963745117, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2572, | |
| "step": 1275 | |
| }, | |
| { | |
| "epoch": 0.24105745124118694, | |
| "grad_norm": 7.9163737297058105, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1206, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.2419990819100978, | |
| "grad_norm": 8.831631660461426, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2485, | |
| "step": 1285 | |
| }, | |
| { | |
| "epoch": 0.2429407125790087, | |
| "grad_norm": 7.537317752838135, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1398, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.24388234324791958, | |
| "grad_norm": 7.471547603607178, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0652, | |
| "step": 1295 | |
| }, | |
| { | |
| "epoch": 0.24482397391683047, | |
| "grad_norm": 7.851377487182617, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0686, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.24576560458574137, | |
| "grad_norm": 8.015359878540039, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2359, | |
| "step": 1305 | |
| }, | |
| { | |
| "epoch": 0.24670723525465224, | |
| "grad_norm": 8.371912002563477, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0857, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 0.24764886592356314, | |
| "grad_norm": 7.510787487030029, | |
| "learning_rate": 1e-05, | |
| "loss": 3.244, | |
| "step": 1315 | |
| }, | |
| { | |
| "epoch": 0.248590496592474, | |
| "grad_norm": 7.756171703338623, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1444, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.2495321272613849, | |
| "grad_norm": 8.038382530212402, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2096, | |
| "step": 1325 | |
| }, | |
| { | |
| "epoch": 0.2504737579302958, | |
| "grad_norm": 7.765430927276611, | |
| "learning_rate": 1e-05, | |
| "loss": 3.139, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 0.2514153885992067, | |
| "grad_norm": 8.474689483642578, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1436, | |
| "step": 1335 | |
| }, | |
| { | |
| "epoch": 0.25235701926811754, | |
| "grad_norm": 9.304710388183594, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1479, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.25329864993702844, | |
| "grad_norm": 10.669320106506348, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0753, | |
| "step": 1345 | |
| }, | |
| { | |
| "epoch": 0.25424028060593934, | |
| "grad_norm": 6.909900665283203, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0703, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.25518191127485024, | |
| "grad_norm": 8.000332832336426, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0837, | |
| "step": 1355 | |
| }, | |
| { | |
| "epoch": 0.25612354194376113, | |
| "grad_norm": 7.857721328735352, | |
| "learning_rate": 1e-05, | |
| "loss": 3.121, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.257065172612672, | |
| "grad_norm": 7.735369682312012, | |
| "learning_rate": 1e-05, | |
| "loss": 3.122, | |
| "step": 1365 | |
| }, | |
| { | |
| "epoch": 0.2580068032815829, | |
| "grad_norm": 7.965452194213867, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1381, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 0.2589484339504938, | |
| "grad_norm": 8.671842575073242, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1744, | |
| "step": 1375 | |
| }, | |
| { | |
| "epoch": 0.25989006461940467, | |
| "grad_norm": 7.975509166717529, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1222, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.26083169528831557, | |
| "grad_norm": 7.30696964263916, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1122, | |
| "step": 1385 | |
| }, | |
| { | |
| "epoch": 0.2617733259572264, | |
| "grad_norm": 8.272529602050781, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0969, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 0.2627149566261373, | |
| "grad_norm": 8.145869255065918, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0817, | |
| "step": 1395 | |
| }, | |
| { | |
| "epoch": 0.2636565872950482, | |
| "grad_norm": 7.628718852996826, | |
| "learning_rate": 1e-05, | |
| "loss": 3.105, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.2645982179639591, | |
| "grad_norm": 6.900229454040527, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2028, | |
| "step": 1405 | |
| }, | |
| { | |
| "epoch": 0.26553984863286995, | |
| "grad_norm": 7.9082207679748535, | |
| "learning_rate": 1e-05, | |
| "loss": 3.014, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 0.26648147930178084, | |
| "grad_norm": 8.532185554504395, | |
| "learning_rate": 1e-05, | |
| "loss": 3.143, | |
| "step": 1415 | |
| }, | |
| { | |
| "epoch": 0.26742310997069174, | |
| "grad_norm": 8.344785690307617, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1865, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.26836474063960264, | |
| "grad_norm": 7.369927883148193, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0858, | |
| "step": 1425 | |
| }, | |
| { | |
| "epoch": 0.26930637130851354, | |
| "grad_norm": 7.539013385772705, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9854, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 0.2702480019774244, | |
| "grad_norm": 8.182465553283691, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0504, | |
| "step": 1435 | |
| }, | |
| { | |
| "epoch": 0.2711896326463353, | |
| "grad_norm": 8.410788536071777, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0354, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.2721312633152462, | |
| "grad_norm": 7.914114475250244, | |
| "learning_rate": 1e-05, | |
| "loss": 3.067, | |
| "step": 1445 | |
| }, | |
| { | |
| "epoch": 0.2730728939841571, | |
| "grad_norm": 8.113722801208496, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1172, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.27401452465306797, | |
| "grad_norm": 7.887296676635742, | |
| "learning_rate": 1e-05, | |
| "loss": 3.139, | |
| "step": 1455 | |
| }, | |
| { | |
| "epoch": 0.2749561553219788, | |
| "grad_norm": 9.083703994750977, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1805, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.2758977859908897, | |
| "grad_norm": 7.937193393707275, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1389, | |
| "step": 1465 | |
| }, | |
| { | |
| "epoch": 0.2768394166598006, | |
| "grad_norm": 8.576871871948242, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0739, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 0.2777810473287115, | |
| "grad_norm": 8.345330238342285, | |
| "learning_rate": 1e-05, | |
| "loss": 3.043, | |
| "step": 1475 | |
| }, | |
| { | |
| "epoch": 0.2787226779976224, | |
| "grad_norm": 8.3008394241333, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2218, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.27966430866653325, | |
| "grad_norm": 7.56035852432251, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1249, | |
| "step": 1485 | |
| }, | |
| { | |
| "epoch": 0.28060593933544414, | |
| "grad_norm": 7.499727249145508, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0944, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 0.28154757000435504, | |
| "grad_norm": 8.389445304870605, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0844, | |
| "step": 1495 | |
| }, | |
| { | |
| "epoch": 0.28248920067326594, | |
| "grad_norm": 9.202953338623047, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2153, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.28343083134217684, | |
| "grad_norm": 7.818172454833984, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0676, | |
| "step": 1505 | |
| }, | |
| { | |
| "epoch": 0.2843724620110877, | |
| "grad_norm": 7.743332386016846, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0716, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 0.2853140926799986, | |
| "grad_norm": 8.619895935058594, | |
| "learning_rate": 1e-05, | |
| "loss": 3.142, | |
| "step": 1515 | |
| }, | |
| { | |
| "epoch": 0.2862557233489095, | |
| "grad_norm": 7.999989986419678, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2461, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.28719735401782037, | |
| "grad_norm": 7.5509772300720215, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1051, | |
| "step": 1525 | |
| }, | |
| { | |
| "epoch": 0.28813898468673127, | |
| "grad_norm": 7.679853916168213, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1131, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 0.2890806153556421, | |
| "grad_norm": 7.623820781707764, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1323, | |
| "step": 1535 | |
| }, | |
| { | |
| "epoch": 0.290022246024553, | |
| "grad_norm": 7.860290050506592, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1595, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.2909638766934639, | |
| "grad_norm": 7.90553092956543, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0943, | |
| "step": 1545 | |
| }, | |
| { | |
| "epoch": 0.2919055073623748, | |
| "grad_norm": 8.377348899841309, | |
| "learning_rate": 1e-05, | |
| "loss": 3.194, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.2928471380312857, | |
| "grad_norm": 7.801243305206299, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0033, | |
| "step": 1555 | |
| }, | |
| { | |
| "epoch": 0.29378876870019655, | |
| "grad_norm": 8.063822746276855, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1661, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.29473039936910744, | |
| "grad_norm": 7.1917314529418945, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1411, | |
| "step": 1565 | |
| }, | |
| { | |
| "epoch": 0.29567203003801834, | |
| "grad_norm": 7.78692102432251, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0471, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 0.29661366070692924, | |
| "grad_norm": 8.212677001953125, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1417, | |
| "step": 1575 | |
| }, | |
| { | |
| "epoch": 0.29755529137584014, | |
| "grad_norm": 6.906081676483154, | |
| "learning_rate": 1e-05, | |
| "loss": 3.13, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.298496922044751, | |
| "grad_norm": 7.585642337799072, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1489, | |
| "step": 1585 | |
| }, | |
| { | |
| "epoch": 0.2994385527136619, | |
| "grad_norm": 7.452507495880127, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1203, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 0.3003801833825728, | |
| "grad_norm": 9.468456268310547, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1018, | |
| "step": 1595 | |
| }, | |
| { | |
| "epoch": 0.30132181405148367, | |
| "grad_norm": 8.016668319702148, | |
| "learning_rate": 1e-05, | |
| "loss": 3.209, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.30226344472039457, | |
| "grad_norm": 7.062180995941162, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2921, | |
| "step": 1605 | |
| }, | |
| { | |
| "epoch": 0.3032050753893054, | |
| "grad_norm": 7.844501495361328, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1812, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 0.3041467060582163, | |
| "grad_norm": 6.8861308097839355, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0092, | |
| "step": 1615 | |
| }, | |
| { | |
| "epoch": 0.3050883367271272, | |
| "grad_norm": 7.58292818069458, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1616, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.3060299673960381, | |
| "grad_norm": 7.547516822814941, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1064, | |
| "step": 1625 | |
| }, | |
| { | |
| "epoch": 0.306971598064949, | |
| "grad_norm": 7.924961090087891, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1486, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 0.30791322873385985, | |
| "grad_norm": 9.10555362701416, | |
| "learning_rate": 1e-05, | |
| "loss": 2.995, | |
| "step": 1635 | |
| }, | |
| { | |
| "epoch": 0.30885485940277074, | |
| "grad_norm": 7.4836931228637695, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1971, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.30979649007168164, | |
| "grad_norm": 9.865035057067871, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1604, | |
| "step": 1645 | |
| }, | |
| { | |
| "epoch": 0.31073812074059254, | |
| "grad_norm": 8.200763702392578, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0831, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.3116797514095034, | |
| "grad_norm": 8.044021606445312, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1665, | |
| "step": 1655 | |
| }, | |
| { | |
| "epoch": 0.3126213820784143, | |
| "grad_norm": 7.945106029510498, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0876, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.3135630127473252, | |
| "grad_norm": 8.18039608001709, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1172, | |
| "step": 1665 | |
| }, | |
| { | |
| "epoch": 0.3145046434162361, | |
| "grad_norm": 7.830636024475098, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1627, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 0.31544627408514697, | |
| "grad_norm": 8.153444290161133, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0749, | |
| "step": 1675 | |
| }, | |
| { | |
| "epoch": 0.3163879047540578, | |
| "grad_norm": 8.141214370727539, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0414, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.3173295354229687, | |
| "grad_norm": 7.207879543304443, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1829, | |
| "step": 1685 | |
| }, | |
| { | |
| "epoch": 0.3182711660918796, | |
| "grad_norm": 6.900830268859863, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0745, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 0.3192127967607905, | |
| "grad_norm": 7.9972243309021, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1691, | |
| "step": 1695 | |
| }, | |
| { | |
| "epoch": 0.3201544274297014, | |
| "grad_norm": 8.405791282653809, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0937, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.32109605809861225, | |
| "grad_norm": 7.26245641708374, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0187, | |
| "step": 1705 | |
| }, | |
| { | |
| "epoch": 0.32203768876752314, | |
| "grad_norm": 8.00014591217041, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0608, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 0.32297931943643404, | |
| "grad_norm": 7.747166156768799, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1721, | |
| "step": 1715 | |
| }, | |
| { | |
| "epoch": 0.32392095010534494, | |
| "grad_norm": 8.84945011138916, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0493, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 0.32486258077425584, | |
| "grad_norm": 8.65839958190918, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0396, | |
| "step": 1725 | |
| }, | |
| { | |
| "epoch": 0.3258042114431667, | |
| "grad_norm": 8.27173900604248, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0956, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 0.3267458421120776, | |
| "grad_norm": 7.819243907928467, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1334, | |
| "step": 1735 | |
| }, | |
| { | |
| "epoch": 0.3276874727809885, | |
| "grad_norm": 8.155767440795898, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1312, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 0.3286291034498994, | |
| "grad_norm": 7.6382155418396, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9995, | |
| "step": 1745 | |
| }, | |
| { | |
| "epoch": 0.32957073411881027, | |
| "grad_norm": 7.503223896026611, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1386, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.3305123647877211, | |
| "grad_norm": 8.2285737991333, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0442, | |
| "step": 1755 | |
| }, | |
| { | |
| "epoch": 0.331453995456632, | |
| "grad_norm": 7.820169925689697, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0377, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 0.3323956261255429, | |
| "grad_norm": 7.6100969314575195, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1311, | |
| "step": 1765 | |
| }, | |
| { | |
| "epoch": 0.3333372567944538, | |
| "grad_norm": 7.763628005981445, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0599, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 0.3342788874633647, | |
| "grad_norm": 8.461652755737305, | |
| "learning_rate": 1e-05, | |
| "loss": 3.222, | |
| "step": 1775 | |
| }, | |
| { | |
| "epoch": 0.33522051813227555, | |
| "grad_norm": 7.414519309997559, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0587, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 0.33616214880118644, | |
| "grad_norm": 7.964784145355225, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2038, | |
| "step": 1785 | |
| }, | |
| { | |
| "epoch": 0.33710377947009734, | |
| "grad_norm": 7.656503200531006, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1324, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 0.33804541013900824, | |
| "grad_norm": 8.035988807678223, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1101, | |
| "step": 1795 | |
| }, | |
| { | |
| "epoch": 0.33898704080791914, | |
| "grad_norm": 7.202066898345947, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0915, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.33992867147683, | |
| "grad_norm": 8.242351531982422, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2059, | |
| "step": 1805 | |
| }, | |
| { | |
| "epoch": 0.3408703021457409, | |
| "grad_norm": 7.794888019561768, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1364, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 0.3418119328146518, | |
| "grad_norm": 7.34774923324585, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0507, | |
| "step": 1815 | |
| }, | |
| { | |
| "epoch": 0.3427535634835627, | |
| "grad_norm": 7.673720359802246, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0383, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 0.34369519415247357, | |
| "grad_norm": 7.948644638061523, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1888, | |
| "step": 1825 | |
| }, | |
| { | |
| "epoch": 0.3446368248213844, | |
| "grad_norm": 7.747712135314941, | |
| "learning_rate": 1e-05, | |
| "loss": 3.089, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 0.3455784554902953, | |
| "grad_norm": 7.607177257537842, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1111, | |
| "step": 1835 | |
| }, | |
| { | |
| "epoch": 0.3465200861592062, | |
| "grad_norm": 6.7083916664123535, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0944, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 0.3474617168281171, | |
| "grad_norm": 8.02619743347168, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1129, | |
| "step": 1845 | |
| }, | |
| { | |
| "epoch": 0.348403347497028, | |
| "grad_norm": 8.047721862792969, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1843, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 0.34934497816593885, | |
| "grad_norm": 7.408081531524658, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0514, | |
| "step": 1855 | |
| }, | |
| { | |
| "epoch": 0.35028660883484974, | |
| "grad_norm": 7.935153484344482, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1919, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 0.35122823950376064, | |
| "grad_norm": 7.98247766494751, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1084, | |
| "step": 1865 | |
| }, | |
| { | |
| "epoch": 0.35216987017267154, | |
| "grad_norm": 6.453451156616211, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0292, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 0.35311150084158244, | |
| "grad_norm": 7.293237686157227, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0266, | |
| "step": 1875 | |
| }, | |
| { | |
| "epoch": 0.3540531315104933, | |
| "grad_norm": 7.254806041717529, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1074, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 0.3549947621794042, | |
| "grad_norm": 7.4231743812561035, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1157, | |
| "step": 1885 | |
| }, | |
| { | |
| "epoch": 0.3559363928483151, | |
| "grad_norm": 7.126735210418701, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0207, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 0.356878023517226, | |
| "grad_norm": 7.252379417419434, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1247, | |
| "step": 1895 | |
| }, | |
| { | |
| "epoch": 0.3578196541861368, | |
| "grad_norm": 8.155769348144531, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1562, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.3587612848550477, | |
| "grad_norm": 7.091341972351074, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0676, | |
| "step": 1905 | |
| }, | |
| { | |
| "epoch": 0.3597029155239586, | |
| "grad_norm": 7.706187725067139, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1094, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 0.3606445461928695, | |
| "grad_norm": 7.110264301300049, | |
| "learning_rate": 1e-05, | |
| "loss": 3.12, | |
| "step": 1915 | |
| }, | |
| { | |
| "epoch": 0.3615861768617804, | |
| "grad_norm": 6.751636505126953, | |
| "learning_rate": 1e-05, | |
| "loss": 3.189, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 0.36252780753069125, | |
| "grad_norm": 7.525967597961426, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1649, | |
| "step": 1925 | |
| }, | |
| { | |
| "epoch": 0.36346943819960215, | |
| "grad_norm": 7.6558356285095215, | |
| "learning_rate": 1e-05, | |
| "loss": 3.103, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 0.36441106886851304, | |
| "grad_norm": 7.773464679718018, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1464, | |
| "step": 1935 | |
| }, | |
| { | |
| "epoch": 0.36535269953742394, | |
| "grad_norm": 8.413071632385254, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0434, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 0.36629433020633484, | |
| "grad_norm": 9.84329605102539, | |
| "learning_rate": 1e-05, | |
| "loss": 3.077, | |
| "step": 1945 | |
| }, | |
| { | |
| "epoch": 0.3672359608752457, | |
| "grad_norm": 7.708522319793701, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0714, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 0.3681775915441566, | |
| "grad_norm": 7.770474910736084, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1796, | |
| "step": 1955 | |
| }, | |
| { | |
| "epoch": 0.3691192222130675, | |
| "grad_norm": 6.997617244720459, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1673, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 0.3700608528819784, | |
| "grad_norm": 7.298404216766357, | |
| "learning_rate": 1e-05, | |
| "loss": 2.955, | |
| "step": 1965 | |
| }, | |
| { | |
| "epoch": 0.3710024835508893, | |
| "grad_norm": 7.35360050201416, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9934, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 0.3719441142198001, | |
| "grad_norm": 7.945688247680664, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0656, | |
| "step": 1975 | |
| }, | |
| { | |
| "epoch": 0.372885744888711, | |
| "grad_norm": 8.263771057128906, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 0.3738273755576219, | |
| "grad_norm": 7.973668575286865, | |
| "learning_rate": 1e-05, | |
| "loss": 3.129, | |
| "step": 1985 | |
| }, | |
| { | |
| "epoch": 0.3747690062265328, | |
| "grad_norm": 7.491722583770752, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1392, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 0.3757106368954437, | |
| "grad_norm": 7.867580413818359, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1261, | |
| "step": 1995 | |
| }, | |
| { | |
| "epoch": 0.37665226756435455, | |
| "grad_norm": 7.192398548126221, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1165, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.37665226756435455, | |
| "eval_accuracy": 0.40667947257976045, | |
| "eval_loss": 3.105938673019409, | |
| "eval_runtime": 1037.0565, | |
| "eval_samples_per_second": 36.411, | |
| "eval_steps_per_second": 9.103, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.37759389823326545, | |
| "grad_norm": 7.978450775146484, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0121, | |
| "step": 2005 | |
| }, | |
| { | |
| "epoch": 0.37853552890217634, | |
| "grad_norm": 7.5898847579956055, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0836, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 0.37947715957108724, | |
| "grad_norm": 7.317259311676025, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0859, | |
| "step": 2015 | |
| }, | |
| { | |
| "epoch": 0.38041879023999814, | |
| "grad_norm": 7.094460964202881, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0992, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 0.381360420908909, | |
| "grad_norm": 7.557990074157715, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0912, | |
| "step": 2025 | |
| }, | |
| { | |
| "epoch": 0.3823020515778199, | |
| "grad_norm": 6.940727710723877, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0669, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 0.3832436822467308, | |
| "grad_norm": 8.001688957214355, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1704, | |
| "step": 2035 | |
| }, | |
| { | |
| "epoch": 0.3841853129156417, | |
| "grad_norm": 7.38444185256958, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0961, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 0.3851269435845526, | |
| "grad_norm": 6.947928428649902, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1942, | |
| "step": 2045 | |
| }, | |
| { | |
| "epoch": 0.3860685742534634, | |
| "grad_norm": 7.699880123138428, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0334, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 0.3870102049223743, | |
| "grad_norm": 7.490096092224121, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1301, | |
| "step": 2055 | |
| }, | |
| { | |
| "epoch": 0.3879518355912852, | |
| "grad_norm": 7.5343451499938965, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0588, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 0.3888934662601961, | |
| "grad_norm": 9.956011772155762, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0914, | |
| "step": 2065 | |
| }, | |
| { | |
| "epoch": 0.389835096929107, | |
| "grad_norm": 7.387354373931885, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1279, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 0.39077672759801785, | |
| "grad_norm": 7.3654351234436035, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0307, | |
| "step": 2075 | |
| }, | |
| { | |
| "epoch": 0.39171835826692875, | |
| "grad_norm": 7.671773910522461, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0223, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 0.39265998893583964, | |
| "grad_norm": 8.229500770568848, | |
| "learning_rate": 1e-05, | |
| "loss": 3.107, | |
| "step": 2085 | |
| }, | |
| { | |
| "epoch": 0.39360161960475054, | |
| "grad_norm": 7.7551798820495605, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1048, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 0.39454325027366144, | |
| "grad_norm": 6.759220600128174, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0824, | |
| "step": 2095 | |
| }, | |
| { | |
| "epoch": 0.3954848809425723, | |
| "grad_norm": 8.242779731750488, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0604, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.3964265116114832, | |
| "grad_norm": 7.663089752197266, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1625, | |
| "step": 2105 | |
| }, | |
| { | |
| "epoch": 0.3973681422803941, | |
| "grad_norm": 8.07699203491211, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0494, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 0.398309772949305, | |
| "grad_norm": 7.712186336517334, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1758, | |
| "step": 2115 | |
| }, | |
| { | |
| "epoch": 0.39925140361821587, | |
| "grad_norm": 9.279346466064453, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1234, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 0.4001930342871267, | |
| "grad_norm": 6.985004901885986, | |
| "learning_rate": 1e-05, | |
| "loss": 3.013, | |
| "step": 2125 | |
| }, | |
| { | |
| "epoch": 0.4011346649560376, | |
| "grad_norm": 7.505834102630615, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1199, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 0.4020762956249485, | |
| "grad_norm": 7.714130401611328, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0762, | |
| "step": 2135 | |
| }, | |
| { | |
| "epoch": 0.4030179262938594, | |
| "grad_norm": 7.358199119567871, | |
| "learning_rate": 1e-05, | |
| "loss": 3.053, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 0.40395955696277025, | |
| "grad_norm": 7.569924354553223, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1715, | |
| "step": 2145 | |
| }, | |
| { | |
| "epoch": 0.40490118763168115, | |
| "grad_norm": 7.3484697341918945, | |
| "learning_rate": 1e-05, | |
| "loss": 3.093, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 0.40584281830059205, | |
| "grad_norm": 7.914135456085205, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0454, | |
| "step": 2155 | |
| }, | |
| { | |
| "epoch": 0.40678444896950294, | |
| "grad_norm": 7.99644660949707, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1014, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 0.40772607963841384, | |
| "grad_norm": 8.003348350524902, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1274, | |
| "step": 2165 | |
| }, | |
| { | |
| "epoch": 0.4086677103073247, | |
| "grad_norm": 7.18101692199707, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0354, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 0.4096093409762356, | |
| "grad_norm": 7.236114501953125, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0585, | |
| "step": 2175 | |
| }, | |
| { | |
| "epoch": 0.4105509716451465, | |
| "grad_norm": 7.164278984069824, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0792, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 0.4114926023140574, | |
| "grad_norm": 7.827737808227539, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1141, | |
| "step": 2185 | |
| }, | |
| { | |
| "epoch": 0.4124342329829683, | |
| "grad_norm": 10.366262435913086, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0372, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 0.4133758636518791, | |
| "grad_norm": 8.009645462036133, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0653, | |
| "step": 2195 | |
| }, | |
| { | |
| "epoch": 0.41431749432079, | |
| "grad_norm": 8.994948387145996, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1723, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.4152591249897009, | |
| "grad_norm": 6.850546360015869, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1088, | |
| "step": 2205 | |
| }, | |
| { | |
| "epoch": 0.4162007556586118, | |
| "grad_norm": 7.697965145111084, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1864, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 0.4171423863275227, | |
| "grad_norm": 7.226153373718262, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1957, | |
| "step": 2215 | |
| }, | |
| { | |
| "epoch": 0.41808401699643355, | |
| "grad_norm": 7.277571678161621, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1946, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 0.41902564766534445, | |
| "grad_norm": 7.740057945251465, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0261, | |
| "step": 2225 | |
| }, | |
| { | |
| "epoch": 0.41996727833425535, | |
| "grad_norm": 7.6060028076171875, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0848, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 0.42090890900316624, | |
| "grad_norm": 8.158476829528809, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0151, | |
| "step": 2235 | |
| }, | |
| { | |
| "epoch": 0.42185053967207714, | |
| "grad_norm": 7.340221405029297, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9477, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 0.422792170340988, | |
| "grad_norm": 6.894491195678711, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0536, | |
| "step": 2245 | |
| }, | |
| { | |
| "epoch": 0.4237338010098989, | |
| "grad_norm": 7.707582473754883, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0879, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 0.4246754316788098, | |
| "grad_norm": 7.966876029968262, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0493, | |
| "step": 2255 | |
| }, | |
| { | |
| "epoch": 0.4256170623477207, | |
| "grad_norm": 7.2440185546875, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9738, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 0.4265586930166316, | |
| "grad_norm": 7.6277337074279785, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2125, | |
| "step": 2265 | |
| }, | |
| { | |
| "epoch": 0.4275003236855424, | |
| "grad_norm": 7.30360746383667, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1131, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 0.4284419543544533, | |
| "grad_norm": 7.502569675445557, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0646, | |
| "step": 2275 | |
| }, | |
| { | |
| "epoch": 0.4293835850233642, | |
| "grad_norm": 7.918311595916748, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1952, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 0.4303252156922751, | |
| "grad_norm": 7.016907691955566, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1168, | |
| "step": 2285 | |
| }, | |
| { | |
| "epoch": 0.431266846361186, | |
| "grad_norm": 7.167459964752197, | |
| "learning_rate": 1e-05, | |
| "loss": 3.126, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 0.43220847703009685, | |
| "grad_norm": 7.548356533050537, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0791, | |
| "step": 2295 | |
| }, | |
| { | |
| "epoch": 0.43315010769900775, | |
| "grad_norm": 6.640120506286621, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1876, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.43409173836791864, | |
| "grad_norm": 7.116699695587158, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0427, | |
| "step": 2305 | |
| }, | |
| { | |
| "epoch": 0.43503336903682954, | |
| "grad_norm": 7.0854668617248535, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1455, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 0.43597499970574044, | |
| "grad_norm": 7.1804423332214355, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0863, | |
| "step": 2315 | |
| }, | |
| { | |
| "epoch": 0.4369166303746513, | |
| "grad_norm": 7.1510210037231445, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1586, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 0.4378582610435622, | |
| "grad_norm": 7.871683120727539, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1336, | |
| "step": 2325 | |
| }, | |
| { | |
| "epoch": 0.4387998917124731, | |
| "grad_norm": 7.599919319152832, | |
| "learning_rate": 1e-05, | |
| "loss": 3.05, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 0.439741522381384, | |
| "grad_norm": 8.056229591369629, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0782, | |
| "step": 2335 | |
| }, | |
| { | |
| "epoch": 0.4406831530502949, | |
| "grad_norm": 7.520589351654053, | |
| "learning_rate": 1e-05, | |
| "loss": 3.113, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 0.4416247837192057, | |
| "grad_norm": 6.8444952964782715, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0913, | |
| "step": 2345 | |
| }, | |
| { | |
| "epoch": 0.4425664143881166, | |
| "grad_norm": 7.817233562469482, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1279, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 0.4435080450570275, | |
| "grad_norm": 6.604560375213623, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1026, | |
| "step": 2355 | |
| }, | |
| { | |
| "epoch": 0.4444496757259384, | |
| "grad_norm": 7.087409496307373, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1523, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 0.4453913063948493, | |
| "grad_norm": 7.123996257781982, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0798, | |
| "step": 2365 | |
| }, | |
| { | |
| "epoch": 0.44633293706376015, | |
| "grad_norm": 7.211167812347412, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1813, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 0.44727456773267105, | |
| "grad_norm": 7.160671710968018, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0597, | |
| "step": 2375 | |
| }, | |
| { | |
| "epoch": 0.44821619840158194, | |
| "grad_norm": 7.699546813964844, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0837, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 0.44915782907049284, | |
| "grad_norm": 7.012664794921875, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9755, | |
| "step": 2385 | |
| }, | |
| { | |
| "epoch": 0.45009945973940374, | |
| "grad_norm": 9.324268341064453, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0906, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 0.4510410904083146, | |
| "grad_norm": 7.166225433349609, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0489, | |
| "step": 2395 | |
| }, | |
| { | |
| "epoch": 0.4519827210772255, | |
| "grad_norm": 7.982870578765869, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1248, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.4529243517461364, | |
| "grad_norm": 7.513890743255615, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1008, | |
| "step": 2405 | |
| }, | |
| { | |
| "epoch": 0.4538659824150473, | |
| "grad_norm": 6.794632911682129, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1117, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 0.4548076130839581, | |
| "grad_norm": 8.143996238708496, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1321, | |
| "step": 2415 | |
| }, | |
| { | |
| "epoch": 0.455749243752869, | |
| "grad_norm": 7.631880283355713, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1185, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 0.4566908744217799, | |
| "grad_norm": 7.6734466552734375, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0592, | |
| "step": 2425 | |
| }, | |
| { | |
| "epoch": 0.4576325050906908, | |
| "grad_norm": 6.934384346008301, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1555, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 0.4585741357596017, | |
| "grad_norm": 7.925302505493164, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0891, | |
| "step": 2435 | |
| }, | |
| { | |
| "epoch": 0.45951576642851255, | |
| "grad_norm": 7.213337421417236, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9779, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 0.46045739709742345, | |
| "grad_norm": 7.504617691040039, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0613, | |
| "step": 2445 | |
| }, | |
| { | |
| "epoch": 0.46139902776633435, | |
| "grad_norm": 7.625884056091309, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9949, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 0.46234065843524524, | |
| "grad_norm": 7.4358696937561035, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1089, | |
| "step": 2455 | |
| }, | |
| { | |
| "epoch": 0.46328228910415614, | |
| "grad_norm": 7.1684489250183105, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9965, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 0.464223919773067, | |
| "grad_norm": 7.672368049621582, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0439, | |
| "step": 2465 | |
| }, | |
| { | |
| "epoch": 0.4651655504419779, | |
| "grad_norm": 7.329806327819824, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0025, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 0.4661071811108888, | |
| "grad_norm": 7.104738235473633, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0738, | |
| "step": 2475 | |
| }, | |
| { | |
| "epoch": 0.4670488117797997, | |
| "grad_norm": 6.561854839324951, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9749, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 0.4679904424487106, | |
| "grad_norm": 7.665432453155518, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0647, | |
| "step": 2485 | |
| }, | |
| { | |
| "epoch": 0.4689320731176214, | |
| "grad_norm": 7.229370594024658, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0231, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 0.4698737037865323, | |
| "grad_norm": 7.576310157775879, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0132, | |
| "step": 2495 | |
| }, | |
| { | |
| "epoch": 0.4708153344554432, | |
| "grad_norm": 6.800790786743164, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0343, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.4717569651243541, | |
| "grad_norm": 7.0642266273498535, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0681, | |
| "step": 2505 | |
| }, | |
| { | |
| "epoch": 0.472698595793265, | |
| "grad_norm": 6.809929370880127, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1242, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 0.47364022646217585, | |
| "grad_norm": 7.387331008911133, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0638, | |
| "step": 2515 | |
| }, | |
| { | |
| "epoch": 0.47458185713108675, | |
| "grad_norm": 7.186639308929443, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0472, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 0.47552348779999765, | |
| "grad_norm": 7.330045223236084, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1256, | |
| "step": 2525 | |
| }, | |
| { | |
| "epoch": 0.47646511846890854, | |
| "grad_norm": 7.658766746520996, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0334, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 0.47740674913781944, | |
| "grad_norm": 7.821640968322754, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1666, | |
| "step": 2535 | |
| }, | |
| { | |
| "epoch": 0.4783483798067303, | |
| "grad_norm": 8.806328773498535, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1129, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 0.4792900104756412, | |
| "grad_norm": 7.119672775268555, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0084, | |
| "step": 2545 | |
| }, | |
| { | |
| "epoch": 0.4802316411445521, | |
| "grad_norm": 7.20904541015625, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0855, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 0.481173271813463, | |
| "grad_norm": 7.136516571044922, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9989, | |
| "step": 2555 | |
| }, | |
| { | |
| "epoch": 0.4821149024823739, | |
| "grad_norm": 6.163409233093262, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9992, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 0.4830565331512847, | |
| "grad_norm": 7.667194843292236, | |
| "learning_rate": 1e-05, | |
| "loss": 2.996, | |
| "step": 2565 | |
| }, | |
| { | |
| "epoch": 0.4839981638201956, | |
| "grad_norm": 7.100743293762207, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0643, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 0.4849397944891065, | |
| "grad_norm": 7.50702428817749, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1647, | |
| "step": 2575 | |
| }, | |
| { | |
| "epoch": 0.4858814251580174, | |
| "grad_norm": 7.247680187225342, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0211, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 0.4868230558269283, | |
| "grad_norm": 7.403180122375488, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0319, | |
| "step": 2585 | |
| }, | |
| { | |
| "epoch": 0.48776468649583915, | |
| "grad_norm": 7.5679473876953125, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0748, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 0.48870631716475005, | |
| "grad_norm": 7.389653205871582, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0973, | |
| "step": 2595 | |
| }, | |
| { | |
| "epoch": 0.48964794783366095, | |
| "grad_norm": 12.070070266723633, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0398, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.49058957850257184, | |
| "grad_norm": 7.190162658691406, | |
| "learning_rate": 1e-05, | |
| "loss": 3.16, | |
| "step": 2605 | |
| }, | |
| { | |
| "epoch": 0.49153120917148274, | |
| "grad_norm": 6.753963470458984, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0305, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 0.4924728398403936, | |
| "grad_norm": 7.3352789878845215, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0226, | |
| "step": 2615 | |
| }, | |
| { | |
| "epoch": 0.4934144705093045, | |
| "grad_norm": 6.455969333648682, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0626, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 0.4943561011782154, | |
| "grad_norm": 6.789318561553955, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1405, | |
| "step": 2625 | |
| }, | |
| { | |
| "epoch": 0.4952977318471263, | |
| "grad_norm": 7.1296186447143555, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9996, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 0.4962393625160372, | |
| "grad_norm": 7.188891887664795, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0108, | |
| "step": 2635 | |
| }, | |
| { | |
| "epoch": 0.497180993184948, | |
| "grad_norm": 8.604632377624512, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1121, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 0.4981226238538589, | |
| "grad_norm": 8.063715934753418, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1481, | |
| "step": 2645 | |
| }, | |
| { | |
| "epoch": 0.4990642545227698, | |
| "grad_norm": 7.208702087402344, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1499, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 0.5000058851916807, | |
| "grad_norm": 6.898314952850342, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0624, | |
| "step": 2655 | |
| }, | |
| { | |
| "epoch": 0.5009475158605916, | |
| "grad_norm": 7.062436103820801, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9495, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 0.5018891465295025, | |
| "grad_norm": 7.108069896697998, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9833, | |
| "step": 2665 | |
| }, | |
| { | |
| "epoch": 0.5028307771984134, | |
| "grad_norm": 8.10451602935791, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1446, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 0.5037724078673242, | |
| "grad_norm": 7.409706115722656, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1812, | |
| "step": 2675 | |
| }, | |
| { | |
| "epoch": 0.5047140385362351, | |
| "grad_norm": 7.078268527984619, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0981, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 0.505655669205146, | |
| "grad_norm": 6.625565052032471, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0404, | |
| "step": 2685 | |
| }, | |
| { | |
| "epoch": 0.5065972998740569, | |
| "grad_norm": 7.314438819885254, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0807, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 0.5075389305429678, | |
| "grad_norm": 6.884581565856934, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0895, | |
| "step": 2695 | |
| }, | |
| { | |
| "epoch": 0.5084805612118787, | |
| "grad_norm": 7.702692031860352, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0731, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.5094221918807895, | |
| "grad_norm": 7.680057525634766, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0837, | |
| "step": 2705 | |
| }, | |
| { | |
| "epoch": 0.5103638225497005, | |
| "grad_norm": 7.11198616027832, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9995, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 0.5113054532186113, | |
| "grad_norm": 7.569001197814941, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1655, | |
| "step": 2715 | |
| }, | |
| { | |
| "epoch": 0.5122470838875223, | |
| "grad_norm": 7.120611667633057, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0245, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 0.5131887145564331, | |
| "grad_norm": 7.674874305725098, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0993, | |
| "step": 2725 | |
| }, | |
| { | |
| "epoch": 0.514130345225344, | |
| "grad_norm": 7.820282459259033, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0897, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 0.5150719758942549, | |
| "grad_norm": 7.010729789733887, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9622, | |
| "step": 2735 | |
| }, | |
| { | |
| "epoch": 0.5160136065631657, | |
| "grad_norm": 7.443545818328857, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1474, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 0.5169552372320767, | |
| "grad_norm": 8.485078811645508, | |
| "learning_rate": 1e-05, | |
| "loss": 3.095, | |
| "step": 2745 | |
| }, | |
| { | |
| "epoch": 0.5178968679009875, | |
| "grad_norm": 7.487083435058594, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1509, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 0.5188384985698984, | |
| "grad_norm": 7.581494331359863, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0771, | |
| "step": 2755 | |
| }, | |
| { | |
| "epoch": 0.5197801292388093, | |
| "grad_norm": 7.041471481323242, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9578, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 0.5207217599077202, | |
| "grad_norm": 6.6093034744262695, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0864, | |
| "step": 2765 | |
| }, | |
| { | |
| "epoch": 0.5216633905766311, | |
| "grad_norm": 6.595880031585693, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9654, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 0.522605021245542, | |
| "grad_norm": 7.0556511878967285, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9616, | |
| "step": 2775 | |
| }, | |
| { | |
| "epoch": 0.5235466519144528, | |
| "grad_norm": 7.266999244689941, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1041, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 0.5244882825833638, | |
| "grad_norm": 7.564467430114746, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1307, | |
| "step": 2785 | |
| }, | |
| { | |
| "epoch": 0.5254299132522746, | |
| "grad_norm": 6.620129108428955, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0366, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 0.5263715439211855, | |
| "grad_norm": 6.7112321853637695, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0846, | |
| "step": 2795 | |
| }, | |
| { | |
| "epoch": 0.5273131745900964, | |
| "grad_norm": 8.041632652282715, | |
| "learning_rate": 1e-05, | |
| "loss": 3.161, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.5282548052590073, | |
| "grad_norm": 7.796072006225586, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0004, | |
| "step": 2805 | |
| }, | |
| { | |
| "epoch": 0.5291964359279182, | |
| "grad_norm": 7.121657371520996, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0108, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 0.530138066596829, | |
| "grad_norm": 7.073033809661865, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0985, | |
| "step": 2815 | |
| }, | |
| { | |
| "epoch": 0.5310796972657399, | |
| "grad_norm": 6.580442905426025, | |
| "learning_rate": 1e-05, | |
| "loss": 3.081, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 0.5320213279346508, | |
| "grad_norm": 7.59156608581543, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1838, | |
| "step": 2825 | |
| }, | |
| { | |
| "epoch": 0.5329629586035617, | |
| "grad_norm": 6.721426963806152, | |
| "learning_rate": 1e-05, | |
| "loss": 3.12, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 0.5339045892724726, | |
| "grad_norm": 6.488240718841553, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0373, | |
| "step": 2835 | |
| }, | |
| { | |
| "epoch": 0.5348462199413835, | |
| "grad_norm": 7.797455787658691, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0716, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 0.5357878506102943, | |
| "grad_norm": 7.475494861602783, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9321, | |
| "step": 2845 | |
| }, | |
| { | |
| "epoch": 0.5367294812792053, | |
| "grad_norm": 7.374056339263916, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1087, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 0.5376711119481161, | |
| "grad_norm": 6.967336654663086, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0788, | |
| "step": 2855 | |
| }, | |
| { | |
| "epoch": 0.5386127426170271, | |
| "grad_norm": 7.319347858428955, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1399, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 0.5395543732859379, | |
| "grad_norm": 7.5937957763671875, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0926, | |
| "step": 2865 | |
| }, | |
| { | |
| "epoch": 0.5404960039548488, | |
| "grad_norm": 7.879610538482666, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0853, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 0.5414376346237597, | |
| "grad_norm": 7.16407585144043, | |
| "learning_rate": 1e-05, | |
| "loss": 2.8899, | |
| "step": 2875 | |
| }, | |
| { | |
| "epoch": 0.5423792652926706, | |
| "grad_norm": 7.709092617034912, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1303, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 0.5433208959615815, | |
| "grad_norm": 7.1035356521606445, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1267, | |
| "step": 2885 | |
| }, | |
| { | |
| "epoch": 0.5442625266304923, | |
| "grad_norm": 6.998748302459717, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1023, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 0.5452041572994032, | |
| "grad_norm": 7.0011186599731445, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1474, | |
| "step": 2895 | |
| }, | |
| { | |
| "epoch": 0.5461457879683141, | |
| "grad_norm": 6.540005207061768, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9699, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.547087418637225, | |
| "grad_norm": 7.625404357910156, | |
| "learning_rate": 1e-05, | |
| "loss": 3.2087, | |
| "step": 2905 | |
| }, | |
| { | |
| "epoch": 0.5480290493061359, | |
| "grad_norm": 8.491080284118652, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0275, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 0.5489706799750468, | |
| "grad_norm": 7.326952934265137, | |
| "learning_rate": 1e-05, | |
| "loss": 3.119, | |
| "step": 2915 | |
| }, | |
| { | |
| "epoch": 0.5499123106439576, | |
| "grad_norm": 6.557469367980957, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0491, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 0.5508539413128686, | |
| "grad_norm": 8.412656784057617, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9752, | |
| "step": 2925 | |
| }, | |
| { | |
| "epoch": 0.5517955719817794, | |
| "grad_norm": 7.805161476135254, | |
| "learning_rate": 1e-05, | |
| "loss": 3.188, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 0.5527372026506904, | |
| "grad_norm": 7.706869602203369, | |
| "learning_rate": 1e-05, | |
| "loss": 3.01, | |
| "step": 2935 | |
| }, | |
| { | |
| "epoch": 0.5536788333196012, | |
| "grad_norm": 6.797297477722168, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9473, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 0.5546204639885121, | |
| "grad_norm": 7.4034624099731445, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0252, | |
| "step": 2945 | |
| }, | |
| { | |
| "epoch": 0.555562094657423, | |
| "grad_norm": 8.384190559387207, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0013, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 0.5565037253263339, | |
| "grad_norm": 7.6911187171936035, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1842, | |
| "step": 2955 | |
| }, | |
| { | |
| "epoch": 0.5574453559952448, | |
| "grad_norm": 6.948124408721924, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9562, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 0.5583869866641556, | |
| "grad_norm": 7.233328819274902, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0687, | |
| "step": 2965 | |
| }, | |
| { | |
| "epoch": 0.5593286173330665, | |
| "grad_norm": 7.2016425132751465, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0819, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 0.5602702480019774, | |
| "grad_norm": 6.626002311706543, | |
| "learning_rate": 1e-05, | |
| "loss": 3.078, | |
| "step": 2975 | |
| }, | |
| { | |
| "epoch": 0.5612118786708883, | |
| "grad_norm": 7.238675117492676, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1172, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 0.5621535093397992, | |
| "grad_norm": 6.863603115081787, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0753, | |
| "step": 2985 | |
| }, | |
| { | |
| "epoch": 0.5630951400087101, | |
| "grad_norm": 7.731668949127197, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9837, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 0.5640367706776209, | |
| "grad_norm": 6.4692702293396, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1187, | |
| "step": 2995 | |
| }, | |
| { | |
| "epoch": 0.5649784013465319, | |
| "grad_norm": 7.253861427307129, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9771, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.5649784013465319, | |
| "eval_accuracy": 0.4138860273852878, | |
| "eval_loss": 3.0637450218200684, | |
| "eval_runtime": 1049.1644, | |
| "eval_samples_per_second": 35.991, | |
| "eval_steps_per_second": 8.998, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.5659200320154427, | |
| "grad_norm": 7.105886459350586, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0877, | |
| "step": 3005 | |
| }, | |
| { | |
| "epoch": 0.5668616626843537, | |
| "grad_norm": 6.724785327911377, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0216, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 0.5678032933532645, | |
| "grad_norm": 6.603315353393555, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1276, | |
| "step": 3015 | |
| }, | |
| { | |
| "epoch": 0.5687449240221754, | |
| "grad_norm": 6.701154708862305, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9892, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 0.5696865546910863, | |
| "grad_norm": 6.806232929229736, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0323, | |
| "step": 3025 | |
| }, | |
| { | |
| "epoch": 0.5706281853599972, | |
| "grad_norm": 7.004578113555908, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0609, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 0.5715698160289081, | |
| "grad_norm": 6.511326789855957, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1092, | |
| "step": 3035 | |
| }, | |
| { | |
| "epoch": 0.572511446697819, | |
| "grad_norm": 7.045777320861816, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1217, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 0.5734530773667298, | |
| "grad_norm": 6.780973434448242, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0726, | |
| "step": 3045 | |
| }, | |
| { | |
| "epoch": 0.5743947080356407, | |
| "grad_norm": 7.139321804046631, | |
| "learning_rate": 1e-05, | |
| "loss": 3.065, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 0.5753363387045516, | |
| "grad_norm": 7.129504203796387, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1518, | |
| "step": 3055 | |
| }, | |
| { | |
| "epoch": 0.5762779693734625, | |
| "grad_norm": 8.042350769042969, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0006, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 0.5772196000423734, | |
| "grad_norm": 6.995969772338867, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0849, | |
| "step": 3065 | |
| }, | |
| { | |
| "epoch": 0.5781612307112842, | |
| "grad_norm": 6.369194030761719, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9541, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 0.5791028613801952, | |
| "grad_norm": 6.521533012390137, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0632, | |
| "step": 3075 | |
| }, | |
| { | |
| "epoch": 0.580044492049106, | |
| "grad_norm": 7.185681343078613, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1407, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 0.580986122718017, | |
| "grad_norm": 6.984232425689697, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1047, | |
| "step": 3085 | |
| }, | |
| { | |
| "epoch": 0.5819277533869278, | |
| "grad_norm": 7.491443634033203, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0165, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 0.5828693840558387, | |
| "grad_norm": 6.652406215667725, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1079, | |
| "step": 3095 | |
| }, | |
| { | |
| "epoch": 0.5838110147247496, | |
| "grad_norm": 7.801830291748047, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1699, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.5847526453936605, | |
| "grad_norm": 7.229469299316406, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0419, | |
| "step": 3105 | |
| }, | |
| { | |
| "epoch": 0.5856942760625714, | |
| "grad_norm": 6.888019561767578, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9814, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 0.5866359067314822, | |
| "grad_norm": 7.353707790374756, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0864, | |
| "step": 3115 | |
| }, | |
| { | |
| "epoch": 0.5875775374003931, | |
| "grad_norm": 7.126093864440918, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1298, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 0.588519168069304, | |
| "grad_norm": 7.8923211097717285, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0013, | |
| "step": 3125 | |
| }, | |
| { | |
| "epoch": 0.5894607987382149, | |
| "grad_norm": 6.3733649253845215, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9838, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 0.5904024294071258, | |
| "grad_norm": 7.295039176940918, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1155, | |
| "step": 3135 | |
| }, | |
| { | |
| "epoch": 0.5913440600760367, | |
| "grad_norm": 7.227429389953613, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1303, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 0.5922856907449475, | |
| "grad_norm": 7.039018630981445, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9929, | |
| "step": 3145 | |
| }, | |
| { | |
| "epoch": 0.5932273214138585, | |
| "grad_norm": 8.034615516662598, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0979, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 0.5941689520827693, | |
| "grad_norm": 7.214796543121338, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0098, | |
| "step": 3155 | |
| }, | |
| { | |
| "epoch": 0.5951105827516803, | |
| "grad_norm": 7.179065227508545, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0168, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 0.5960522134205911, | |
| "grad_norm": 7.219829559326172, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9791, | |
| "step": 3165 | |
| }, | |
| { | |
| "epoch": 0.596993844089502, | |
| "grad_norm": 7.285015106201172, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9036, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 0.5979354747584129, | |
| "grad_norm": 8.418420791625977, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9316, | |
| "step": 3175 | |
| }, | |
| { | |
| "epoch": 0.5988771054273238, | |
| "grad_norm": 7.653911590576172, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0915, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 0.5998187360962347, | |
| "grad_norm": 6.934592247009277, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9547, | |
| "step": 3185 | |
| }, | |
| { | |
| "epoch": 0.6007603667651455, | |
| "grad_norm": 6.826767444610596, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9792, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 0.6017019974340564, | |
| "grad_norm": 7.79368782043457, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0121, | |
| "step": 3195 | |
| }, | |
| { | |
| "epoch": 0.6026436281029673, | |
| "grad_norm": 7.679154872894287, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1288, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.6035852587718782, | |
| "grad_norm": 7.1232781410217285, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0423, | |
| "step": 3205 | |
| }, | |
| { | |
| "epoch": 0.6045268894407891, | |
| "grad_norm": 7.5240983963012695, | |
| "learning_rate": 1e-05, | |
| "loss": 3.056, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 0.6054685201097, | |
| "grad_norm": 7.4613237380981445, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0777, | |
| "step": 3215 | |
| }, | |
| { | |
| "epoch": 0.6064101507786108, | |
| "grad_norm": 6.664481163024902, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0872, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 0.6073517814475218, | |
| "grad_norm": 6.648809432983398, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1206, | |
| "step": 3225 | |
| }, | |
| { | |
| "epoch": 0.6082934121164326, | |
| "grad_norm": 6.968840599060059, | |
| "learning_rate": 1e-05, | |
| "loss": 3.052, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 0.6092350427853436, | |
| "grad_norm": 7.363081932067871, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1386, | |
| "step": 3235 | |
| }, | |
| { | |
| "epoch": 0.6101766734542544, | |
| "grad_norm": 6.684266567230225, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9426, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 0.6111183041231653, | |
| "grad_norm": 7.6497650146484375, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9756, | |
| "step": 3245 | |
| }, | |
| { | |
| "epoch": 0.6120599347920762, | |
| "grad_norm": 7.052670955657959, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0383, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 0.613001565460987, | |
| "grad_norm": 6.5792083740234375, | |
| "learning_rate": 1e-05, | |
| "loss": 2.8681, | |
| "step": 3255 | |
| }, | |
| { | |
| "epoch": 0.613943196129898, | |
| "grad_norm": 7.075092315673828, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0287, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 0.6148848267988088, | |
| "grad_norm": 7.760746002197266, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0866, | |
| "step": 3265 | |
| }, | |
| { | |
| "epoch": 0.6158264574677197, | |
| "grad_norm": 7.239394664764404, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9918, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 0.6167680881366306, | |
| "grad_norm": 7.3641276359558105, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9678, | |
| "step": 3275 | |
| }, | |
| { | |
| "epoch": 0.6177097188055415, | |
| "grad_norm": 10.26201057434082, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1028, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 0.6186513494744523, | |
| "grad_norm": 6.868218421936035, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0633, | |
| "step": 3285 | |
| }, | |
| { | |
| "epoch": 0.6195929801433633, | |
| "grad_norm": 6.771929740905762, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0126, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 0.6205346108122741, | |
| "grad_norm": 7.173276901245117, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0549, | |
| "step": 3295 | |
| }, | |
| { | |
| "epoch": 0.6214762414811851, | |
| "grad_norm": 6.571986198425293, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0469, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.6224178721500959, | |
| "grad_norm": 8.061853408813477, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0797, | |
| "step": 3305 | |
| }, | |
| { | |
| "epoch": 0.6233595028190068, | |
| "grad_norm": 6.856041431427002, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0835, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 0.6243011334879177, | |
| "grad_norm": 6.356860637664795, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0321, | |
| "step": 3315 | |
| }, | |
| { | |
| "epoch": 0.6252427641568286, | |
| "grad_norm": 7.281395435333252, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9796, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 0.6261843948257395, | |
| "grad_norm": 7.561255931854248, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0791, | |
| "step": 3325 | |
| }, | |
| { | |
| "epoch": 0.6271260254946504, | |
| "grad_norm": 5.9859185218811035, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0435, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 0.6280676561635612, | |
| "grad_norm": 7.344315528869629, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0616, | |
| "step": 3335 | |
| }, | |
| { | |
| "epoch": 0.6290092868324721, | |
| "grad_norm": 7.146785736083984, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0322, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 0.629950917501383, | |
| "grad_norm": 7.084141731262207, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0367, | |
| "step": 3345 | |
| }, | |
| { | |
| "epoch": 0.6308925481702939, | |
| "grad_norm": 7.505505084991455, | |
| "learning_rate": 1e-05, | |
| "loss": 3.052, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 0.6318341788392048, | |
| "grad_norm": 7.202043533325195, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0902, | |
| "step": 3355 | |
| }, | |
| { | |
| "epoch": 0.6327758095081156, | |
| "grad_norm": 6.772159099578857, | |
| "learning_rate": 1e-05, | |
| "loss": 3.015, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 0.6337174401770266, | |
| "grad_norm": 6.495903968811035, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0254, | |
| "step": 3365 | |
| }, | |
| { | |
| "epoch": 0.6346590708459374, | |
| "grad_norm": 7.2963457107543945, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0161, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 0.6356007015148484, | |
| "grad_norm": 7.240870475769043, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0329, | |
| "step": 3375 | |
| }, | |
| { | |
| "epoch": 0.6365423321837592, | |
| "grad_norm": 6.670111656188965, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9956, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 0.6374839628526701, | |
| "grad_norm": 7.213159561157227, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9224, | |
| "step": 3385 | |
| }, | |
| { | |
| "epoch": 0.638425593521581, | |
| "grad_norm": 7.563082218170166, | |
| "learning_rate": 1e-05, | |
| "loss": 3.14, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 0.6393672241904919, | |
| "grad_norm": 7.692990779876709, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9937, | |
| "step": 3395 | |
| }, | |
| { | |
| "epoch": 0.6403088548594028, | |
| "grad_norm": 7.73374605178833, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0108, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.6412504855283137, | |
| "grad_norm": 6.7092976570129395, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0887, | |
| "step": 3405 | |
| }, | |
| { | |
| "epoch": 0.6421921161972245, | |
| "grad_norm": 7.405087471008301, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9877, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 0.6431337468661354, | |
| "grad_norm": 7.017783164978027, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0624, | |
| "step": 3415 | |
| }, | |
| { | |
| "epoch": 0.6440753775350463, | |
| "grad_norm": 6.923324108123779, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9223, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 0.6450170082039572, | |
| "grad_norm": 7.183239936828613, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1286, | |
| "step": 3425 | |
| }, | |
| { | |
| "epoch": 0.6459586388728681, | |
| "grad_norm": 7.303028106689453, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0108, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 0.6469002695417789, | |
| "grad_norm": 6.967830181121826, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0991, | |
| "step": 3435 | |
| }, | |
| { | |
| "epoch": 0.6478419002106899, | |
| "grad_norm": 7.331003189086914, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1454, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 0.6487835308796007, | |
| "grad_norm": 7.24896764755249, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9953, | |
| "step": 3445 | |
| }, | |
| { | |
| "epoch": 0.6497251615485117, | |
| "grad_norm": 8.16205883026123, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9901, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 0.6506667922174225, | |
| "grad_norm": 7.7972412109375, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1011, | |
| "step": 3455 | |
| }, | |
| { | |
| "epoch": 0.6516084228863334, | |
| "grad_norm": 6.8451714515686035, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9883, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 0.6525500535552443, | |
| "grad_norm": 6.7832465171813965, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9701, | |
| "step": 3465 | |
| }, | |
| { | |
| "epoch": 0.6534916842241552, | |
| "grad_norm": 7.421736717224121, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0205, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 0.6544333148930661, | |
| "grad_norm": 7.19790506362915, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9935, | |
| "step": 3475 | |
| }, | |
| { | |
| "epoch": 0.655374945561977, | |
| "grad_norm": 7.348209857940674, | |
| "learning_rate": 1e-05, | |
| "loss": 2.8969, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 0.6563165762308878, | |
| "grad_norm": 7.604790687561035, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0523, | |
| "step": 3485 | |
| }, | |
| { | |
| "epoch": 0.6572582068997987, | |
| "grad_norm": 6.750668525695801, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0571, | |
| "step": 3490 | |
| }, | |
| { | |
| "epoch": 0.6581998375687096, | |
| "grad_norm": 6.694459915161133, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0252, | |
| "step": 3495 | |
| }, | |
| { | |
| "epoch": 0.6591414682376205, | |
| "grad_norm": 7.250037670135498, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0203, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.6600830989065314, | |
| "grad_norm": 6.481886863708496, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1012, | |
| "step": 3505 | |
| }, | |
| { | |
| "epoch": 0.6610247295754422, | |
| "grad_norm": 7.44242525100708, | |
| "learning_rate": 1e-05, | |
| "loss": 3.031, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 0.6619663602443532, | |
| "grad_norm": 6.313452243804932, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0551, | |
| "step": 3515 | |
| }, | |
| { | |
| "epoch": 0.662907990913264, | |
| "grad_norm": 7.033621788024902, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0459, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 0.663849621582175, | |
| "grad_norm": 6.128735065460205, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9017, | |
| "step": 3525 | |
| }, | |
| { | |
| "epoch": 0.6647912522510858, | |
| "grad_norm": 6.252425670623779, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0382, | |
| "step": 3530 | |
| }, | |
| { | |
| "epoch": 0.6657328829199967, | |
| "grad_norm": 6.670257091522217, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1805, | |
| "step": 3535 | |
| }, | |
| { | |
| "epoch": 0.6666745135889076, | |
| "grad_norm": 6.33909797668457, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0568, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 0.6676161442578185, | |
| "grad_norm": 7.0296630859375, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0804, | |
| "step": 3545 | |
| }, | |
| { | |
| "epoch": 0.6685577749267294, | |
| "grad_norm": 6.852359294891357, | |
| "learning_rate": 1e-05, | |
| "loss": 3.003, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 0.6694994055956403, | |
| "grad_norm": 7.319458961486816, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9733, | |
| "step": 3555 | |
| }, | |
| { | |
| "epoch": 0.6704410362645511, | |
| "grad_norm": 7.213205337524414, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0515, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 0.671382666933462, | |
| "grad_norm": 6.743834972381592, | |
| "learning_rate": 1e-05, | |
| "loss": 2.8611, | |
| "step": 3565 | |
| }, | |
| { | |
| "epoch": 0.6723242976023729, | |
| "grad_norm": 7.25566291809082, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0633, | |
| "step": 3570 | |
| }, | |
| { | |
| "epoch": 0.6732659282712838, | |
| "grad_norm": 6.3295578956604, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0196, | |
| "step": 3575 | |
| }, | |
| { | |
| "epoch": 0.6742075589401947, | |
| "grad_norm": 7.30181884765625, | |
| "learning_rate": 1e-05, | |
| "loss": 3.052, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 0.6751491896091055, | |
| "grad_norm": 6.492465496063232, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9969, | |
| "step": 3585 | |
| }, | |
| { | |
| "epoch": 0.6760908202780165, | |
| "grad_norm": 7.620621204376221, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0297, | |
| "step": 3590 | |
| }, | |
| { | |
| "epoch": 0.6770324509469273, | |
| "grad_norm": 6.799111843109131, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9475, | |
| "step": 3595 | |
| }, | |
| { | |
| "epoch": 0.6779740816158383, | |
| "grad_norm": 7.955085277557373, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9165, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.6789157122847491, | |
| "grad_norm": 6.649621486663818, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0665, | |
| "step": 3605 | |
| }, | |
| { | |
| "epoch": 0.67985734295366, | |
| "grad_norm": 7.301611423492432, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0468, | |
| "step": 3610 | |
| }, | |
| { | |
| "epoch": 0.6807989736225709, | |
| "grad_norm": 6.720338821411133, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9307, | |
| "step": 3615 | |
| }, | |
| { | |
| "epoch": 0.6817406042914818, | |
| "grad_norm": 6.150376796722412, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0759, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 0.6826822349603927, | |
| "grad_norm": 6.534663200378418, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0509, | |
| "step": 3625 | |
| }, | |
| { | |
| "epoch": 0.6836238656293036, | |
| "grad_norm": 6.672440052032471, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9175, | |
| "step": 3630 | |
| }, | |
| { | |
| "epoch": 0.6845654962982144, | |
| "grad_norm": 6.849086761474609, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1881, | |
| "step": 3635 | |
| }, | |
| { | |
| "epoch": 0.6855071269671253, | |
| "grad_norm": 7.532481670379639, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0881, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 0.6864487576360362, | |
| "grad_norm": 7.62110710144043, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0303, | |
| "step": 3645 | |
| }, | |
| { | |
| "epoch": 0.6873903883049471, | |
| "grad_norm": 6.990483283996582, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1092, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 0.688332018973858, | |
| "grad_norm": 7.549264907836914, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0515, | |
| "step": 3655 | |
| }, | |
| { | |
| "epoch": 0.6892736496427688, | |
| "grad_norm": 7.345057010650635, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0485, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 0.6902152803116798, | |
| "grad_norm": 7.437522888183594, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0337, | |
| "step": 3665 | |
| }, | |
| { | |
| "epoch": 0.6911569109805906, | |
| "grad_norm": 6.142207145690918, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9561, | |
| "step": 3670 | |
| }, | |
| { | |
| "epoch": 0.6920985416495016, | |
| "grad_norm": 6.626426696777344, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1157, | |
| "step": 3675 | |
| }, | |
| { | |
| "epoch": 0.6930401723184124, | |
| "grad_norm": 6.62837553024292, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0824, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 0.6939818029873233, | |
| "grad_norm": 7.575382232666016, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0039, | |
| "step": 3685 | |
| }, | |
| { | |
| "epoch": 0.6949234336562342, | |
| "grad_norm": 7.333104133605957, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0786, | |
| "step": 3690 | |
| }, | |
| { | |
| "epoch": 0.695865064325145, | |
| "grad_norm": 7.4377923011779785, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9884, | |
| "step": 3695 | |
| }, | |
| { | |
| "epoch": 0.696806694994056, | |
| "grad_norm": 6.842844486236572, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9574, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.6977483256629669, | |
| "grad_norm": 6.345764636993408, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0264, | |
| "step": 3705 | |
| }, | |
| { | |
| "epoch": 0.6986899563318777, | |
| "grad_norm": 7.15539026260376, | |
| "learning_rate": 1e-05, | |
| "loss": 3.059, | |
| "step": 3710 | |
| }, | |
| { | |
| "epoch": 0.6996315870007886, | |
| "grad_norm": 7.269350528717041, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0736, | |
| "step": 3715 | |
| }, | |
| { | |
| "epoch": 0.7005732176696995, | |
| "grad_norm": 6.788912296295166, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9954, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 0.7015148483386104, | |
| "grad_norm": 6.4122185707092285, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9234, | |
| "step": 3725 | |
| }, | |
| { | |
| "epoch": 0.7024564790075213, | |
| "grad_norm": 7.263458251953125, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9974, | |
| "step": 3730 | |
| }, | |
| { | |
| "epoch": 0.7033981096764321, | |
| "grad_norm": 6.88037633895874, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0908, | |
| "step": 3735 | |
| }, | |
| { | |
| "epoch": 0.7043397403453431, | |
| "grad_norm": 7.234920978546143, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0355, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 0.7052813710142539, | |
| "grad_norm": 7.346100807189941, | |
| "learning_rate": 1e-05, | |
| "loss": 2.938, | |
| "step": 3745 | |
| }, | |
| { | |
| "epoch": 0.7062230016831649, | |
| "grad_norm": 6.825830936431885, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1014, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 0.7071646323520757, | |
| "grad_norm": 7.167881965637207, | |
| "learning_rate": 1e-05, | |
| "loss": 3.064, | |
| "step": 3755 | |
| }, | |
| { | |
| "epoch": 0.7081062630209866, | |
| "grad_norm": 6.6397294998168945, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0352, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 0.7090478936898975, | |
| "grad_norm": 7.553011894226074, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0674, | |
| "step": 3765 | |
| }, | |
| { | |
| "epoch": 0.7099895243588084, | |
| "grad_norm": 7.612958908081055, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0081, | |
| "step": 3770 | |
| }, | |
| { | |
| "epoch": 0.7109311550277193, | |
| "grad_norm": 6.634482383728027, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0408, | |
| "step": 3775 | |
| }, | |
| { | |
| "epoch": 0.7118727856966302, | |
| "grad_norm": 6.5582356452941895, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0377, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 0.712814416365541, | |
| "grad_norm": 6.804215431213379, | |
| "learning_rate": 1e-05, | |
| "loss": 3.001, | |
| "step": 3785 | |
| }, | |
| { | |
| "epoch": 0.713756047034452, | |
| "grad_norm": 7.458028316497803, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9582, | |
| "step": 3790 | |
| }, | |
| { | |
| "epoch": 0.7146976777033628, | |
| "grad_norm": 6.639705181121826, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0887, | |
| "step": 3795 | |
| }, | |
| { | |
| "epoch": 0.7156393083722736, | |
| "grad_norm": 6.617265701293945, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0967, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.7165809390411846, | |
| "grad_norm": 8.8301362991333, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0354, | |
| "step": 3805 | |
| }, | |
| { | |
| "epoch": 0.7175225697100954, | |
| "grad_norm": 6.7238264083862305, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0144, | |
| "step": 3810 | |
| }, | |
| { | |
| "epoch": 0.7184642003790064, | |
| "grad_norm": 6.8456926345825195, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9888, | |
| "step": 3815 | |
| }, | |
| { | |
| "epoch": 0.7194058310479172, | |
| "grad_norm": 8.741981506347656, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1182, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 0.7203474617168281, | |
| "grad_norm": 6.671243667602539, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0422, | |
| "step": 3825 | |
| }, | |
| { | |
| "epoch": 0.721289092385739, | |
| "grad_norm": 7.100765228271484, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0205, | |
| "step": 3830 | |
| }, | |
| { | |
| "epoch": 0.7222307230546499, | |
| "grad_norm": 7.116576671600342, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0333, | |
| "step": 3835 | |
| }, | |
| { | |
| "epoch": 0.7231723537235608, | |
| "grad_norm": 6.72273588180542, | |
| "learning_rate": 1e-05, | |
| "loss": 3.081, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 0.7241139843924717, | |
| "grad_norm": 6.83878231048584, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9635, | |
| "step": 3845 | |
| }, | |
| { | |
| "epoch": 0.7250556150613825, | |
| "grad_norm": 7.474968433380127, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0792, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 0.7259972457302935, | |
| "grad_norm": 7.929852485656738, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9793, | |
| "step": 3855 | |
| }, | |
| { | |
| "epoch": 0.7269388763992043, | |
| "grad_norm": 6.638747215270996, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0311, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 0.7278805070681152, | |
| "grad_norm": 6.270215034484863, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9846, | |
| "step": 3865 | |
| }, | |
| { | |
| "epoch": 0.7288221377370261, | |
| "grad_norm": 7.199500560760498, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0651, | |
| "step": 3870 | |
| }, | |
| { | |
| "epoch": 0.7297637684059369, | |
| "grad_norm": 6.566064834594727, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9806, | |
| "step": 3875 | |
| }, | |
| { | |
| "epoch": 0.7307053990748479, | |
| "grad_norm": 7.1128830909729, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0375, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 0.7316470297437587, | |
| "grad_norm": 6.685820579528809, | |
| "learning_rate": 1e-05, | |
| "loss": 3.058, | |
| "step": 3885 | |
| }, | |
| { | |
| "epoch": 0.7325886604126697, | |
| "grad_norm": 6.155799388885498, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0514, | |
| "step": 3890 | |
| }, | |
| { | |
| "epoch": 0.7335302910815805, | |
| "grad_norm": 6.671477794647217, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9818, | |
| "step": 3895 | |
| }, | |
| { | |
| "epoch": 0.7344719217504914, | |
| "grad_norm": 7.42489767074585, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1053, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.7354135524194023, | |
| "grad_norm": 6.796725749969482, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9821, | |
| "step": 3905 | |
| }, | |
| { | |
| "epoch": 0.7363551830883132, | |
| "grad_norm": 7.848191261291504, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9765, | |
| "step": 3910 | |
| }, | |
| { | |
| "epoch": 0.7372968137572241, | |
| "grad_norm": 6.922936916351318, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1376, | |
| "step": 3915 | |
| }, | |
| { | |
| "epoch": 0.738238444426135, | |
| "grad_norm": 6.367747783660889, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1063, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 0.7391800750950458, | |
| "grad_norm": 6.697858810424805, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9815, | |
| "step": 3925 | |
| }, | |
| { | |
| "epoch": 0.7401217057639567, | |
| "grad_norm": 6.511905193328857, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9989, | |
| "step": 3930 | |
| }, | |
| { | |
| "epoch": 0.7410633364328676, | |
| "grad_norm": 7.045902729034424, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0436, | |
| "step": 3935 | |
| }, | |
| { | |
| "epoch": 0.7420049671017785, | |
| "grad_norm": 6.697847843170166, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9933, | |
| "step": 3940 | |
| }, | |
| { | |
| "epoch": 0.7429465977706894, | |
| "grad_norm": 8.465171813964844, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1475, | |
| "step": 3945 | |
| }, | |
| { | |
| "epoch": 0.7438882284396002, | |
| "grad_norm": 7.368391990661621, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9639, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 0.7448298591085112, | |
| "grad_norm": 9.309432983398438, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0413, | |
| "step": 3955 | |
| }, | |
| { | |
| "epoch": 0.745771489777422, | |
| "grad_norm": 7.960738658905029, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9888, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 0.746713120446333, | |
| "grad_norm": 7.475559711456299, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0662, | |
| "step": 3965 | |
| }, | |
| { | |
| "epoch": 0.7476547511152438, | |
| "grad_norm": 7.966568946838379, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0452, | |
| "step": 3970 | |
| }, | |
| { | |
| "epoch": 0.7485963817841547, | |
| "grad_norm": 7.25874137878418, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9464, | |
| "step": 3975 | |
| }, | |
| { | |
| "epoch": 0.7495380124530656, | |
| "grad_norm": 7.146042346954346, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0599, | |
| "step": 3980 | |
| }, | |
| { | |
| "epoch": 0.7504796431219765, | |
| "grad_norm": 7.571315765380859, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0561, | |
| "step": 3985 | |
| }, | |
| { | |
| "epoch": 0.7514212737908874, | |
| "grad_norm": 7.106105327606201, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0022, | |
| "step": 3990 | |
| }, | |
| { | |
| "epoch": 0.7523629044597983, | |
| "grad_norm": 7.547785758972168, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1197, | |
| "step": 3995 | |
| }, | |
| { | |
| "epoch": 0.7533045351287091, | |
| "grad_norm": 6.921874046325684, | |
| "learning_rate": 1e-05, | |
| "loss": 2.946, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.7533045351287091, | |
| "eval_accuracy": 0.4179943871149811, | |
| "eval_loss": 3.0350847244262695, | |
| "eval_runtime": 1060.3761, | |
| "eval_samples_per_second": 35.61, | |
| "eval_steps_per_second": 8.903, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.75424616579762, | |
| "grad_norm": 7.115399360656738, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9728, | |
| "step": 4005 | |
| }, | |
| { | |
| "epoch": 0.7551877964665309, | |
| "grad_norm": 6.1880269050598145, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0188, | |
| "step": 4010 | |
| }, | |
| { | |
| "epoch": 0.7561294271354418, | |
| "grad_norm": 6.350123405456543, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9826, | |
| "step": 4015 | |
| }, | |
| { | |
| "epoch": 0.7570710578043527, | |
| "grad_norm": 6.536036968231201, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0136, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 0.7580126884732635, | |
| "grad_norm": 6.593567848205566, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9847, | |
| "step": 4025 | |
| }, | |
| { | |
| "epoch": 0.7589543191421745, | |
| "grad_norm": 6.684474468231201, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1012, | |
| "step": 4030 | |
| }, | |
| { | |
| "epoch": 0.7598959498110853, | |
| "grad_norm": 6.77077579498291, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0014, | |
| "step": 4035 | |
| }, | |
| { | |
| "epoch": 0.7608375804799963, | |
| "grad_norm": 7.389578819274902, | |
| "learning_rate": 1e-05, | |
| "loss": 3.014, | |
| "step": 4040 | |
| }, | |
| { | |
| "epoch": 0.7617792111489071, | |
| "grad_norm": 6.702242374420166, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9964, | |
| "step": 4045 | |
| }, | |
| { | |
| "epoch": 0.762720841817818, | |
| "grad_norm": 7.013911247253418, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0237, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 0.7636624724867289, | |
| "grad_norm": 6.621683597564697, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9814, | |
| "step": 4055 | |
| }, | |
| { | |
| "epoch": 0.7646041031556398, | |
| "grad_norm": 6.537116050720215, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0512, | |
| "step": 4060 | |
| }, | |
| { | |
| "epoch": 0.7655457338245507, | |
| "grad_norm": 7.781442165374756, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0855, | |
| "step": 4065 | |
| }, | |
| { | |
| "epoch": 0.7664873644934616, | |
| "grad_norm": 7.431787967681885, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9926, | |
| "step": 4070 | |
| }, | |
| { | |
| "epoch": 0.7674289951623724, | |
| "grad_norm": 7.550678730010986, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9967, | |
| "step": 4075 | |
| }, | |
| { | |
| "epoch": 0.7683706258312833, | |
| "grad_norm": 7.120694637298584, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0271, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 0.7693122565001942, | |
| "grad_norm": 6.2324748039245605, | |
| "learning_rate": 1e-05, | |
| "loss": 2.979, | |
| "step": 4085 | |
| }, | |
| { | |
| "epoch": 0.7702538871691051, | |
| "grad_norm": 7.045401573181152, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0638, | |
| "step": 4090 | |
| }, | |
| { | |
| "epoch": 0.771195517838016, | |
| "grad_norm": 7.148534774780273, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9982, | |
| "step": 4095 | |
| }, | |
| { | |
| "epoch": 0.7721371485069268, | |
| "grad_norm": 6.389626979827881, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9228, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 0.7730787791758378, | |
| "grad_norm": 7.732367992401123, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0264, | |
| "step": 4105 | |
| }, | |
| { | |
| "epoch": 0.7740204098447486, | |
| "grad_norm": 6.456219673156738, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1302, | |
| "step": 4110 | |
| }, | |
| { | |
| "epoch": 0.7749620405136596, | |
| "grad_norm": 6.384737014770508, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9592, | |
| "step": 4115 | |
| }, | |
| { | |
| "epoch": 0.7759036711825704, | |
| "grad_norm": 7.74468469619751, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0449, | |
| "step": 4120 | |
| }, | |
| { | |
| "epoch": 0.7768453018514813, | |
| "grad_norm": 6.250748634338379, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0414, | |
| "step": 4125 | |
| }, | |
| { | |
| "epoch": 0.7777869325203922, | |
| "grad_norm": 6.97851037979126, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0961, | |
| "step": 4130 | |
| }, | |
| { | |
| "epoch": 0.7787285631893031, | |
| "grad_norm": 7.248284816741943, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9958, | |
| "step": 4135 | |
| }, | |
| { | |
| "epoch": 0.779670193858214, | |
| "grad_norm": 6.732598304748535, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9861, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 0.7806118245271249, | |
| "grad_norm": 6.4773173332214355, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9421, | |
| "step": 4145 | |
| }, | |
| { | |
| "epoch": 0.7815534551960357, | |
| "grad_norm": 6.1240620613098145, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9885, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 0.7824950858649466, | |
| "grad_norm": 8.18076229095459, | |
| "learning_rate": 1e-05, | |
| "loss": 2.8914, | |
| "step": 4155 | |
| }, | |
| { | |
| "epoch": 0.7834367165338575, | |
| "grad_norm": 6.479029178619385, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9848, | |
| "step": 4160 | |
| }, | |
| { | |
| "epoch": 0.7843783472027684, | |
| "grad_norm": 7.001804828643799, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9447, | |
| "step": 4165 | |
| }, | |
| { | |
| "epoch": 0.7853199778716793, | |
| "grad_norm": 7.894457817077637, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9969, | |
| "step": 4170 | |
| }, | |
| { | |
| "epoch": 0.7862616085405901, | |
| "grad_norm": 6.224025726318359, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1038, | |
| "step": 4175 | |
| }, | |
| { | |
| "epoch": 0.7872032392095011, | |
| "grad_norm": 7.672967910766602, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1315, | |
| "step": 4180 | |
| }, | |
| { | |
| "epoch": 0.7881448698784119, | |
| "grad_norm": 6.595861434936523, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0766, | |
| "step": 4185 | |
| }, | |
| { | |
| "epoch": 0.7890865005473229, | |
| "grad_norm": 6.691225051879883, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0528, | |
| "step": 4190 | |
| }, | |
| { | |
| "epoch": 0.7900281312162337, | |
| "grad_norm": 6.80654764175415, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0533, | |
| "step": 4195 | |
| }, | |
| { | |
| "epoch": 0.7909697618851446, | |
| "grad_norm": 7.244470119476318, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0414, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.7919113925540555, | |
| "grad_norm": 6.781643390655518, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0211, | |
| "step": 4205 | |
| }, | |
| { | |
| "epoch": 0.7928530232229664, | |
| "grad_norm": 6.594189167022705, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0595, | |
| "step": 4210 | |
| }, | |
| { | |
| "epoch": 0.7937946538918773, | |
| "grad_norm": 6.970313549041748, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0505, | |
| "step": 4215 | |
| }, | |
| { | |
| "epoch": 0.7947362845607882, | |
| "grad_norm": 7.164793491363525, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9477, | |
| "step": 4220 | |
| }, | |
| { | |
| "epoch": 0.795677915229699, | |
| "grad_norm": 6.909482955932617, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9963, | |
| "step": 4225 | |
| }, | |
| { | |
| "epoch": 0.79661954589861, | |
| "grad_norm": 6.645529747009277, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0773, | |
| "step": 4230 | |
| }, | |
| { | |
| "epoch": 0.7975611765675208, | |
| "grad_norm": 6.645266532897949, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0217, | |
| "step": 4235 | |
| }, | |
| { | |
| "epoch": 0.7985028072364317, | |
| "grad_norm": 6.193962574005127, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9698, | |
| "step": 4240 | |
| }, | |
| { | |
| "epoch": 0.7994444379053426, | |
| "grad_norm": 6.405771255493164, | |
| "learning_rate": 1e-05, | |
| "loss": 3.012, | |
| "step": 4245 | |
| }, | |
| { | |
| "epoch": 0.8003860685742534, | |
| "grad_norm": 6.351288795471191, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9623, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 0.8013276992431644, | |
| "grad_norm": 6.759338855743408, | |
| "learning_rate": 1e-05, | |
| "loss": 3.004, | |
| "step": 4255 | |
| }, | |
| { | |
| "epoch": 0.8022693299120752, | |
| "grad_norm": 6.525925636291504, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0707, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 0.8032109605809862, | |
| "grad_norm": 7.245208263397217, | |
| "learning_rate": 1e-05, | |
| "loss": 3.033, | |
| "step": 4265 | |
| }, | |
| { | |
| "epoch": 0.804152591249897, | |
| "grad_norm": 7.056334972381592, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9285, | |
| "step": 4270 | |
| }, | |
| { | |
| "epoch": 0.8050942219188079, | |
| "grad_norm": 7.097318172454834, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9806, | |
| "step": 4275 | |
| }, | |
| { | |
| "epoch": 0.8060358525877188, | |
| "grad_norm": 6.930456638336182, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0609, | |
| "step": 4280 | |
| }, | |
| { | |
| "epoch": 0.8069774832566297, | |
| "grad_norm": 6.811648368835449, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1415, | |
| "step": 4285 | |
| }, | |
| { | |
| "epoch": 0.8079191139255405, | |
| "grad_norm": 6.777252197265625, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0246, | |
| "step": 4290 | |
| }, | |
| { | |
| "epoch": 0.8088607445944515, | |
| "grad_norm": 6.139156818389893, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9366, | |
| "step": 4295 | |
| }, | |
| { | |
| "epoch": 0.8098023752633623, | |
| "grad_norm": 6.63328742980957, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0745, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 0.8107440059322732, | |
| "grad_norm": 7.7045183181762695, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0592, | |
| "step": 4305 | |
| }, | |
| { | |
| "epoch": 0.8116856366011841, | |
| "grad_norm": 6.513607978820801, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9483, | |
| "step": 4310 | |
| }, | |
| { | |
| "epoch": 0.8126272672700949, | |
| "grad_norm": 6.670443534851074, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0345, | |
| "step": 4315 | |
| }, | |
| { | |
| "epoch": 0.8135688979390059, | |
| "grad_norm": 6.660811901092529, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9852, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 0.8145105286079167, | |
| "grad_norm": 7.2338361740112305, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0945, | |
| "step": 4325 | |
| }, | |
| { | |
| "epoch": 0.8154521592768277, | |
| "grad_norm": 6.750393390655518, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9639, | |
| "step": 4330 | |
| }, | |
| { | |
| "epoch": 0.8163937899457385, | |
| "grad_norm": 7.605077743530273, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0707, | |
| "step": 4335 | |
| }, | |
| { | |
| "epoch": 0.8173354206146494, | |
| "grad_norm": 7.224457263946533, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0978, | |
| "step": 4340 | |
| }, | |
| { | |
| "epoch": 0.8182770512835603, | |
| "grad_norm": 6.255330562591553, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0075, | |
| "step": 4345 | |
| }, | |
| { | |
| "epoch": 0.8192186819524712, | |
| "grad_norm": 6.431050777435303, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1404, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 0.8201603126213821, | |
| "grad_norm": 7.618997097015381, | |
| "learning_rate": 1e-05, | |
| "loss": 2.8578, | |
| "step": 4355 | |
| }, | |
| { | |
| "epoch": 0.821101943290293, | |
| "grad_norm": 6.874431133270264, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9544, | |
| "step": 4360 | |
| }, | |
| { | |
| "epoch": 0.8220435739592038, | |
| "grad_norm": 6.51139497756958, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0135, | |
| "step": 4365 | |
| }, | |
| { | |
| "epoch": 0.8229852046281148, | |
| "grad_norm": 6.711349964141846, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0015, | |
| "step": 4370 | |
| }, | |
| { | |
| "epoch": 0.8239268352970256, | |
| "grad_norm": 7.047435283660889, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0321, | |
| "step": 4375 | |
| }, | |
| { | |
| "epoch": 0.8248684659659365, | |
| "grad_norm": 7.979480266571045, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1258, | |
| "step": 4380 | |
| }, | |
| { | |
| "epoch": 0.8258100966348474, | |
| "grad_norm": 6.74650239944458, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1225, | |
| "step": 4385 | |
| }, | |
| { | |
| "epoch": 0.8267517273037582, | |
| "grad_norm": 5.72318696975708, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9543, | |
| "step": 4390 | |
| }, | |
| { | |
| "epoch": 0.8276933579726692, | |
| "grad_norm": 6.581451416015625, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9904, | |
| "step": 4395 | |
| }, | |
| { | |
| "epoch": 0.82863498864158, | |
| "grad_norm": 6.8295769691467285, | |
| "learning_rate": 1e-05, | |
| "loss": 2.994, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 0.829576619310491, | |
| "grad_norm": 7.150579929351807, | |
| "learning_rate": 1e-05, | |
| "loss": 3.024, | |
| "step": 4405 | |
| }, | |
| { | |
| "epoch": 0.8305182499794018, | |
| "grad_norm": 6.949995994567871, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0905, | |
| "step": 4410 | |
| }, | |
| { | |
| "epoch": 0.8314598806483127, | |
| "grad_norm": 7.133937835693359, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0603, | |
| "step": 4415 | |
| }, | |
| { | |
| "epoch": 0.8324015113172236, | |
| "grad_norm": 6.495321750640869, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0027, | |
| "step": 4420 | |
| }, | |
| { | |
| "epoch": 0.8333431419861345, | |
| "grad_norm": 6.963568210601807, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9662, | |
| "step": 4425 | |
| }, | |
| { | |
| "epoch": 0.8342847726550454, | |
| "grad_norm": 7.362428188323975, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9869, | |
| "step": 4430 | |
| }, | |
| { | |
| "epoch": 0.8352264033239563, | |
| "grad_norm": 6.788746356964111, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0145, | |
| "step": 4435 | |
| }, | |
| { | |
| "epoch": 0.8361680339928671, | |
| "grad_norm": 7.730389595031738, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1651, | |
| "step": 4440 | |
| }, | |
| { | |
| "epoch": 0.837109664661778, | |
| "grad_norm": 7.5514116287231445, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0095, | |
| "step": 4445 | |
| }, | |
| { | |
| "epoch": 0.8380512953306889, | |
| "grad_norm": 6.9045257568359375, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0569, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 0.8389929259995998, | |
| "grad_norm": 7.573888778686523, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9869, | |
| "step": 4455 | |
| }, | |
| { | |
| "epoch": 0.8399345566685107, | |
| "grad_norm": 7.188941478729248, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0701, | |
| "step": 4460 | |
| }, | |
| { | |
| "epoch": 0.8408761873374215, | |
| "grad_norm": 6.4194817543029785, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1, | |
| "step": 4465 | |
| }, | |
| { | |
| "epoch": 0.8418178180063325, | |
| "grad_norm": 6.900866508483887, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9665, | |
| "step": 4470 | |
| }, | |
| { | |
| "epoch": 0.8427594486752433, | |
| "grad_norm": 7.416345596313477, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0122, | |
| "step": 4475 | |
| }, | |
| { | |
| "epoch": 0.8437010793441543, | |
| "grad_norm": 7.08629846572876, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1027, | |
| "step": 4480 | |
| }, | |
| { | |
| "epoch": 0.8446427100130651, | |
| "grad_norm": 6.6842803955078125, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9872, | |
| "step": 4485 | |
| }, | |
| { | |
| "epoch": 0.845584340681976, | |
| "grad_norm": 7.190927982330322, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9336, | |
| "step": 4490 | |
| }, | |
| { | |
| "epoch": 0.8465259713508869, | |
| "grad_norm": 7.118557929992676, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9181, | |
| "step": 4495 | |
| }, | |
| { | |
| "epoch": 0.8474676020197978, | |
| "grad_norm": 6.6827006340026855, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0798, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.8484092326887087, | |
| "grad_norm": 6.979405403137207, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9763, | |
| "step": 4505 | |
| }, | |
| { | |
| "epoch": 0.8493508633576196, | |
| "grad_norm": 6.2492780685424805, | |
| "learning_rate": 1e-05, | |
| "loss": 3.021, | |
| "step": 4510 | |
| }, | |
| { | |
| "epoch": 0.8502924940265304, | |
| "grad_norm": 7.063942909240723, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9505, | |
| "step": 4515 | |
| }, | |
| { | |
| "epoch": 0.8512341246954414, | |
| "grad_norm": 6.5366716384887695, | |
| "learning_rate": 1e-05, | |
| "loss": 2.8787, | |
| "step": 4520 | |
| }, | |
| { | |
| "epoch": 0.8521757553643522, | |
| "grad_norm": 6.5581769943237305, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1491, | |
| "step": 4525 | |
| }, | |
| { | |
| "epoch": 0.8531173860332631, | |
| "grad_norm": 6.441840171813965, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0369, | |
| "step": 4530 | |
| }, | |
| { | |
| "epoch": 0.854059016702174, | |
| "grad_norm": 7.320792198181152, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0956, | |
| "step": 4535 | |
| }, | |
| { | |
| "epoch": 0.8550006473710848, | |
| "grad_norm": 7.050429821014404, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0892, | |
| "step": 4540 | |
| }, | |
| { | |
| "epoch": 0.8559422780399958, | |
| "grad_norm": 6.61350679397583, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9785, | |
| "step": 4545 | |
| }, | |
| { | |
| "epoch": 0.8568839087089066, | |
| "grad_norm": 6.520270347595215, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0025, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 0.8578255393778176, | |
| "grad_norm": 6.5053791999816895, | |
| "learning_rate": 1e-05, | |
| "loss": 2.8551, | |
| "step": 4555 | |
| }, | |
| { | |
| "epoch": 0.8587671700467284, | |
| "grad_norm": 6.498419761657715, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9634, | |
| "step": 4560 | |
| }, | |
| { | |
| "epoch": 0.8597088007156393, | |
| "grad_norm": 7.01683235168457, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1556, | |
| "step": 4565 | |
| }, | |
| { | |
| "epoch": 0.8606504313845502, | |
| "grad_norm": 6.802071571350098, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0464, | |
| "step": 4570 | |
| }, | |
| { | |
| "epoch": 0.8615920620534611, | |
| "grad_norm": 6.664157390594482, | |
| "learning_rate": 1e-05, | |
| "loss": 3.012, | |
| "step": 4575 | |
| }, | |
| { | |
| "epoch": 0.862533692722372, | |
| "grad_norm": 6.3640456199646, | |
| "learning_rate": 1e-05, | |
| "loss": 2.934, | |
| "step": 4580 | |
| }, | |
| { | |
| "epoch": 0.8634753233912829, | |
| "grad_norm": 6.966064929962158, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9922, | |
| "step": 4585 | |
| }, | |
| { | |
| "epoch": 0.8644169540601937, | |
| "grad_norm": 6.394281387329102, | |
| "learning_rate": 1e-05, | |
| "loss": 2.904, | |
| "step": 4590 | |
| }, | |
| { | |
| "epoch": 0.8653585847291047, | |
| "grad_norm": 8.291240692138672, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0824, | |
| "step": 4595 | |
| }, | |
| { | |
| "epoch": 0.8663002153980155, | |
| "grad_norm": 6.770839214324951, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0608, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 0.8672418460669264, | |
| "grad_norm": 7.17354679107666, | |
| "learning_rate": 1e-05, | |
| "loss": 3.007, | |
| "step": 4605 | |
| }, | |
| { | |
| "epoch": 0.8681834767358373, | |
| "grad_norm": 6.194829940795898, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9562, | |
| "step": 4610 | |
| }, | |
| { | |
| "epoch": 0.8691251074047481, | |
| "grad_norm": 6.281735420227051, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9716, | |
| "step": 4615 | |
| }, | |
| { | |
| "epoch": 0.8700667380736591, | |
| "grad_norm": 7.197927951812744, | |
| "learning_rate": 1e-05, | |
| "loss": 3.008, | |
| "step": 4620 | |
| }, | |
| { | |
| "epoch": 0.8710083687425699, | |
| "grad_norm": 6.667356967926025, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1845, | |
| "step": 4625 | |
| }, | |
| { | |
| "epoch": 0.8719499994114809, | |
| "grad_norm": 7.937153339385986, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0967, | |
| "step": 4630 | |
| }, | |
| { | |
| "epoch": 0.8728916300803917, | |
| "grad_norm": 6.7486252784729, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0172, | |
| "step": 4635 | |
| }, | |
| { | |
| "epoch": 0.8738332607493026, | |
| "grad_norm": 7.023075103759766, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0151, | |
| "step": 4640 | |
| }, | |
| { | |
| "epoch": 0.8747748914182135, | |
| "grad_norm": 6.361885070800781, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0446, | |
| "step": 4645 | |
| }, | |
| { | |
| "epoch": 0.8757165220871244, | |
| "grad_norm": 6.498104572296143, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0111, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 0.8766581527560353, | |
| "grad_norm": 6.258391380310059, | |
| "learning_rate": 1e-05, | |
| "loss": 3.006, | |
| "step": 4655 | |
| }, | |
| { | |
| "epoch": 0.8775997834249462, | |
| "grad_norm": 7.305341720581055, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9864, | |
| "step": 4660 | |
| }, | |
| { | |
| "epoch": 0.878541414093857, | |
| "grad_norm": 6.587819576263428, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9275, | |
| "step": 4665 | |
| }, | |
| { | |
| "epoch": 0.879483044762768, | |
| "grad_norm": 6.967113018035889, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0478, | |
| "step": 4670 | |
| }, | |
| { | |
| "epoch": 0.8804246754316788, | |
| "grad_norm": 6.257725715637207, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9357, | |
| "step": 4675 | |
| }, | |
| { | |
| "epoch": 0.8813663061005897, | |
| "grad_norm": 7.08759069442749, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9292, | |
| "step": 4680 | |
| }, | |
| { | |
| "epoch": 0.8823079367695006, | |
| "grad_norm": 6.993216514587402, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9556, | |
| "step": 4685 | |
| }, | |
| { | |
| "epoch": 0.8832495674384114, | |
| "grad_norm": 6.2222771644592285, | |
| "learning_rate": 1e-05, | |
| "loss": 2.95, | |
| "step": 4690 | |
| }, | |
| { | |
| "epoch": 0.8841911981073224, | |
| "grad_norm": 5.418234825134277, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9276, | |
| "step": 4695 | |
| }, | |
| { | |
| "epoch": 0.8851328287762332, | |
| "grad_norm": 7.821156024932861, | |
| "learning_rate": 1e-05, | |
| "loss": 2.957, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 0.8860744594451442, | |
| "grad_norm": 7.50763463973999, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0653, | |
| "step": 4705 | |
| }, | |
| { | |
| "epoch": 0.887016090114055, | |
| "grad_norm": 5.917377471923828, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0135, | |
| "step": 4710 | |
| }, | |
| { | |
| "epoch": 0.8879577207829659, | |
| "grad_norm": 6.086848258972168, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9784, | |
| "step": 4715 | |
| }, | |
| { | |
| "epoch": 0.8888993514518768, | |
| "grad_norm": 7.318466663360596, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9473, | |
| "step": 4720 | |
| }, | |
| { | |
| "epoch": 0.8898409821207877, | |
| "grad_norm": 6.821482181549072, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0772, | |
| "step": 4725 | |
| }, | |
| { | |
| "epoch": 0.8907826127896986, | |
| "grad_norm": 6.979645252227783, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0302, | |
| "step": 4730 | |
| }, | |
| { | |
| "epoch": 0.8917242434586095, | |
| "grad_norm": 7.538966655731201, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0771, | |
| "step": 4735 | |
| }, | |
| { | |
| "epoch": 0.8926658741275203, | |
| "grad_norm": 6.447031497955322, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0053, | |
| "step": 4740 | |
| }, | |
| { | |
| "epoch": 0.8936075047964313, | |
| "grad_norm": 6.903289794921875, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0288, | |
| "step": 4745 | |
| }, | |
| { | |
| "epoch": 0.8945491354653421, | |
| "grad_norm": 6.24967098236084, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0103, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 0.895490766134253, | |
| "grad_norm": 6.205604553222656, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0107, | |
| "step": 4755 | |
| }, | |
| { | |
| "epoch": 0.8964323968031639, | |
| "grad_norm": 6.532459735870361, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0275, | |
| "step": 4760 | |
| }, | |
| { | |
| "epoch": 0.8973740274720747, | |
| "grad_norm": 6.63011360168457, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9595, | |
| "step": 4765 | |
| }, | |
| { | |
| "epoch": 0.8983156581409857, | |
| "grad_norm": 6.74818229675293, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0929, | |
| "step": 4770 | |
| }, | |
| { | |
| "epoch": 0.8992572888098965, | |
| "grad_norm": 6.410697937011719, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1027, | |
| "step": 4775 | |
| }, | |
| { | |
| "epoch": 0.9001989194788075, | |
| "grad_norm": 6.183323860168457, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9818, | |
| "step": 4780 | |
| }, | |
| { | |
| "epoch": 0.9011405501477183, | |
| "grad_norm": 6.527472496032715, | |
| "learning_rate": 1e-05, | |
| "loss": 2.996, | |
| "step": 4785 | |
| }, | |
| { | |
| "epoch": 0.9020821808166292, | |
| "grad_norm": 6.853177547454834, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0178, | |
| "step": 4790 | |
| }, | |
| { | |
| "epoch": 0.9030238114855401, | |
| "grad_norm": 7.206171035766602, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0277, | |
| "step": 4795 | |
| }, | |
| { | |
| "epoch": 0.903965442154451, | |
| "grad_norm": 7.369021892547607, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9376, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 0.9049070728233618, | |
| "grad_norm": 7.516411781311035, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0268, | |
| "step": 4805 | |
| }, | |
| { | |
| "epoch": 0.9058487034922728, | |
| "grad_norm": 6.495131015777588, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0405, | |
| "step": 4810 | |
| }, | |
| { | |
| "epoch": 0.9067903341611836, | |
| "grad_norm": 7.050095558166504, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1039, | |
| "step": 4815 | |
| }, | |
| { | |
| "epoch": 0.9077319648300946, | |
| "grad_norm": 6.463887691497803, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0336, | |
| "step": 4820 | |
| }, | |
| { | |
| "epoch": 0.9086735954990054, | |
| "grad_norm": 6.821866512298584, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1019, | |
| "step": 4825 | |
| }, | |
| { | |
| "epoch": 0.9096152261679162, | |
| "grad_norm": 9.358695983886719, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9516, | |
| "step": 4830 | |
| }, | |
| { | |
| "epoch": 0.9105568568368272, | |
| "grad_norm": 7.6679863929748535, | |
| "learning_rate": 1e-05, | |
| "loss": 2.979, | |
| "step": 4835 | |
| }, | |
| { | |
| "epoch": 0.911498487505738, | |
| "grad_norm": 6.348465919494629, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0011, | |
| "step": 4840 | |
| }, | |
| { | |
| "epoch": 0.912440118174649, | |
| "grad_norm": 6.158372402191162, | |
| "learning_rate": 1e-05, | |
| "loss": 2.8976, | |
| "step": 4845 | |
| }, | |
| { | |
| "epoch": 0.9133817488435598, | |
| "grad_norm": 6.827818870544434, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0961, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 0.9143233795124707, | |
| "grad_norm": 6.508996486663818, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9493, | |
| "step": 4855 | |
| }, | |
| { | |
| "epoch": 0.9152650101813816, | |
| "grad_norm": 6.057750701904297, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9642, | |
| "step": 4860 | |
| }, | |
| { | |
| "epoch": 0.9162066408502925, | |
| "grad_norm": 6.342018127441406, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9965, | |
| "step": 4865 | |
| }, | |
| { | |
| "epoch": 0.9171482715192034, | |
| "grad_norm": 7.247106552124023, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0096, | |
| "step": 4870 | |
| }, | |
| { | |
| "epoch": 0.9180899021881143, | |
| "grad_norm": 7.869892120361328, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0186, | |
| "step": 4875 | |
| }, | |
| { | |
| "epoch": 0.9190315328570251, | |
| "grad_norm": 6.655325412750244, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9532, | |
| "step": 4880 | |
| }, | |
| { | |
| "epoch": 0.919973163525936, | |
| "grad_norm": 7.238323211669922, | |
| "learning_rate": 1e-05, | |
| "loss": 2.964, | |
| "step": 4885 | |
| }, | |
| { | |
| "epoch": 0.9209147941948469, | |
| "grad_norm": 6.327097415924072, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9265, | |
| "step": 4890 | |
| }, | |
| { | |
| "epoch": 0.9218564248637579, | |
| "grad_norm": 6.87644624710083, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9748, | |
| "step": 4895 | |
| }, | |
| { | |
| "epoch": 0.9227980555326687, | |
| "grad_norm": 7.045034408569336, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9084, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 0.9237396862015795, | |
| "grad_norm": 7.708536624908447, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0383, | |
| "step": 4905 | |
| }, | |
| { | |
| "epoch": 0.9246813168704905, | |
| "grad_norm": 6.771285057067871, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9808, | |
| "step": 4910 | |
| }, | |
| { | |
| "epoch": 0.9256229475394013, | |
| "grad_norm": 6.201882362365723, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0511, | |
| "step": 4915 | |
| }, | |
| { | |
| "epoch": 0.9265645782083123, | |
| "grad_norm": 6.08769416809082, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0086, | |
| "step": 4920 | |
| }, | |
| { | |
| "epoch": 0.9275062088772231, | |
| "grad_norm": 6.0023393630981445, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0951, | |
| "step": 4925 | |
| }, | |
| { | |
| "epoch": 0.928447839546134, | |
| "grad_norm": 6.94964075088501, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0002, | |
| "step": 4930 | |
| }, | |
| { | |
| "epoch": 0.9293894702150449, | |
| "grad_norm": 6.350240707397461, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9763, | |
| "step": 4935 | |
| }, | |
| { | |
| "epoch": 0.9303311008839558, | |
| "grad_norm": 7.173480033874512, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9946, | |
| "step": 4940 | |
| }, | |
| { | |
| "epoch": 0.9312727315528667, | |
| "grad_norm": 6.645884990692139, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0732, | |
| "step": 4945 | |
| }, | |
| { | |
| "epoch": 0.9322143622217776, | |
| "grad_norm": 6.9946370124816895, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0506, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 0.9331559928906884, | |
| "grad_norm": 6.4223456382751465, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9862, | |
| "step": 4955 | |
| }, | |
| { | |
| "epoch": 0.9340976235595994, | |
| "grad_norm": 6.558778762817383, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9428, | |
| "step": 4960 | |
| }, | |
| { | |
| "epoch": 0.9350392542285102, | |
| "grad_norm": 7.190298080444336, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0667, | |
| "step": 4965 | |
| }, | |
| { | |
| "epoch": 0.9359808848974212, | |
| "grad_norm": 6.2120184898376465, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9411, | |
| "step": 4970 | |
| }, | |
| { | |
| "epoch": 0.936922515566332, | |
| "grad_norm": 7.766112804412842, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0971, | |
| "step": 4975 | |
| }, | |
| { | |
| "epoch": 0.9378641462352428, | |
| "grad_norm": 7.066897392272949, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1249, | |
| "step": 4980 | |
| }, | |
| { | |
| "epoch": 0.9388057769041538, | |
| "grad_norm": 6.099008560180664, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0034, | |
| "step": 4985 | |
| }, | |
| { | |
| "epoch": 0.9397474075730646, | |
| "grad_norm": 6.696260452270508, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0705, | |
| "step": 4990 | |
| }, | |
| { | |
| "epoch": 0.9406890382419756, | |
| "grad_norm": 6.140739440917969, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9683, | |
| "step": 4995 | |
| }, | |
| { | |
| "epoch": 0.9416306689108864, | |
| "grad_norm": 7.1208600997924805, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9827, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.9416306689108864, | |
| "eval_accuracy": 0.4198751663662041, | |
| "eval_loss": 3.008730888366699, | |
| "eval_runtime": 1038.4567, | |
| "eval_samples_per_second": 36.362, | |
| "eval_steps_per_second": 9.09, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.9425722995797973, | |
| "grad_norm": 6.645845890045166, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9817, | |
| "step": 5005 | |
| }, | |
| { | |
| "epoch": 0.9435139302487082, | |
| "grad_norm": 6.95515251159668, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0721, | |
| "step": 5010 | |
| }, | |
| { | |
| "epoch": 0.9444555609176191, | |
| "grad_norm": 6.197011947631836, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0179, | |
| "step": 5015 | |
| }, | |
| { | |
| "epoch": 0.94539719158653, | |
| "grad_norm": 6.877534866333008, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0148, | |
| "step": 5020 | |
| }, | |
| { | |
| "epoch": 0.9463388222554409, | |
| "grad_norm": 6.991530418395996, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0165, | |
| "step": 5025 | |
| }, | |
| { | |
| "epoch": 0.9472804529243517, | |
| "grad_norm": 7.398330211639404, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9934, | |
| "step": 5030 | |
| }, | |
| { | |
| "epoch": 0.9482220835932627, | |
| "grad_norm": 7.08344030380249, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0171, | |
| "step": 5035 | |
| }, | |
| { | |
| "epoch": 0.9491637142621735, | |
| "grad_norm": 6.782553195953369, | |
| "learning_rate": 1e-05, | |
| "loss": 2.932, | |
| "step": 5040 | |
| }, | |
| { | |
| "epoch": 0.9501053449310845, | |
| "grad_norm": 6.938860893249512, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9003, | |
| "step": 5045 | |
| }, | |
| { | |
| "epoch": 0.9510469755999953, | |
| "grad_norm": 6.443044185638428, | |
| "learning_rate": 1e-05, | |
| "loss": 3.066, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 0.9519886062689061, | |
| "grad_norm": 7.133784770965576, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0384, | |
| "step": 5055 | |
| }, | |
| { | |
| "epoch": 0.9529302369378171, | |
| "grad_norm": 7.460980415344238, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1658, | |
| "step": 5060 | |
| }, | |
| { | |
| "epoch": 0.9538718676067279, | |
| "grad_norm": 6.758791923522949, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0877, | |
| "step": 5065 | |
| }, | |
| { | |
| "epoch": 0.9548134982756389, | |
| "grad_norm": 6.621381759643555, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9102, | |
| "step": 5070 | |
| }, | |
| { | |
| "epoch": 0.9557551289445497, | |
| "grad_norm": 7.349249362945557, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9883, | |
| "step": 5075 | |
| }, | |
| { | |
| "epoch": 0.9566967596134606, | |
| "grad_norm": 7.027257919311523, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0714, | |
| "step": 5080 | |
| }, | |
| { | |
| "epoch": 0.9576383902823715, | |
| "grad_norm": 7.285141468048096, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9703, | |
| "step": 5085 | |
| }, | |
| { | |
| "epoch": 0.9585800209512824, | |
| "grad_norm": 6.913536548614502, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0111, | |
| "step": 5090 | |
| }, | |
| { | |
| "epoch": 0.9595216516201933, | |
| "grad_norm": 6.543398857116699, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0237, | |
| "step": 5095 | |
| }, | |
| { | |
| "epoch": 0.9604632822891042, | |
| "grad_norm": 6.569692611694336, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9803, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 0.961404912958015, | |
| "grad_norm": 6.915118217468262, | |
| "learning_rate": 1e-05, | |
| "loss": 2.941, | |
| "step": 5105 | |
| }, | |
| { | |
| "epoch": 0.962346543626926, | |
| "grad_norm": 7.568630695343018, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1096, | |
| "step": 5110 | |
| }, | |
| { | |
| "epoch": 0.9632881742958368, | |
| "grad_norm": 6.865966320037842, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0179, | |
| "step": 5115 | |
| }, | |
| { | |
| "epoch": 0.9642298049647477, | |
| "grad_norm": 6.584255695343018, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1118, | |
| "step": 5120 | |
| }, | |
| { | |
| "epoch": 0.9651714356336586, | |
| "grad_norm": 6.9417314529418945, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9616, | |
| "step": 5125 | |
| }, | |
| { | |
| "epoch": 0.9661130663025694, | |
| "grad_norm": 7.483455181121826, | |
| "learning_rate": 1e-05, | |
| "loss": 2.8741, | |
| "step": 5130 | |
| }, | |
| { | |
| "epoch": 0.9670546969714804, | |
| "grad_norm": 6.249847412109375, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0775, | |
| "step": 5135 | |
| }, | |
| { | |
| "epoch": 0.9679963276403912, | |
| "grad_norm": 7.1288018226623535, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0476, | |
| "step": 5140 | |
| }, | |
| { | |
| "epoch": 0.9689379583093022, | |
| "grad_norm": 6.932041168212891, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0245, | |
| "step": 5145 | |
| }, | |
| { | |
| "epoch": 0.969879588978213, | |
| "grad_norm": 7.062252044677734, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9943, | |
| "step": 5150 | |
| }, | |
| { | |
| "epoch": 0.9708212196471239, | |
| "grad_norm": 7.1132121086120605, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0385, | |
| "step": 5155 | |
| }, | |
| { | |
| "epoch": 0.9717628503160348, | |
| "grad_norm": 6.818399429321289, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0848, | |
| "step": 5160 | |
| }, | |
| { | |
| "epoch": 0.9727044809849457, | |
| "grad_norm": 6.7855610847473145, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0152, | |
| "step": 5165 | |
| }, | |
| { | |
| "epoch": 0.9736461116538566, | |
| "grad_norm": 6.324885368347168, | |
| "learning_rate": 1e-05, | |
| "loss": 2.91, | |
| "step": 5170 | |
| }, | |
| { | |
| "epoch": 0.9745877423227675, | |
| "grad_norm": 6.335012435913086, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0045, | |
| "step": 5175 | |
| }, | |
| { | |
| "epoch": 0.9755293729916783, | |
| "grad_norm": 6.50994873046875, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9847, | |
| "step": 5180 | |
| }, | |
| { | |
| "epoch": 0.9764710036605893, | |
| "grad_norm": 6.5178680419921875, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9757, | |
| "step": 5185 | |
| }, | |
| { | |
| "epoch": 0.9774126343295001, | |
| "grad_norm": 7.039647102355957, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9804, | |
| "step": 5190 | |
| }, | |
| { | |
| "epoch": 0.978354264998411, | |
| "grad_norm": 7.136976718902588, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9841, | |
| "step": 5195 | |
| }, | |
| { | |
| "epoch": 0.9792958956673219, | |
| "grad_norm": 6.711559772491455, | |
| "learning_rate": 1e-05, | |
| "loss": 3.1176, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 0.9802375263362327, | |
| "grad_norm": 6.82883882522583, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9664, | |
| "step": 5205 | |
| }, | |
| { | |
| "epoch": 0.9811791570051437, | |
| "grad_norm": 7.072768688201904, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9684, | |
| "step": 5210 | |
| }, | |
| { | |
| "epoch": 0.9821207876740545, | |
| "grad_norm": 6.364959716796875, | |
| "learning_rate": 1e-05, | |
| "loss": 2.8905, | |
| "step": 5215 | |
| }, | |
| { | |
| "epoch": 0.9830624183429655, | |
| "grad_norm": 6.766102313995361, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0164, | |
| "step": 5220 | |
| }, | |
| { | |
| "epoch": 0.9840040490118763, | |
| "grad_norm": 8.253533363342285, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9722, | |
| "step": 5225 | |
| }, | |
| { | |
| "epoch": 0.9849456796807872, | |
| "grad_norm": 7.092304229736328, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9922, | |
| "step": 5230 | |
| }, | |
| { | |
| "epoch": 0.9858873103496981, | |
| "grad_norm": 6.579424858093262, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9465, | |
| "step": 5235 | |
| }, | |
| { | |
| "epoch": 0.986828941018609, | |
| "grad_norm": 7.015688896179199, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9417, | |
| "step": 5240 | |
| }, | |
| { | |
| "epoch": 0.9877705716875199, | |
| "grad_norm": 7.938333511352539, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0678, | |
| "step": 5245 | |
| }, | |
| { | |
| "epoch": 0.9887122023564308, | |
| "grad_norm": 6.95517635345459, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9374, | |
| "step": 5250 | |
| }, | |
| { | |
| "epoch": 0.9896538330253416, | |
| "grad_norm": 6.90593957901001, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9633, | |
| "step": 5255 | |
| }, | |
| { | |
| "epoch": 0.9905954636942526, | |
| "grad_norm": 7.352856636047363, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0061, | |
| "step": 5260 | |
| }, | |
| { | |
| "epoch": 0.9915370943631634, | |
| "grad_norm": 7.105273723602295, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9133, | |
| "step": 5265 | |
| }, | |
| { | |
| "epoch": 0.9924787250320743, | |
| "grad_norm": 6.221724033355713, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9518, | |
| "step": 5270 | |
| }, | |
| { | |
| "epoch": 0.9934203557009852, | |
| "grad_norm": 6.9957380294799805, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0659, | |
| "step": 5275 | |
| }, | |
| { | |
| "epoch": 0.994361986369896, | |
| "grad_norm": 6.70432186126709, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0431, | |
| "step": 5280 | |
| }, | |
| { | |
| "epoch": 0.995303617038807, | |
| "grad_norm": 6.815949440002441, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9397, | |
| "step": 5285 | |
| }, | |
| { | |
| "epoch": 0.9962452477077178, | |
| "grad_norm": 6.253696918487549, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9773, | |
| "step": 5290 | |
| }, | |
| { | |
| "epoch": 0.9971868783766287, | |
| "grad_norm": 6.235177040100098, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9691, | |
| "step": 5295 | |
| }, | |
| { | |
| "epoch": 0.9981285090455396, | |
| "grad_norm": 8.551080703735352, | |
| "learning_rate": 1e-05, | |
| "loss": 3.0107, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 0.9990701397144505, | |
| "grad_norm": 6.636978626251221, | |
| "learning_rate": 1e-05, | |
| "loss": 2.9, | |
| "step": 5305 | |
| }, | |
| { | |
| "epoch": 0.9998234442495793, | |
| "step": 5309, | |
| "total_flos": 7.961680948018545e+18, | |
| "train_loss": 3.093570432952174, | |
| "train_runtime": 64923.0121, | |
| "train_samples_per_second": 5.234, | |
| "train_steps_per_second": 0.082 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 5309, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 200, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 7.961680948018545e+18, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |