| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.7558488302339532, | |
| "eval_steps": 500, | |
| "global_step": 252, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.002999400119976005, | |
| "grad_norm": 12.213529318427376, | |
| "learning_rate": 1.0000000000000002e-06, | |
| "loss": 1.6991, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.00599880023995201, | |
| "grad_norm": 11.149836011452198, | |
| "learning_rate": 2.0000000000000003e-06, | |
| "loss": 1.5997, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.008998200359928014, | |
| "grad_norm": 5.296165512046405, | |
| "learning_rate": 3e-06, | |
| "loss": 1.3037, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.01199760047990402, | |
| "grad_norm": 3.7277911972297235, | |
| "learning_rate": 4.000000000000001e-06, | |
| "loss": 1.3472, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.014997000599880024, | |
| "grad_norm": 6.970291148182751, | |
| "learning_rate": 5e-06, | |
| "loss": 1.5234, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.017996400719856028, | |
| "grad_norm": 5.825870871741097, | |
| "learning_rate": 6e-06, | |
| "loss": 1.2331, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.020995800839832032, | |
| "grad_norm": 4.789226733983147, | |
| "learning_rate": 7e-06, | |
| "loss": 1.2847, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.02399520095980804, | |
| "grad_norm": 5.735809923380043, | |
| "learning_rate": 8.000000000000001e-06, | |
| "loss": 1.3148, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.026994601079784044, | |
| "grad_norm": 4.146594143559111, | |
| "learning_rate": 9e-06, | |
| "loss": 1.3241, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.029994001199760048, | |
| "grad_norm": 4.3548572301609365, | |
| "learning_rate": 1e-05, | |
| "loss": 1.2048, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.032993401319736056, | |
| "grad_norm": 3.186860649490824, | |
| "learning_rate": 9.999942663491213e-06, | |
| "loss": 1.271, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.035992801439712056, | |
| "grad_norm": 5.683449148461711, | |
| "learning_rate": 9.999770655279843e-06, | |
| "loss": 1.4619, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.038992201559688064, | |
| "grad_norm": 2.918639590322084, | |
| "learning_rate": 9.99948397931083e-06, | |
| "loss": 1.2698, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.041991601679664065, | |
| "grad_norm": 3.1049642688975228, | |
| "learning_rate": 9.999082642158972e-06, | |
| "loss": 1.4018, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.04499100179964007, | |
| "grad_norm": 2.396816425806027, | |
| "learning_rate": 9.99856665302878e-06, | |
| "loss": 1.1897, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.04799040191961608, | |
| "grad_norm": 2.6334601975355283, | |
| "learning_rate": 9.997936023754258e-06, | |
| "loss": 1.3976, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.05098980203959208, | |
| "grad_norm": 2.619689679258694, | |
| "learning_rate": 9.997190768798639e-06, | |
| "loss": 1.286, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.05398920215956809, | |
| "grad_norm": 2.7084119507983377, | |
| "learning_rate": 9.99633090525405e-06, | |
| "loss": 1.3328, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.05698860227954409, | |
| "grad_norm": 2.3481903809919404, | |
| "learning_rate": 9.995356452841122e-06, | |
| "loss": 1.1662, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.059988002399520096, | |
| "grad_norm": 4.651416427067284, | |
| "learning_rate": 9.994267433908533e-06, | |
| "loss": 1.5033, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.0629874025194961, | |
| "grad_norm": 2.625598226472764, | |
| "learning_rate": 9.9930638734325e-06, | |
| "loss": 1.2022, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.06598680263947211, | |
| "grad_norm": 2.4104347021265973, | |
| "learning_rate": 9.991745799016206e-06, | |
| "loss": 1.1979, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.06898620275944811, | |
| "grad_norm": 3.1305838351835478, | |
| "learning_rate": 9.990313240889167e-06, | |
| "loss": 1.3105, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.07198560287942411, | |
| "grad_norm": 3.18309366674778, | |
| "learning_rate": 9.988766231906532e-06, | |
| "loss": 1.3079, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.07498500299940011, | |
| "grad_norm": 3.5781057246328563, | |
| "learning_rate": 9.987104807548341e-06, | |
| "loss": 1.3255, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.07798440311937613, | |
| "grad_norm": 2.259604288692876, | |
| "learning_rate": 9.985329005918702e-06, | |
| "loss": 1.242, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.08098380323935213, | |
| "grad_norm": 5.16511455231151, | |
| "learning_rate": 9.983438867744923e-06, | |
| "loss": 1.5295, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.08398320335932813, | |
| "grad_norm": 2.5507458276752955, | |
| "learning_rate": 9.981434436376572e-06, | |
| "loss": 1.1835, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.08698260347930414, | |
| "grad_norm": 2.374301337319529, | |
| "learning_rate": 9.97931575778449e-06, | |
| "loss": 1.2602, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.08998200359928014, | |
| "grad_norm": 2.869182508074545, | |
| "learning_rate": 9.977082880559725e-06, | |
| "loss": 1.2714, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.09298140371925614, | |
| "grad_norm": 3.352349678924823, | |
| "learning_rate": 9.974735855912436e-06, | |
| "loss": 1.2895, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.09598080383923216, | |
| "grad_norm": 2.402498887438619, | |
| "learning_rate": 9.972274737670702e-06, | |
| "loss": 1.2169, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.09898020395920816, | |
| "grad_norm": 2.3026043458108356, | |
| "learning_rate": 9.969699582279294e-06, | |
| "loss": 1.2486, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.10197960407918416, | |
| "grad_norm": 2.7234482697971796, | |
| "learning_rate": 9.967010448798376e-06, | |
| "loss": 1.3611, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.10497900419916016, | |
| "grad_norm": 2.3328312481588096, | |
| "learning_rate": 9.964207398902163e-06, | |
| "loss": 1.2112, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.10797840431913618, | |
| "grad_norm": 2.3738667556158304, | |
| "learning_rate": 9.961290496877492e-06, | |
| "loss": 1.1489, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.11097780443911218, | |
| "grad_norm": 2.381064080119671, | |
| "learning_rate": 9.958259809622353e-06, | |
| "loss": 1.3426, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.11397720455908818, | |
| "grad_norm": 2.789399877084244, | |
| "learning_rate": 9.955115406644357e-06, | |
| "loss": 1.2355, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.11697660467906419, | |
| "grad_norm": 2.2360091447066885, | |
| "learning_rate": 9.951857360059141e-06, | |
| "loss": 1.298, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.11997600479904019, | |
| "grad_norm": 2.3420608415672644, | |
| "learning_rate": 9.948485744588709e-06, | |
| "loss": 1.2597, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.12297540491901619, | |
| "grad_norm": 2.0514216630297786, | |
| "learning_rate": 9.945000637559728e-06, | |
| "loss": 1.2464, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.1259748050389922, | |
| "grad_norm": 2.1083906964286765, | |
| "learning_rate": 9.941402118901743e-06, | |
| "loss": 1.2223, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.1289742051589682, | |
| "grad_norm": 2.584585224378793, | |
| "learning_rate": 9.937690271145356e-06, | |
| "loss": 1.2575, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.13197360527894422, | |
| "grad_norm": 2.3779090684986652, | |
| "learning_rate": 9.93386517942032e-06, | |
| "loss": 1.1451, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.1349730053989202, | |
| "grad_norm": 2.108339055225494, | |
| "learning_rate": 9.9299269314536e-06, | |
| "loss": 1.2248, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.13797240551889622, | |
| "grad_norm": 2.121719885374642, | |
| "learning_rate": 9.92587561756735e-06, | |
| "loss": 1.207, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.14097180563887224, | |
| "grad_norm": 2.205771486341789, | |
| "learning_rate": 9.921711330676848e-06, | |
| "loss": 1.2321, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.14397120575884823, | |
| "grad_norm": 2.288650968193418, | |
| "learning_rate": 9.917434166288364e-06, | |
| "loss": 1.244, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.14697060587882424, | |
| "grad_norm": 2.3703349873199917, | |
| "learning_rate": 9.913044222496966e-06, | |
| "loss": 1.2758, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.14997000599880023, | |
| "grad_norm": 2.342816783402187, | |
| "learning_rate": 9.908541599984276e-06, | |
| "loss": 1.2799, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.15296940611877624, | |
| "grad_norm": 2.290011867632185, | |
| "learning_rate": 9.903926402016153e-06, | |
| "loss": 1.137, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.15596880623875226, | |
| "grad_norm": 2.2455094517388825, | |
| "learning_rate": 9.899198734440335e-06, | |
| "loss": 1.2133, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.15896820635872824, | |
| "grad_norm": 2.263487895355519, | |
| "learning_rate": 9.894358705684002e-06, | |
| "loss": 1.1898, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.16196760647870426, | |
| "grad_norm": 2.094119128560065, | |
| "learning_rate": 9.889406426751296e-06, | |
| "loss": 1.1847, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.16496700659868027, | |
| "grad_norm": 2.45640150031141, | |
| "learning_rate": 9.88434201122077e-06, | |
| "loss": 1.2662, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.16796640671865626, | |
| "grad_norm": 3.7462847341409606, | |
| "learning_rate": 9.879165575242788e-06, | |
| "loss": 1.1912, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.17096580683863227, | |
| "grad_norm": 2.382097038596032, | |
| "learning_rate": 9.873877237536854e-06, | |
| "loss": 1.2652, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.1739652069586083, | |
| "grad_norm": 2.320080713494513, | |
| "learning_rate": 9.868477119388897e-06, | |
| "loss": 1.2682, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.17696460707858427, | |
| "grad_norm": 2.1974546622057938, | |
| "learning_rate": 9.862965344648485e-06, | |
| "loss": 1.1956, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.1799640071985603, | |
| "grad_norm": 2.40241845961337, | |
| "learning_rate": 9.85734203972599e-06, | |
| "loss": 1.2948, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.1829634073185363, | |
| "grad_norm": 2.1855343450625053, | |
| "learning_rate": 9.851607333589677e-06, | |
| "loss": 1.1967, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.1859628074385123, | |
| "grad_norm": 2.772912562685191, | |
| "learning_rate": 9.84576135776276e-06, | |
| "loss": 1.1864, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.1889622075584883, | |
| "grad_norm": 2.420057020746961, | |
| "learning_rate": 9.839804246320374e-06, | |
| "loss": 1.2534, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.19196160767846432, | |
| "grad_norm": 2.265555099135192, | |
| "learning_rate": 9.833736135886513e-06, | |
| "loss": 1.2563, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.1949610077984403, | |
| "grad_norm": 2.292076182842636, | |
| "learning_rate": 9.827557165630879e-06, | |
| "loss": 1.193, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.19796040791841632, | |
| "grad_norm": 2.032874490713336, | |
| "learning_rate": 9.821267477265705e-06, | |
| "loss": 1.231, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.20095980803839233, | |
| "grad_norm": 2.1341683347971543, | |
| "learning_rate": 9.814867215042503e-06, | |
| "loss": 1.2317, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.20395920815836832, | |
| "grad_norm": 2.2463443930345077, | |
| "learning_rate": 9.808356525748748e-06, | |
| "loss": 1.2911, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.20695860827834434, | |
| "grad_norm": 2.0325804061530848, | |
| "learning_rate": 9.801735558704516e-06, | |
| "loss": 1.254, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.20995800839832032, | |
| "grad_norm": 2.0568500164008086, | |
| "learning_rate": 9.795004465759067e-06, | |
| "loss": 1.206, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.21295740851829634, | |
| "grad_norm": 2.082577302356669, | |
| "learning_rate": 9.78816340128734e-06, | |
| "loss": 1.2027, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.21595680863827235, | |
| "grad_norm": 2.048998263808703, | |
| "learning_rate": 9.781212522186442e-06, | |
| "loss": 1.2288, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.21895620875824834, | |
| "grad_norm": 2.49581568276726, | |
| "learning_rate": 9.774151987872029e-06, | |
| "loss": 1.2493, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.22195560887822435, | |
| "grad_norm": 1.8323511888260988, | |
| "learning_rate": 9.766981960274653e-06, | |
| "loss": 1.2133, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.22495500899820037, | |
| "grad_norm": 1.9244897973012691, | |
| "learning_rate": 9.759702603836059e-06, | |
| "loss": 1.1556, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.22795440911817635, | |
| "grad_norm": 2.2391757657702476, | |
| "learning_rate": 9.752314085505396e-06, | |
| "loss": 1.139, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.23095380923815237, | |
| "grad_norm": 2.1062513435343955, | |
| "learning_rate": 9.744816574735405e-06, | |
| "loss": 1.1568, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.23395320935812838, | |
| "grad_norm": 2.244216493001031, | |
| "learning_rate": 9.737210243478522e-06, | |
| "loss": 1.1672, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.23695260947810437, | |
| "grad_norm": 2.2301876958310585, | |
| "learning_rate": 9.72949526618294e-06, | |
| "loss": 1.1805, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.23995200959808038, | |
| "grad_norm": 2.087760281351449, | |
| "learning_rate": 9.721671819788603e-06, | |
| "loss": 1.2595, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.2429514097180564, | |
| "grad_norm": 2.040923837775886, | |
| "learning_rate": 9.713740083723152e-06, | |
| "loss": 1.1853, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.24595080983803239, | |
| "grad_norm": 1.9835044903865013, | |
| "learning_rate": 9.705700239897809e-06, | |
| "loss": 1.1165, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.2489502099580084, | |
| "grad_norm": 2.4048211037733433, | |
| "learning_rate": 9.697552472703205e-06, | |
| "loss": 1.2442, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.2519496100779844, | |
| "grad_norm": 1.9274434259225983, | |
| "learning_rate": 9.689296969005151e-06, | |
| "loss": 1.2233, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.2549490101979604, | |
| "grad_norm": 2.0930221605237813, | |
| "learning_rate": 9.680933918140348e-06, | |
| "loss": 1.2042, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.2579484103179364, | |
| "grad_norm": 2.050947291358715, | |
| "learning_rate": 9.672463511912056e-06, | |
| "loss": 1.1526, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.26094781043791243, | |
| "grad_norm": 2.1905662044973218, | |
| "learning_rate": 9.66388594458568e-06, | |
| "loss": 1.2428, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.26394721055788845, | |
| "grad_norm": 2.194939852549527, | |
| "learning_rate": 9.655201412884328e-06, | |
| "loss": 1.1842, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.2669466106778644, | |
| "grad_norm": 1.9191574847419615, | |
| "learning_rate": 9.64641011598429e-06, | |
| "loss": 1.2005, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.2699460107978404, | |
| "grad_norm": 2.0406272846772597, | |
| "learning_rate": 9.637512255510475e-06, | |
| "loss": 1.1275, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.27294541091781643, | |
| "grad_norm": 1.9546291627174466, | |
| "learning_rate": 9.628508035531785e-06, | |
| "loss": 1.2274, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.27594481103779245, | |
| "grad_norm": 1.9691367121837948, | |
| "learning_rate": 9.619397662556434e-06, | |
| "loss": 1.207, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.27894421115776846, | |
| "grad_norm": 2.3613406410056323, | |
| "learning_rate": 9.610181345527217e-06, | |
| "loss": 1.2273, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.2819436112777445, | |
| "grad_norm": 1.869890124546451, | |
| "learning_rate": 9.600859295816708e-06, | |
| "loss": 1.2272, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.28494301139772044, | |
| "grad_norm": 2.1493924596390834, | |
| "learning_rate": 9.591431727222425e-06, | |
| "loss": 1.1787, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.28794241151769645, | |
| "grad_norm": 2.055643425325773, | |
| "learning_rate": 9.581898855961911e-06, | |
| "loss": 1.1993, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.29094181163767247, | |
| "grad_norm": 1.84642386683885, | |
| "learning_rate": 9.572260900667794e-06, | |
| "loss": 1.1839, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.2939412117576485, | |
| "grad_norm": 1.890951805519698, | |
| "learning_rate": 9.562518082382751e-06, | |
| "loss": 1.207, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.2969406118776245, | |
| "grad_norm": 2.37835708136971, | |
| "learning_rate": 9.55267062455446e-06, | |
| "loss": 1.1894, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.29994001199760045, | |
| "grad_norm": 1.9298763701114428, | |
| "learning_rate": 9.542718753030463e-06, | |
| "loss": 1.17, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.30293941211757647, | |
| "grad_norm": 1.9738559838328023, | |
| "learning_rate": 9.532662696052986e-06, | |
| "loss": 1.1911, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.3059388122375525, | |
| "grad_norm": 2.097481526316711, | |
| "learning_rate": 9.522502684253709e-06, | |
| "loss": 1.2047, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.3089382123575285, | |
| "grad_norm": 2.2773066391553405, | |
| "learning_rate": 9.512238950648474e-06, | |
| "loss": 1.1683, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.3119376124775045, | |
| "grad_norm": 2.0545409984803435, | |
| "learning_rate": 9.501871730631944e-06, | |
| "loss": 1.1777, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.3149370125974805, | |
| "grad_norm": 2.0765643204952435, | |
| "learning_rate": 9.491401261972194e-06, | |
| "loss": 1.2306, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.3179364127174565, | |
| "grad_norm": 1.9891174989072355, | |
| "learning_rate": 9.480827784805278e-06, | |
| "loss": 1.2351, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.3209358128374325, | |
| "grad_norm": 1.889574475483626, | |
| "learning_rate": 9.4701515416297e-06, | |
| "loss": 1.1742, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.3239352129574085, | |
| "grad_norm": 2.0131114525216955, | |
| "learning_rate": 9.459372777300863e-06, | |
| "loss": 1.127, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.32693461307738453, | |
| "grad_norm": 2.0094971545889506, | |
| "learning_rate": 9.448491739025455e-06, | |
| "loss": 1.1203, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.32993401319736054, | |
| "grad_norm": 2.007024349964355, | |
| "learning_rate": 9.437508676355774e-06, | |
| "loss": 1.2005, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.33293341331733656, | |
| "grad_norm": 1.834949376512209, | |
| "learning_rate": 9.426423841184007e-06, | |
| "loss": 1.1451, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.3359328134373125, | |
| "grad_norm": 2.0659739662077863, | |
| "learning_rate": 9.415237487736452e-06, | |
| "loss": 1.1978, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.33893221355728853, | |
| "grad_norm": 2.0021385422334714, | |
| "learning_rate": 9.403949872567696e-06, | |
| "loss": 1.1969, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.34193161367726455, | |
| "grad_norm": 2.131713368907865, | |
| "learning_rate": 9.392561254554712e-06, | |
| "loss": 1.0958, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.34493101379724056, | |
| "grad_norm": 2.009265355301516, | |
| "learning_rate": 9.381071894890942e-06, | |
| "loss": 1.2154, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.3479304139172166, | |
| "grad_norm": 1.9134985847126271, | |
| "learning_rate": 9.369482057080293e-06, | |
| "loss": 1.2242, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.3509298140371926, | |
| "grad_norm": 1.9567206999616176, | |
| "learning_rate": 9.3577920069311e-06, | |
| "loss": 1.172, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.35392921415716855, | |
| "grad_norm": 1.9433526958741427, | |
| "learning_rate": 9.346002012550027e-06, | |
| "loss": 1.1785, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.35692861427714456, | |
| "grad_norm": 3.7758127026886776, | |
| "learning_rate": 9.334112344335924e-06, | |
| "loss": 1.2509, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.3599280143971206, | |
| "grad_norm": 2.2288097190307754, | |
| "learning_rate": 9.322123274973613e-06, | |
| "loss": 1.2034, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.3629274145170966, | |
| "grad_norm": 2.210018938006893, | |
| "learning_rate": 9.310035079427651e-06, | |
| "loss": 1.2089, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.3659268146370726, | |
| "grad_norm": 1.8381206653711355, | |
| "learning_rate": 9.297848034936007e-06, | |
| "loss": 1.186, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.36892621475704857, | |
| "grad_norm": 2.1078020542808913, | |
| "learning_rate": 9.285562421003716e-06, | |
| "loss": 1.2042, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.3719256148770246, | |
| "grad_norm": 2.2181514311801496, | |
| "learning_rate": 9.273178519396459e-06, | |
| "loss": 1.1561, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.3749250149970006, | |
| "grad_norm": 2.127427295141141, | |
| "learning_rate": 9.260696614134115e-06, | |
| "loss": 1.1935, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.3779244151169766, | |
| "grad_norm": 2.179542019285163, | |
| "learning_rate": 9.24811699148423e-06, | |
| "loss": 1.2001, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.3809238152369526, | |
| "grad_norm": 2.95737675688332, | |
| "learning_rate": 9.235439939955458e-06, | |
| "loss": 1.1333, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.38392321535692864, | |
| "grad_norm": 2.408346714267433, | |
| "learning_rate": 9.222665750290953e-06, | |
| "loss": 1.2035, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.3869226154769046, | |
| "grad_norm": 1.9803537052769442, | |
| "learning_rate": 9.209794715461691e-06, | |
| "loss": 1.2133, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.3899220155968806, | |
| "grad_norm": 2.0753391828539844, | |
| "learning_rate": 9.196827130659752e-06, | |
| "loss": 1.2389, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.3929214157168566, | |
| "grad_norm": 2.057624573882372, | |
| "learning_rate": 9.18376329329155e-06, | |
| "loss": 1.1612, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.39592081583683264, | |
| "grad_norm": 2.095421490379302, | |
| "learning_rate": 9.170603502971017e-06, | |
| "loss": 1.2319, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.39892021595680865, | |
| "grad_norm": 1.9629007266889844, | |
| "learning_rate": 9.157348061512728e-06, | |
| "loss": 1.2061, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.40191961607678467, | |
| "grad_norm": 1.9702164374790287, | |
| "learning_rate": 9.143997272924974e-06, | |
| "loss": 1.197, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.40491901619676063, | |
| "grad_norm": 2.094524351597277, | |
| "learning_rate": 9.1305514434028e-06, | |
| "loss": 1.1508, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.40791841631673664, | |
| "grad_norm": 2.078336809149187, | |
| "learning_rate": 9.117010881320973e-06, | |
| "loss": 1.2006, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.41091781643671266, | |
| "grad_norm": 2.1047922013296385, | |
| "learning_rate": 9.103375897226919e-06, | |
| "loss": 1.2225, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.41391721655668867, | |
| "grad_norm": 1.9800892010976672, | |
| "learning_rate": 9.089646803833589e-06, | |
| "loss": 1.2117, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.4169166166766647, | |
| "grad_norm": 1.948925596250933, | |
| "learning_rate": 9.075823916012298e-06, | |
| "loss": 1.1917, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.41991601679664065, | |
| "grad_norm": 1.8720535653109143, | |
| "learning_rate": 9.061907550785498e-06, | |
| "loss": 1.2048, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.42291541691661666, | |
| "grad_norm": 2.1914812005829316, | |
| "learning_rate": 9.047898027319508e-06, | |
| "loss": 1.2393, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.4259148170365927, | |
| "grad_norm": 3.541065061637063, | |
| "learning_rate": 9.033795666917191e-06, | |
| "loss": 1.1728, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.4289142171565687, | |
| "grad_norm": 1.958943800312053, | |
| "learning_rate": 9.019600793010596e-06, | |
| "loss": 1.1204, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.4319136172765447, | |
| "grad_norm": 1.8982949525316917, | |
| "learning_rate": 9.005313731153525e-06, | |
| "loss": 1.136, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.4349130173965207, | |
| "grad_norm": 2.1962092793560775, | |
| "learning_rate": 8.990934809014079e-06, | |
| "loss": 1.2025, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.4379124175164967, | |
| "grad_norm": 2.148888648427292, | |
| "learning_rate": 8.976464356367133e-06, | |
| "loss": 1.1971, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.4409118176364727, | |
| "grad_norm": 2.1656007482087243, | |
| "learning_rate": 8.961902705086785e-06, | |
| "loss": 1.1531, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.4439112177564487, | |
| "grad_norm": 2.245423572348511, | |
| "learning_rate": 8.947250189138732e-06, | |
| "loss": 1.205, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.4469106178764247, | |
| "grad_norm": 2.2015971794373037, | |
| "learning_rate": 8.932507144572616e-06, | |
| "loss": 1.1663, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.44991001799640074, | |
| "grad_norm": 2.040881991969036, | |
| "learning_rate": 8.917673909514321e-06, | |
| "loss": 1.2318, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.45290941811637675, | |
| "grad_norm": 2.0365534777961884, | |
| "learning_rate": 8.902750824158213e-06, | |
| "loss": 1.1634, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.4559088182363527, | |
| "grad_norm": 2.0931596144968903, | |
| "learning_rate": 8.887738230759334e-06, | |
| "loss": 1.1547, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.4589082183563287, | |
| "grad_norm": 2.040316251155918, | |
| "learning_rate": 8.872636473625564e-06, | |
| "loss": 1.1948, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.46190761847630474, | |
| "grad_norm": 2.0411032824079514, | |
| "learning_rate": 8.857445899109716e-06, | |
| "loss": 1.1292, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.46490701859628075, | |
| "grad_norm": 2.9762584994335923, | |
| "learning_rate": 8.84216685560159e-06, | |
| "loss": 1.1308, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.46790641871625677, | |
| "grad_norm": 2.097668580242586, | |
| "learning_rate": 8.826799693519996e-06, | |
| "loss": 1.1631, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.4709058188362327, | |
| "grad_norm": 2.0737253370280384, | |
| "learning_rate": 8.811344765304698e-06, | |
| "loss": 1.1484, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.47390521895620874, | |
| "grad_norm": 2.1733316818144033, | |
| "learning_rate": 8.795802425408352e-06, | |
| "loss": 1.1992, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.47690461907618475, | |
| "grad_norm": 3.0100700234706927, | |
| "learning_rate": 8.780173030288359e-06, | |
| "loss": 1.2115, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.47990401919616077, | |
| "grad_norm": 2.0471682767814374, | |
| "learning_rate": 8.7644569383987e-06, | |
| "loss": 1.2007, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.4829034193161368, | |
| "grad_norm": 2.069677080263759, | |
| "learning_rate": 8.748654510181709e-06, | |
| "loss": 1.1719, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.4859028194361128, | |
| "grad_norm": 2.136133148960194, | |
| "learning_rate": 8.732766108059814e-06, | |
| "loss": 1.1097, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.48890221955608876, | |
| "grad_norm": 2.2078781415683704, | |
| "learning_rate": 8.716792096427217e-06, | |
| "loss": 1.2822, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.49190161967606477, | |
| "grad_norm": 1.8108893344256605, | |
| "learning_rate": 8.700732841641542e-06, | |
| "loss": 1.1984, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.4949010197960408, | |
| "grad_norm": 1.9202600268802827, | |
| "learning_rate": 8.68458871201543e-06, | |
| "loss": 1.1822, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.4979004199160168, | |
| "grad_norm": 2.043987291062948, | |
| "learning_rate": 8.668360077808093e-06, | |
| "loss": 1.1588, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.5008998200359928, | |
| "grad_norm": 1.8608294617929213, | |
| "learning_rate": 8.652047311216823e-06, | |
| "loss": 1.1739, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.5038992201559688, | |
| "grad_norm": 2.0614264560161812, | |
| "learning_rate": 8.635650786368452e-06, | |
| "loss": 1.2053, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.5068986202759448, | |
| "grad_norm": 1.8415134777925706, | |
| "learning_rate": 8.61917087931078e-06, | |
| "loss": 1.1629, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.5098980203959208, | |
| "grad_norm": 2.067056790142203, | |
| "learning_rate": 8.602607968003935e-06, | |
| "loss": 1.1656, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.5128974205158968, | |
| "grad_norm": 1.9885496712381465, | |
| "learning_rate": 8.585962432311728e-06, | |
| "loss": 1.1061, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.5158968206358728, | |
| "grad_norm": 1.8542146492076237, | |
| "learning_rate": 8.569234653992916e-06, | |
| "loss": 1.2416, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.5188962207558488, | |
| "grad_norm": 2.2130247278463453, | |
| "learning_rate": 8.552425016692464e-06, | |
| "loss": 1.2651, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.5218956208758249, | |
| "grad_norm": 1.9214372548459986, | |
| "learning_rate": 8.535533905932739e-06, | |
| "loss": 1.2038, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.5248950209958009, | |
| "grad_norm": 1.8202998106553783, | |
| "learning_rate": 8.518561709104667e-06, | |
| "loss": 1.1806, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.5278944211157769, | |
| "grad_norm": 2.033480383542316, | |
| "learning_rate": 8.501508815458856e-06, | |
| "loss": 1.1138, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.5308938212357528, | |
| "grad_norm": 1.8173739641856819, | |
| "learning_rate": 8.484375616096658e-06, | |
| "loss": 1.0463, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.5338932213557288, | |
| "grad_norm": 1.9159018093250713, | |
| "learning_rate": 8.467162503961209e-06, | |
| "loss": 1.1438, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.5368926214757048, | |
| "grad_norm": 2.0146669190076416, | |
| "learning_rate": 8.449869873828411e-06, | |
| "loss": 1.2241, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.5398920215956808, | |
| "grad_norm": 2.07880204207024, | |
| "learning_rate": 8.432498122297879e-06, | |
| "loss": 1.2059, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.5428914217156569, | |
| "grad_norm": 2.1530579237214593, | |
| "learning_rate": 8.415047647783847e-06, | |
| "loss": 1.1945, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.5458908218356329, | |
| "grad_norm": 1.9176382821131852, | |
| "learning_rate": 8.39751885050603e-06, | |
| "loss": 1.2025, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.5488902219556089, | |
| "grad_norm": 2.032474526366641, | |
| "learning_rate": 8.379912132480441e-06, | |
| "loss": 1.1962, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.5518896220755849, | |
| "grad_norm": 2.454514968342902, | |
| "learning_rate": 8.36222789751018e-06, | |
| "loss": 1.2075, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.5548890221955609, | |
| "grad_norm": 2.033627242661126, | |
| "learning_rate": 8.344466551176163e-06, | |
| "loss": 1.2134, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.5578884223155369, | |
| "grad_norm": 2.0441091460883585, | |
| "learning_rate": 8.326628500827826e-06, | |
| "loss": 1.2038, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.5608878224355129, | |
| "grad_norm": 1.9807070431779146, | |
| "learning_rate": 8.308714155573785e-06, | |
| "loss": 1.12, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.563887222555489, | |
| "grad_norm": 1.9349184602121337, | |
| "learning_rate": 8.290723926272439e-06, | |
| "loss": 1.1802, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.5668866226754649, | |
| "grad_norm": 1.8384106015909107, | |
| "learning_rate": 8.27265822552257e-06, | |
| "loss": 1.1826, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.5698860227954409, | |
| "grad_norm": 1.9156655051619718, | |
| "learning_rate": 8.254517467653858e-06, | |
| "loss": 1.1257, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.5728854229154169, | |
| "grad_norm": 2.1724786856863973, | |
| "learning_rate": 8.236302068717393e-06, | |
| "loss": 1.1839, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.5758848230353929, | |
| "grad_norm": 1.8396024645135292, | |
| "learning_rate": 8.218012446476128e-06, | |
| "loss": 1.16, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.5788842231553689, | |
| "grad_norm": 1.9408487099445306, | |
| "learning_rate": 8.199649020395299e-06, | |
| "loss": 1.2241, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.5818836232753449, | |
| "grad_norm": 1.9108055995475, | |
| "learning_rate": 8.1812122116328e-06, | |
| "loss": 1.164, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.584883023395321, | |
| "grad_norm": 2.098400399811293, | |
| "learning_rate": 8.16270244302953e-06, | |
| "loss": 1.1549, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.587882423515297, | |
| "grad_norm": 2.1600698972560015, | |
| "learning_rate": 8.144120139099697e-06, | |
| "loss": 1.2257, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.590881823635273, | |
| "grad_norm": 2.4157151117755826, | |
| "learning_rate": 8.125465726021068e-06, | |
| "loss": 1.2377, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.593881223755249, | |
| "grad_norm": 2.0852805021246246, | |
| "learning_rate": 8.106739631625216e-06, | |
| "loss": 1.2469, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.596880623875225, | |
| "grad_norm": 1.9748280355919685, | |
| "learning_rate": 8.08794228538769e-06, | |
| "loss": 1.1286, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.5998800239952009, | |
| "grad_norm": 1.9281839006734618, | |
| "learning_rate": 8.06907411841817e-06, | |
| "loss": 1.1319, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.6028794241151769, | |
| "grad_norm": 1.8690571721258256, | |
| "learning_rate": 8.050135563450587e-06, | |
| "loss": 1.1856, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.6058788242351529, | |
| "grad_norm": 1.9748712035222622, | |
| "learning_rate": 8.031127054833192e-06, | |
| "loss": 1.1948, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.608878224355129, | |
| "grad_norm": 1.991151818988732, | |
| "learning_rate": 8.01204902851859e-06, | |
| "loss": 1.1659, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.611877624475105, | |
| "grad_norm": 2.958946009978542, | |
| "learning_rate": 7.992901922053751e-06, | |
| "loss": 1.1856, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.614877024595081, | |
| "grad_norm": 2.065918241043226, | |
| "learning_rate": 7.973686174569971e-06, | |
| "loss": 1.0692, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.617876424715057, | |
| "grad_norm": 2.123952469982477, | |
| "learning_rate": 7.954402226772804e-06, | |
| "loss": 1.1462, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.620875824835033, | |
| "grad_norm": 2.384554268129999, | |
| "learning_rate": 7.93505052093194e-06, | |
| "loss": 1.2331, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.623875224955009, | |
| "grad_norm": 2.213816272262928, | |
| "learning_rate": 7.915631500871084e-06, | |
| "loss": 1.2179, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.626874625074985, | |
| "grad_norm": 2.041537464539173, | |
| "learning_rate": 7.896145611957759e-06, | |
| "loss": 1.1869, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.629874025194961, | |
| "grad_norm": 1.9032099687526678, | |
| "learning_rate": 7.876593301093104e-06, | |
| "loss": 1.1041, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.6328734253149371, | |
| "grad_norm": 2.0485786710751452, | |
| "learning_rate": 7.856975016701616e-06, | |
| "loss": 1.2018, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.635872825434913, | |
| "grad_norm": 1.985579363766555, | |
| "learning_rate": 7.837291208720867e-06, | |
| "loss": 1.2284, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.638872225554889, | |
| "grad_norm": 1.8372693078207498, | |
| "learning_rate": 7.81754232859119e-06, | |
| "loss": 1.2194, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.641871625674865, | |
| "grad_norm": 2.034191799766837, | |
| "learning_rate": 7.797728829245321e-06, | |
| "loss": 1.2119, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.644871025794841, | |
| "grad_norm": 2.172768388345461, | |
| "learning_rate": 7.777851165098012e-06, | |
| "loss": 1.2264, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.647870425914817, | |
| "grad_norm": 2.036813209788241, | |
| "learning_rate": 7.757909792035608e-06, | |
| "loss": 1.2233, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.650869826034793, | |
| "grad_norm": 1.7687192297667345, | |
| "learning_rate": 7.737905167405596e-06, | |
| "loss": 1.2008, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.6538692261547691, | |
| "grad_norm": 1.9034513819160714, | |
| "learning_rate": 7.717837750006106e-06, | |
| "loss": 1.2116, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.6568686262747451, | |
| "grad_norm": 1.896055349909721, | |
| "learning_rate": 7.697708000075404e-06, | |
| "loss": 1.0943, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.6598680263947211, | |
| "grad_norm": 2.01228606334524, | |
| "learning_rate": 7.67751637928132e-06, | |
| "loss": 1.1236, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.6628674265146971, | |
| "grad_norm": 2.156992216970816, | |
| "learning_rate": 7.657263350710676e-06, | |
| "loss": 1.1291, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 0.6658668266346731, | |
| "grad_norm": 1.9699773673672667, | |
| "learning_rate": 7.636949378858647e-06, | |
| "loss": 1.1458, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.668866226754649, | |
| "grad_norm": 2.1989186993052754, | |
| "learning_rate": 7.616574929618126e-06, | |
| "loss": 1.1121, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 0.671865626874625, | |
| "grad_norm": 1.993731506914194, | |
| "learning_rate": 7.596140470269029e-06, | |
| "loss": 1.0735, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.674865026994601, | |
| "grad_norm": 1.8252968956264437, | |
| "learning_rate": 7.575646469467576e-06, | |
| "loss": 1.149, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.6778644271145771, | |
| "grad_norm": 2.055158428727496, | |
| "learning_rate": 7.555093397235553e-06, | |
| "loss": 1.154, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.6808638272345531, | |
| "grad_norm": 1.9915560854905294, | |
| "learning_rate": 7.5344817249495195e-06, | |
| "loss": 1.0942, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 0.6838632273545291, | |
| "grad_norm": 2.06764585971135, | |
| "learning_rate": 7.51381192533001e-06, | |
| "loss": 1.1884, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.6868626274745051, | |
| "grad_norm": 1.9341247237628625, | |
| "learning_rate": 7.493084472430683e-06, | |
| "loss": 1.1791, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 0.6898620275944811, | |
| "grad_norm": 1.9931838011874805, | |
| "learning_rate": 7.472299841627452e-06, | |
| "loss": 1.1299, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.6928614277144571, | |
| "grad_norm": 1.9296797551564173, | |
| "learning_rate": 7.451458509607583e-06, | |
| "loss": 1.1513, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.6958608278344331, | |
| "grad_norm": 2.0306426071035544, | |
| "learning_rate": 7.430560954358764e-06, | |
| "loss": 1.1837, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.6988602279544092, | |
| "grad_norm": 1.8564265209276956, | |
| "learning_rate": 7.409607655158139e-06, | |
| "loss": 1.2164, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 0.7018596280743852, | |
| "grad_norm": 2.2341388452646727, | |
| "learning_rate": 7.388599092561315e-06, | |
| "loss": 1.1252, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.7048590281943611, | |
| "grad_norm": 2.1161776966895496, | |
| "learning_rate": 7.367535748391349e-06, | |
| "loss": 1.213, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.7078584283143371, | |
| "grad_norm": 1.9243533852197758, | |
| "learning_rate": 7.3464181057276864e-06, | |
| "loss": 1.1046, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.7108578284343131, | |
| "grad_norm": 1.9237683937108236, | |
| "learning_rate": 7.325246648895089e-06, | |
| "loss": 1.1577, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.7138572285542891, | |
| "grad_norm": 1.7984916602185241, | |
| "learning_rate": 7.304021863452525e-06, | |
| "loss": 1.2052, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.7168566286742651, | |
| "grad_norm": 1.8721994512639362, | |
| "learning_rate": 7.282744236182033e-06, | |
| "loss": 1.1676, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 0.7198560287942412, | |
| "grad_norm": 1.8334072031704922, | |
| "learning_rate": 7.261414255077561e-06, | |
| "loss": 1.1118, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.7228554289142172, | |
| "grad_norm": 2.2489377670615474, | |
| "learning_rate": 7.240032409333765e-06, | |
| "loss": 1.1682, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 0.7258548290341932, | |
| "grad_norm": 1.8400363638367707, | |
| "learning_rate": 7.218599189334799e-06, | |
| "loss": 1.1638, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.7288542291541692, | |
| "grad_norm": 1.8851878635905157, | |
| "learning_rate": 7.197115086643069e-06, | |
| "loss": 1.1176, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.7318536292741452, | |
| "grad_norm": 1.8023550249649072, | |
| "learning_rate": 7.175580593987952e-06, | |
| "loss": 1.1164, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.7348530293941212, | |
| "grad_norm": 2.694191659057711, | |
| "learning_rate": 7.153996205254495e-06, | |
| "loss": 1.1573, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.7378524295140971, | |
| "grad_norm": 1.964255695607597, | |
| "learning_rate": 7.132362415472099e-06, | |
| "loss": 1.0549, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.7408518296340731, | |
| "grad_norm": 1.8017763616156734, | |
| "learning_rate": 7.1106797208031554e-06, | |
| "loss": 1.1446, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 0.7438512297540492, | |
| "grad_norm": 1.9428097126313124, | |
| "learning_rate": 7.088948618531668e-06, | |
| "loss": 1.1081, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.7468506298740252, | |
| "grad_norm": 2.002394147288212, | |
| "learning_rate": 7.067169607051851e-06, | |
| "loss": 1.1359, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.7498500299940012, | |
| "grad_norm": 1.8198961995762413, | |
| "learning_rate": 7.045343185856701e-06, | |
| "loss": 1.1201, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.7528494301139772, | |
| "grad_norm": 1.9151331196894974, | |
| "learning_rate": 7.02346985552653e-06, | |
| "loss": 1.0402, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 0.7558488302339532, | |
| "grad_norm": 1.9384378002461773, | |
| "learning_rate": 7.001550117717499e-06, | |
| "loss": 1.175, | |
| "step": 252 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 666, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 2, | |
| "save_steps": 84, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 139507934232576.0, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |