| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 5.988118811881188, | |
| "eval_steps": 500, | |
| "global_step": 1134, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.005280528052805281, | |
| "grad_norm": 7.677456378936768, | |
| "learning_rate": 4.385964912280702e-07, | |
| "loss": 0.9639, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.010561056105610561, | |
| "grad_norm": 6.208090305328369, | |
| "learning_rate": 8.771929824561404e-07, | |
| "loss": 0.8832, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.015841584158415842, | |
| "grad_norm": 7.966090202331543, | |
| "learning_rate": 1.3157894736842106e-06, | |
| "loss": 0.9749, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.021122112211221122, | |
| "grad_norm": 7.089982986450195, | |
| "learning_rate": 1.7543859649122807e-06, | |
| "loss": 0.9166, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.026402640264026403, | |
| "grad_norm": 9.171483039855957, | |
| "learning_rate": 2.1929824561403507e-06, | |
| "loss": 1.02, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.031683168316831684, | |
| "grad_norm": 7.198622703552246, | |
| "learning_rate": 2.631578947368421e-06, | |
| "loss": 0.9205, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.036963696369636964, | |
| "grad_norm": 6.673768997192383, | |
| "learning_rate": 3.070175438596491e-06, | |
| "loss": 0.9179, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.042244224422442245, | |
| "grad_norm": 5.332284450531006, | |
| "learning_rate": 3.5087719298245615e-06, | |
| "loss": 0.7845, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.047524752475247525, | |
| "grad_norm": 5.0864458084106445, | |
| "learning_rate": 3.9473684210526315e-06, | |
| "loss": 0.7676, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.052805280528052806, | |
| "grad_norm": 4.446084022521973, | |
| "learning_rate": 4.3859649122807014e-06, | |
| "loss": 0.6778, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.058085808580858087, | |
| "grad_norm": 4.680896759033203, | |
| "learning_rate": 4.824561403508772e-06, | |
| "loss": 0.6388, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.06336633663366337, | |
| "grad_norm": 3.745345115661621, | |
| "learning_rate": 5.263157894736842e-06, | |
| "loss": 0.7111, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.06864686468646865, | |
| "grad_norm": 3.1842403411865234, | |
| "learning_rate": 5.701754385964912e-06, | |
| "loss": 0.6327, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.07392739273927393, | |
| "grad_norm": 2.8893682956695557, | |
| "learning_rate": 6.140350877192982e-06, | |
| "loss": 0.6334, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.07920792079207921, | |
| "grad_norm": 2.5611019134521484, | |
| "learning_rate": 6.578947368421053e-06, | |
| "loss": 0.6284, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.08448844884488449, | |
| "grad_norm": 2.8038105964660645, | |
| "learning_rate": 7.017543859649123e-06, | |
| "loss": 0.6224, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.08976897689768977, | |
| "grad_norm": 2.51741099357605, | |
| "learning_rate": 7.456140350877193e-06, | |
| "loss": 0.6035, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.09504950495049505, | |
| "grad_norm": 2.309906482696533, | |
| "learning_rate": 7.894736842105263e-06, | |
| "loss": 0.4823, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.10033003300330033, | |
| "grad_norm": 2.719104528427124, | |
| "learning_rate": 8.333333333333334e-06, | |
| "loss": 0.5194, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.10561056105610561, | |
| "grad_norm": 2.3248131275177, | |
| "learning_rate": 8.771929824561403e-06, | |
| "loss": 0.5342, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.11089108910891089, | |
| "grad_norm": 2.2312052249908447, | |
| "learning_rate": 9.210526315789474e-06, | |
| "loss": 0.5275, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.11617161716171617, | |
| "grad_norm": 2.045426368713379, | |
| "learning_rate": 9.649122807017545e-06, | |
| "loss": 0.5499, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.12145214521452145, | |
| "grad_norm": 2.17621111869812, | |
| "learning_rate": 1.0087719298245614e-05, | |
| "loss": 0.5891, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.12673267326732673, | |
| "grad_norm": 2.363523244857788, | |
| "learning_rate": 1.0526315789473684e-05, | |
| "loss": 0.667, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.132013201320132, | |
| "grad_norm": 2.3387694358825684, | |
| "learning_rate": 1.0964912280701754e-05, | |
| "loss": 0.5842, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.1372937293729373, | |
| "grad_norm": 2.1026556491851807, | |
| "learning_rate": 1.1403508771929824e-05, | |
| "loss": 0.6199, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.14257425742574256, | |
| "grad_norm": 2.3579792976379395, | |
| "learning_rate": 1.1842105263157895e-05, | |
| "loss": 0.5985, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.14785478547854786, | |
| "grad_norm": 2.136988639831543, | |
| "learning_rate": 1.2280701754385964e-05, | |
| "loss": 0.5759, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.15313531353135312, | |
| "grad_norm": 1.9653126001358032, | |
| "learning_rate": 1.2719298245614037e-05, | |
| "loss": 0.5132, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.15841584158415842, | |
| "grad_norm": 2.0811052322387695, | |
| "learning_rate": 1.3157894736842106e-05, | |
| "loss": 0.514, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.16369636963696368, | |
| "grad_norm": 2.0110175609588623, | |
| "learning_rate": 1.3596491228070177e-05, | |
| "loss": 0.588, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.16897689768976898, | |
| "grad_norm": 2.020909070968628, | |
| "learning_rate": 1.4035087719298246e-05, | |
| "loss": 0.5896, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.17425742574257425, | |
| "grad_norm": 2.001084804534912, | |
| "learning_rate": 1.4473684210526317e-05, | |
| "loss": 0.5028, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.17953795379537954, | |
| "grad_norm": 1.9398471117019653, | |
| "learning_rate": 1.4912280701754386e-05, | |
| "loss": 0.5199, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.1848184818481848, | |
| "grad_norm": 2.3714287281036377, | |
| "learning_rate": 1.5350877192982457e-05, | |
| "loss": 0.6112, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.1900990099009901, | |
| "grad_norm": 2.054084062576294, | |
| "learning_rate": 1.5789473684210526e-05, | |
| "loss": 0.4697, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.19537953795379537, | |
| "grad_norm": 2.1482019424438477, | |
| "learning_rate": 1.62280701754386e-05, | |
| "loss": 0.442, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.20066006600660066, | |
| "grad_norm": 2.005889892578125, | |
| "learning_rate": 1.6666666666666667e-05, | |
| "loss": 0.4276, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.20594059405940593, | |
| "grad_norm": 2.1889655590057373, | |
| "learning_rate": 1.7105263157894737e-05, | |
| "loss": 0.5114, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.21122112211221122, | |
| "grad_norm": 1.9033912420272827, | |
| "learning_rate": 1.7543859649122806e-05, | |
| "loss": 0.5266, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.2165016501650165, | |
| "grad_norm": 2.01960825920105, | |
| "learning_rate": 1.7982456140350878e-05, | |
| "loss": 0.485, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.22178217821782178, | |
| "grad_norm": 2.0285496711730957, | |
| "learning_rate": 1.8421052631578947e-05, | |
| "loss": 0.4915, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.22706270627062705, | |
| "grad_norm": 1.9628126621246338, | |
| "learning_rate": 1.885964912280702e-05, | |
| "loss": 0.4059, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.23234323432343235, | |
| "grad_norm": 2.2826972007751465, | |
| "learning_rate": 1.929824561403509e-05, | |
| "loss": 0.5476, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.2376237623762376, | |
| "grad_norm": 2.3612606525421143, | |
| "learning_rate": 1.9736842105263158e-05, | |
| "loss": 0.5466, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.2429042904290429, | |
| "grad_norm": 1.8648459911346436, | |
| "learning_rate": 2.0175438596491227e-05, | |
| "loss": 0.505, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.24818481848184817, | |
| "grad_norm": 1.9400116205215454, | |
| "learning_rate": 2.06140350877193e-05, | |
| "loss": 0.4488, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.25346534653465347, | |
| "grad_norm": 1.8791626691818237, | |
| "learning_rate": 2.105263157894737e-05, | |
| "loss": 0.4311, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.25874587458745874, | |
| "grad_norm": 1.8579607009887695, | |
| "learning_rate": 2.149122807017544e-05, | |
| "loss": 0.4188, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.264026402640264, | |
| "grad_norm": 1.8994585275650024, | |
| "learning_rate": 2.1929824561403507e-05, | |
| "loss": 0.4455, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.2693069306930693, | |
| "grad_norm": 1.9908592700958252, | |
| "learning_rate": 2.236842105263158e-05, | |
| "loss": 0.4684, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.2745874587458746, | |
| "grad_norm": 2.282810926437378, | |
| "learning_rate": 2.280701754385965e-05, | |
| "loss": 0.6328, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.27986798679867986, | |
| "grad_norm": 2.017083168029785, | |
| "learning_rate": 2.324561403508772e-05, | |
| "loss": 0.4561, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.2851485148514851, | |
| "grad_norm": 2.0290071964263916, | |
| "learning_rate": 2.368421052631579e-05, | |
| "loss": 0.4749, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.29042904290429045, | |
| "grad_norm": 2.1724143028259277, | |
| "learning_rate": 2.412280701754386e-05, | |
| "loss": 0.6216, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.2957095709570957, | |
| "grad_norm": 1.8445512056350708, | |
| "learning_rate": 2.456140350877193e-05, | |
| "loss": 0.4473, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.300990099009901, | |
| "grad_norm": 1.9536579847335815, | |
| "learning_rate": 2.5e-05, | |
| "loss": 0.4746, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.30627062706270625, | |
| "grad_norm": 2.0544443130493164, | |
| "learning_rate": 2.5438596491228074e-05, | |
| "loss": 0.4069, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.31155115511551157, | |
| "grad_norm": 2.315953016281128, | |
| "learning_rate": 2.5877192982456143e-05, | |
| "loss": 0.4296, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.31683168316831684, | |
| "grad_norm": 2.234273672103882, | |
| "learning_rate": 2.6315789473684212e-05, | |
| "loss": 0.4941, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.3221122112211221, | |
| "grad_norm": 2.1946239471435547, | |
| "learning_rate": 2.675438596491228e-05, | |
| "loss": 0.4416, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.32739273927392737, | |
| "grad_norm": 2.0075912475585938, | |
| "learning_rate": 2.7192982456140354e-05, | |
| "loss": 0.4202, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.3326732673267327, | |
| "grad_norm": 4.797417640686035, | |
| "learning_rate": 2.7631578947368426e-05, | |
| "loss": 0.452, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.33795379537953796, | |
| "grad_norm": 3.3362960815429688, | |
| "learning_rate": 2.8070175438596492e-05, | |
| "loss": 0.5405, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.3432343234323432, | |
| "grad_norm": 1.9122159481048584, | |
| "learning_rate": 2.850877192982456e-05, | |
| "loss": 0.3242, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.3485148514851485, | |
| "grad_norm": 1.8941394090652466, | |
| "learning_rate": 2.8947368421052634e-05, | |
| "loss": 0.4896, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.3537953795379538, | |
| "grad_norm": 1.8021477460861206, | |
| "learning_rate": 2.9385964912280706e-05, | |
| "loss": 0.3874, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.3590759075907591, | |
| "grad_norm": 2.036555767059326, | |
| "learning_rate": 2.9824561403508772e-05, | |
| "loss": 0.4886, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.36435643564356435, | |
| "grad_norm": 2.1943323612213135, | |
| "learning_rate": 3.0263157894736844e-05, | |
| "loss": 0.421, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.3696369636963696, | |
| "grad_norm": 2.077173948287964, | |
| "learning_rate": 3.0701754385964913e-05, | |
| "loss": 0.5161, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.37491749174917494, | |
| "grad_norm": 2.0704095363616943, | |
| "learning_rate": 3.1140350877192986e-05, | |
| "loss": 0.4241, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.3801980198019802, | |
| "grad_norm": 2.082000970840454, | |
| "learning_rate": 3.157894736842105e-05, | |
| "loss": 0.4373, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.38547854785478547, | |
| "grad_norm": 1.8969218730926514, | |
| "learning_rate": 3.2017543859649124e-05, | |
| "loss": 0.4705, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.39075907590759074, | |
| "grad_norm": 2.206298351287842, | |
| "learning_rate": 3.24561403508772e-05, | |
| "loss": 0.4938, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.39603960396039606, | |
| "grad_norm": 2.0572750568389893, | |
| "learning_rate": 3.289473684210527e-05, | |
| "loss": 0.5007, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.4013201320132013, | |
| "grad_norm": 1.94302237033844, | |
| "learning_rate": 3.3333333333333335e-05, | |
| "loss": 0.3689, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.4066006600660066, | |
| "grad_norm": 2.0126149654388428, | |
| "learning_rate": 3.377192982456141e-05, | |
| "loss": 0.4881, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.41188118811881186, | |
| "grad_norm": 1.887984037399292, | |
| "learning_rate": 3.421052631578947e-05, | |
| "loss": 0.4761, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.4171617161716172, | |
| "grad_norm": 1.9020264148712158, | |
| "learning_rate": 3.4649122807017546e-05, | |
| "loss": 0.4343, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.42244224422442245, | |
| "grad_norm": 1.942435622215271, | |
| "learning_rate": 3.508771929824561e-05, | |
| "loss": 0.4563, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.4277227722772277, | |
| "grad_norm": 2.269737482070923, | |
| "learning_rate": 3.5526315789473684e-05, | |
| "loss": 0.4508, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.433003300330033, | |
| "grad_norm": 2.0216665267944336, | |
| "learning_rate": 3.5964912280701756e-05, | |
| "loss": 0.4971, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.4382838283828383, | |
| "grad_norm": 2.1765635013580322, | |
| "learning_rate": 3.640350877192983e-05, | |
| "loss": 0.452, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.44356435643564357, | |
| "grad_norm": 2.25856876373291, | |
| "learning_rate": 3.6842105263157895e-05, | |
| "loss": 0.4824, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.44884488448844884, | |
| "grad_norm": 2.2144601345062256, | |
| "learning_rate": 3.728070175438597e-05, | |
| "loss": 0.4479, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.4541254125412541, | |
| "grad_norm": 1.97480309009552, | |
| "learning_rate": 3.771929824561404e-05, | |
| "loss": 0.3071, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.4594059405940594, | |
| "grad_norm": 1.9352009296417236, | |
| "learning_rate": 3.815789473684211e-05, | |
| "loss": 0.4513, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.4646864686468647, | |
| "grad_norm": 2.055535316467285, | |
| "learning_rate": 3.859649122807018e-05, | |
| "loss": 0.4508, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.46996699669966996, | |
| "grad_norm": 1.93705415725708, | |
| "learning_rate": 3.9035087719298244e-05, | |
| "loss": 0.4083, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.4752475247524752, | |
| "grad_norm": 1.9412288665771484, | |
| "learning_rate": 3.9473684210526316e-05, | |
| "loss": 0.3715, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.48052805280528055, | |
| "grad_norm": 2.098421335220337, | |
| "learning_rate": 3.991228070175439e-05, | |
| "loss": 0.414, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.4858085808580858, | |
| "grad_norm": 2.2177186012268066, | |
| "learning_rate": 4.0350877192982455e-05, | |
| "loss": 0.4736, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.4910891089108911, | |
| "grad_norm": 2.056107759475708, | |
| "learning_rate": 4.078947368421053e-05, | |
| "loss": 0.4177, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.49636963696369635, | |
| "grad_norm": 1.9581352472305298, | |
| "learning_rate": 4.12280701754386e-05, | |
| "loss": 0.3688, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.5016501650165016, | |
| "grad_norm": 2.2061662673950195, | |
| "learning_rate": 4.166666666666667e-05, | |
| "loss": 0.4705, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.5069306930693069, | |
| "grad_norm": 1.7467211484909058, | |
| "learning_rate": 4.210526315789474e-05, | |
| "loss": 0.3303, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.5122112211221123, | |
| "grad_norm": 1.9702417850494385, | |
| "learning_rate": 4.254385964912281e-05, | |
| "loss": 0.5031, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.5174917491749175, | |
| "grad_norm": 2.079378604888916, | |
| "learning_rate": 4.298245614035088e-05, | |
| "loss": 0.456, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.5227722772277228, | |
| "grad_norm": 1.8181231021881104, | |
| "learning_rate": 4.342105263157895e-05, | |
| "loss": 0.4231, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.528052805280528, | |
| "grad_norm": 2.1575117111206055, | |
| "learning_rate": 4.3859649122807014e-05, | |
| "loss": 0.457, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.5333333333333333, | |
| "grad_norm": 2.3540713787078857, | |
| "learning_rate": 4.429824561403509e-05, | |
| "loss": 0.397, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.5386138613861386, | |
| "grad_norm": 2.3277106285095215, | |
| "learning_rate": 4.473684210526316e-05, | |
| "loss": 0.474, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.5438943894389439, | |
| "grad_norm": 2.0837771892547607, | |
| "learning_rate": 4.517543859649123e-05, | |
| "loss": 0.4911, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.5491749174917492, | |
| "grad_norm": 2.315387487411499, | |
| "learning_rate": 4.56140350877193e-05, | |
| "loss": 0.5138, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.5544554455445545, | |
| "grad_norm": 1.9372241497039795, | |
| "learning_rate": 4.605263157894737e-05, | |
| "loss": 0.402, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.5597359735973597, | |
| "grad_norm": 2.0722286701202393, | |
| "learning_rate": 4.649122807017544e-05, | |
| "loss": 0.3484, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.565016501650165, | |
| "grad_norm": 1.8825434446334839, | |
| "learning_rate": 4.6929824561403515e-05, | |
| "loss": 0.4057, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.5702970297029702, | |
| "grad_norm": 2.0918331146240234, | |
| "learning_rate": 4.736842105263158e-05, | |
| "loss": 0.4073, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.5755775577557756, | |
| "grad_norm": 2.246974468231201, | |
| "learning_rate": 4.780701754385965e-05, | |
| "loss": 0.4104, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.5808580858085809, | |
| "grad_norm": 1.8505111932754517, | |
| "learning_rate": 4.824561403508772e-05, | |
| "loss": 0.351, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.5861386138613861, | |
| "grad_norm": 2.2233192920684814, | |
| "learning_rate": 4.868421052631579e-05, | |
| "loss": 0.4916, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.5914191419141914, | |
| "grad_norm": 2.3233530521392822, | |
| "learning_rate": 4.912280701754386e-05, | |
| "loss": 0.4886, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.5966996699669967, | |
| "grad_norm": 2.298288345336914, | |
| "learning_rate": 4.956140350877193e-05, | |
| "loss": 0.497, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.601980198019802, | |
| "grad_norm": 1.9848483800888062, | |
| "learning_rate": 5e-05, | |
| "loss": 0.3982, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.6072607260726073, | |
| "grad_norm": 1.7758945226669312, | |
| "learning_rate": 4.995098039215686e-05, | |
| "loss": 0.4284, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.6125412541254125, | |
| "grad_norm": 2.1073226928710938, | |
| "learning_rate": 4.990196078431373e-05, | |
| "loss": 0.4741, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.6178217821782178, | |
| "grad_norm": 2.1385958194732666, | |
| "learning_rate": 4.985294117647059e-05, | |
| "loss": 0.3853, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.6231023102310231, | |
| "grad_norm": 2.053973436355591, | |
| "learning_rate": 4.980392156862745e-05, | |
| "loss": 0.4385, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.6283828382838283, | |
| "grad_norm": 1.7011091709136963, | |
| "learning_rate": 4.975490196078432e-05, | |
| "loss": 0.3604, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.6336633663366337, | |
| "grad_norm": 2.1312694549560547, | |
| "learning_rate": 4.970588235294118e-05, | |
| "loss": 0.4636, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.638943894389439, | |
| "grad_norm": 1.9020744562149048, | |
| "learning_rate": 4.9656862745098046e-05, | |
| "loss": 0.3795, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.6442244224422442, | |
| "grad_norm": 1.9798043966293335, | |
| "learning_rate": 4.960784313725491e-05, | |
| "loss": 0.382, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.6495049504950495, | |
| "grad_norm": 1.8981302976608276, | |
| "learning_rate": 4.955882352941177e-05, | |
| "loss": 0.4038, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.6547854785478547, | |
| "grad_norm": 1.9499566555023193, | |
| "learning_rate": 4.9509803921568634e-05, | |
| "loss": 0.434, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.6600660066006601, | |
| "grad_norm": 1.912457823753357, | |
| "learning_rate": 4.9460784313725495e-05, | |
| "loss": 0.3187, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.6653465346534654, | |
| "grad_norm": 2.0483133792877197, | |
| "learning_rate": 4.9411764705882355e-05, | |
| "loss": 0.3916, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.6706270627062706, | |
| "grad_norm": 1.924310564994812, | |
| "learning_rate": 4.936274509803922e-05, | |
| "loss": 0.378, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.6759075907590759, | |
| "grad_norm": 2.1889538764953613, | |
| "learning_rate": 4.931372549019608e-05, | |
| "loss": 0.4093, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.6811881188118812, | |
| "grad_norm": 1.8973898887634277, | |
| "learning_rate": 4.9264705882352944e-05, | |
| "loss": 0.3571, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.6864686468646864, | |
| "grad_norm": 2.0125250816345215, | |
| "learning_rate": 4.9215686274509804e-05, | |
| "loss": 0.359, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.6917491749174918, | |
| "grad_norm": 1.9622538089752197, | |
| "learning_rate": 4.9166666666666665e-05, | |
| "loss": 0.3238, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.697029702970297, | |
| "grad_norm": 1.7294894456863403, | |
| "learning_rate": 4.911764705882353e-05, | |
| "loss": 0.3555, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.7023102310231023, | |
| "grad_norm": 2.0299930572509766, | |
| "learning_rate": 4.906862745098039e-05, | |
| "loss": 0.4667, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.7075907590759076, | |
| "grad_norm": 1.813370704650879, | |
| "learning_rate": 4.901960784313725e-05, | |
| "loss": 0.3004, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.7128712871287128, | |
| "grad_norm": 2.090129852294922, | |
| "learning_rate": 4.897058823529412e-05, | |
| "loss": 0.3845, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.7181518151815182, | |
| "grad_norm": 2.3778114318847656, | |
| "learning_rate": 4.892156862745098e-05, | |
| "loss": 0.4731, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.7234323432343235, | |
| "grad_norm": 2.1827681064605713, | |
| "learning_rate": 4.887254901960784e-05, | |
| "loss": 0.4078, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.7287128712871287, | |
| "grad_norm": 2.18556547164917, | |
| "learning_rate": 4.882352941176471e-05, | |
| "loss": 0.4042, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.733993399339934, | |
| "grad_norm": 1.9759682416915894, | |
| "learning_rate": 4.877450980392157e-05, | |
| "loss": 0.3694, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.7392739273927392, | |
| "grad_norm": 1.8205828666687012, | |
| "learning_rate": 4.872549019607843e-05, | |
| "loss": 0.244, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.7445544554455445, | |
| "grad_norm": 1.8210268020629883, | |
| "learning_rate": 4.86764705882353e-05, | |
| "loss": 0.4084, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.7498349834983499, | |
| "grad_norm": 2.2197041511535645, | |
| "learning_rate": 4.862745098039216e-05, | |
| "loss": 0.5236, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.7551155115511551, | |
| "grad_norm": 2.137676239013672, | |
| "learning_rate": 4.8578431372549024e-05, | |
| "loss": 0.3302, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.7603960396039604, | |
| "grad_norm": 2.126865863800049, | |
| "learning_rate": 4.8529411764705885e-05, | |
| "loss": 0.4265, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.7656765676567657, | |
| "grad_norm": 1.828809380531311, | |
| "learning_rate": 4.8480392156862745e-05, | |
| "loss": 0.4363, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.7709570957095709, | |
| "grad_norm": 1.7918983697891235, | |
| "learning_rate": 4.843137254901961e-05, | |
| "loss": 0.3568, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.7762376237623763, | |
| "grad_norm": 1.7884886264801025, | |
| "learning_rate": 4.838235294117647e-05, | |
| "loss": 0.3338, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.7815181518151815, | |
| "grad_norm": 1.8494501113891602, | |
| "learning_rate": 4.8333333333333334e-05, | |
| "loss": 0.3311, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.7867986798679868, | |
| "grad_norm": 2.0265438556671143, | |
| "learning_rate": 4.82843137254902e-05, | |
| "loss": 0.4381, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.7920792079207921, | |
| "grad_norm": 2.0035383701324463, | |
| "learning_rate": 4.823529411764706e-05, | |
| "loss": 0.4291, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.7973597359735973, | |
| "grad_norm": 1.8268975019454956, | |
| "learning_rate": 4.818627450980392e-05, | |
| "loss": 0.464, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.8026402640264027, | |
| "grad_norm": 1.9028264284133911, | |
| "learning_rate": 4.813725490196079e-05, | |
| "loss": 0.3807, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.807920792079208, | |
| "grad_norm": 1.8994662761688232, | |
| "learning_rate": 4.808823529411765e-05, | |
| "loss": 0.3904, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.8132013201320132, | |
| "grad_norm": 1.832381248474121, | |
| "learning_rate": 4.803921568627452e-05, | |
| "loss": 0.3856, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.8184818481848185, | |
| "grad_norm": 1.877752661705017, | |
| "learning_rate": 4.799019607843138e-05, | |
| "loss": 0.3185, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.8237623762376237, | |
| "grad_norm": 2.0872726440429688, | |
| "learning_rate": 4.794117647058824e-05, | |
| "loss": 0.3217, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.829042904290429, | |
| "grad_norm": 1.8779263496398926, | |
| "learning_rate": 4.7892156862745105e-05, | |
| "loss": 0.3654, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.8343234323432344, | |
| "grad_norm": 1.996422529220581, | |
| "learning_rate": 4.7843137254901966e-05, | |
| "loss": 0.3524, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.8396039603960396, | |
| "grad_norm": 1.5910488367080688, | |
| "learning_rate": 4.7794117647058826e-05, | |
| "loss": 0.2859, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.8448844884488449, | |
| "grad_norm": 1.9424618482589722, | |
| "learning_rate": 4.774509803921569e-05, | |
| "loss": 0.3422, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.8501650165016502, | |
| "grad_norm": 1.9187934398651123, | |
| "learning_rate": 4.7696078431372554e-05, | |
| "loss": 0.3807, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.8554455445544554, | |
| "grad_norm": 1.7809456586837769, | |
| "learning_rate": 4.7647058823529414e-05, | |
| "loss": 0.2991, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.8607260726072608, | |
| "grad_norm": 1.9575221538543701, | |
| "learning_rate": 4.7598039215686275e-05, | |
| "loss": 0.4688, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.866006600660066, | |
| "grad_norm": 1.8438433408737183, | |
| "learning_rate": 4.7549019607843135e-05, | |
| "loss": 0.3691, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.8712871287128713, | |
| "grad_norm": 1.9522879123687744, | |
| "learning_rate": 4.75e-05, | |
| "loss": 0.3987, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.8765676567656766, | |
| "grad_norm": 2.243354320526123, | |
| "learning_rate": 4.745098039215686e-05, | |
| "loss": 0.4836, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.8818481848184818, | |
| "grad_norm": 1.8762164115905762, | |
| "learning_rate": 4.7401960784313724e-05, | |
| "loss": 0.4691, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.8871287128712871, | |
| "grad_norm": 1.8055609464645386, | |
| "learning_rate": 4.735294117647059e-05, | |
| "loss": 0.2506, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.8924092409240925, | |
| "grad_norm": 1.8521029949188232, | |
| "learning_rate": 4.730392156862745e-05, | |
| "loss": 0.3291, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.8976897689768977, | |
| "grad_norm": 1.939030647277832, | |
| "learning_rate": 4.725490196078431e-05, | |
| "loss": 0.4104, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.902970297029703, | |
| "grad_norm": 1.8853607177734375, | |
| "learning_rate": 4.720588235294118e-05, | |
| "loss": 0.3797, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.9082508250825082, | |
| "grad_norm": 2.0953316688537598, | |
| "learning_rate": 4.715686274509804e-05, | |
| "loss": 0.324, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.9135313531353135, | |
| "grad_norm": 1.9342799186706543, | |
| "learning_rate": 4.71078431372549e-05, | |
| "loss": 0.4306, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.9188118811881189, | |
| "grad_norm": 1.8248006105422974, | |
| "learning_rate": 4.705882352941177e-05, | |
| "loss": 0.3388, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.9240924092409241, | |
| "grad_norm": 1.9689913988113403, | |
| "learning_rate": 4.700980392156863e-05, | |
| "loss": 0.3591, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.9293729372937294, | |
| "grad_norm": 2.017063856124878, | |
| "learning_rate": 4.6960784313725495e-05, | |
| "loss": 0.4723, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.9346534653465347, | |
| "grad_norm": 1.9692254066467285, | |
| "learning_rate": 4.6911764705882356e-05, | |
| "loss": 0.3893, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.9399339933993399, | |
| "grad_norm": 1.9935567378997803, | |
| "learning_rate": 4.6862745098039216e-05, | |
| "loss": 0.3938, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.9452145214521452, | |
| "grad_norm": 1.7153037786483765, | |
| "learning_rate": 4.681372549019608e-05, | |
| "loss": 0.2368, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.9504950495049505, | |
| "grad_norm": 1.4944133758544922, | |
| "learning_rate": 4.6764705882352944e-05, | |
| "loss": 0.2181, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.9557755775577558, | |
| "grad_norm": 1.9143524169921875, | |
| "learning_rate": 4.6715686274509804e-05, | |
| "loss": 0.413, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.9610561056105611, | |
| "grad_norm": 2.162576675415039, | |
| "learning_rate": 4.666666666666667e-05, | |
| "loss": 0.3745, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.9663366336633663, | |
| "grad_norm": 1.8236726522445679, | |
| "learning_rate": 4.661764705882353e-05, | |
| "loss": 0.3056, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.9716171617161716, | |
| "grad_norm": 1.9680614471435547, | |
| "learning_rate": 4.656862745098039e-05, | |
| "loss": 0.3875, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.976897689768977, | |
| "grad_norm": 1.575900912284851, | |
| "learning_rate": 4.651960784313726e-05, | |
| "loss": 0.1831, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.9821782178217822, | |
| "grad_norm": 2.6015613079071045, | |
| "learning_rate": 4.647058823529412e-05, | |
| "loss": 0.4227, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.9874587458745875, | |
| "grad_norm": 2.066946268081665, | |
| "learning_rate": 4.642156862745098e-05, | |
| "loss": 0.4256, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.9927392739273927, | |
| "grad_norm": 2.1683449745178223, | |
| "learning_rate": 4.637254901960785e-05, | |
| "loss": 0.3943, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.998019801980198, | |
| "grad_norm": 1.876991629600525, | |
| "learning_rate": 4.632352941176471e-05, | |
| "loss": 0.4049, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 1.0033003300330032, | |
| "grad_norm": 2.2598772048950195, | |
| "learning_rate": 4.6274509803921576e-05, | |
| "loss": 0.3749, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 1.0085808580858087, | |
| "grad_norm": 1.8292592763900757, | |
| "learning_rate": 4.6225490196078436e-05, | |
| "loss": 0.2654, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 1.0138613861386139, | |
| "grad_norm": 2.6986138820648193, | |
| "learning_rate": 4.61764705882353e-05, | |
| "loss": 0.3662, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 1.019141914191419, | |
| "grad_norm": 1.9084346294403076, | |
| "learning_rate": 4.6127450980392164e-05, | |
| "loss": 0.2929, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 1.0244224422442245, | |
| "grad_norm": 1.6963775157928467, | |
| "learning_rate": 4.607843137254902e-05, | |
| "loss": 0.2481, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 1.0297029702970297, | |
| "grad_norm": 2.524332046508789, | |
| "learning_rate": 4.6029411764705885e-05, | |
| "loss": 0.2213, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 1.034983498349835, | |
| "grad_norm": 1.945142388343811, | |
| "learning_rate": 4.5980392156862746e-05, | |
| "loss": 0.2892, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 1.0402640264026402, | |
| "grad_norm": 2.3402678966522217, | |
| "learning_rate": 4.5931372549019606e-05, | |
| "loss": 0.2232, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 1.0455445544554456, | |
| "grad_norm": 1.7755571603775024, | |
| "learning_rate": 4.588235294117647e-05, | |
| "loss": 0.2416, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 1.0508250825082508, | |
| "grad_norm": 2.110517740249634, | |
| "learning_rate": 4.5833333333333334e-05, | |
| "loss": 0.2557, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 1.056105610561056, | |
| "grad_norm": 1.7219949960708618, | |
| "learning_rate": 4.5784313725490194e-05, | |
| "loss": 0.2274, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 1.0613861386138614, | |
| "grad_norm": 1.9697656631469727, | |
| "learning_rate": 4.573529411764706e-05, | |
| "loss": 0.2588, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 1.0666666666666667, | |
| "grad_norm": 1.5107744932174683, | |
| "learning_rate": 4.568627450980392e-05, | |
| "loss": 0.2175, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 1.0719471947194719, | |
| "grad_norm": 1.8557658195495605, | |
| "learning_rate": 4.563725490196078e-05, | |
| "loss": 0.2901, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 1.0772277227722773, | |
| "grad_norm": 1.764145851135254, | |
| "learning_rate": 4.558823529411765e-05, | |
| "loss": 0.176, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 1.0825082508250825, | |
| "grad_norm": 1.8334012031555176, | |
| "learning_rate": 4.553921568627451e-05, | |
| "loss": 0.2572, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 1.0877887788778877, | |
| "grad_norm": 1.8206666707992554, | |
| "learning_rate": 4.549019607843137e-05, | |
| "loss": 0.248, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 1.0930693069306932, | |
| "grad_norm": 2.0776381492614746, | |
| "learning_rate": 4.544117647058824e-05, | |
| "loss": 0.3113, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 1.0983498349834984, | |
| "grad_norm": 1.8429386615753174, | |
| "learning_rate": 4.53921568627451e-05, | |
| "loss": 0.2849, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 1.1036303630363036, | |
| "grad_norm": 1.7320504188537598, | |
| "learning_rate": 4.5343137254901966e-05, | |
| "loss": 0.2145, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 1.108910891089109, | |
| "grad_norm": 1.7207646369934082, | |
| "learning_rate": 4.5294117647058826e-05, | |
| "loss": 0.2013, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 1.1141914191419142, | |
| "grad_norm": 2.134873151779175, | |
| "learning_rate": 4.524509803921569e-05, | |
| "loss": 0.3293, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 1.1194719471947194, | |
| "grad_norm": 1.7931280136108398, | |
| "learning_rate": 4.5196078431372554e-05, | |
| "loss": 0.2292, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 1.1247524752475249, | |
| "grad_norm": 2.1878650188446045, | |
| "learning_rate": 4.5147058823529415e-05, | |
| "loss": 0.3173, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 1.13003300330033, | |
| "grad_norm": 1.8994349241256714, | |
| "learning_rate": 4.5098039215686275e-05, | |
| "loss": 0.2742, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 1.1353135313531353, | |
| "grad_norm": 1.9204659461975098, | |
| "learning_rate": 4.504901960784314e-05, | |
| "loss": 0.2713, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 1.1405940594059407, | |
| "grad_norm": 1.9061977863311768, | |
| "learning_rate": 4.5e-05, | |
| "loss": 0.2433, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 1.145874587458746, | |
| "grad_norm": 2.254232168197632, | |
| "learning_rate": 4.495098039215686e-05, | |
| "loss": 0.3523, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 1.1511551155115511, | |
| "grad_norm": 1.9071446657180786, | |
| "learning_rate": 4.490196078431373e-05, | |
| "loss": 0.2196, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 1.1564356435643564, | |
| "grad_norm": 1.8089710474014282, | |
| "learning_rate": 4.485294117647059e-05, | |
| "loss": 0.2907, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 1.1617161716171618, | |
| "grad_norm": 1.9056932926177979, | |
| "learning_rate": 4.480392156862745e-05, | |
| "loss": 0.2564, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 1.166996699669967, | |
| "grad_norm": 1.9336401224136353, | |
| "learning_rate": 4.475490196078432e-05, | |
| "loss": 0.205, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 1.1722772277227722, | |
| "grad_norm": 1.9126192331314087, | |
| "learning_rate": 4.470588235294118e-05, | |
| "loss": 0.2647, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 1.1775577557755776, | |
| "grad_norm": 1.8508714437484741, | |
| "learning_rate": 4.4656862745098047e-05, | |
| "loss": 0.262, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 1.1828382838283829, | |
| "grad_norm": 1.478278398513794, | |
| "learning_rate": 4.460784313725491e-05, | |
| "loss": 0.1711, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 1.188118811881188, | |
| "grad_norm": 1.7818694114685059, | |
| "learning_rate": 4.455882352941177e-05, | |
| "loss": 0.2043, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 1.1933993399339933, | |
| "grad_norm": 1.916344404220581, | |
| "learning_rate": 4.450980392156863e-05, | |
| "loss": 0.3064, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 1.1986798679867987, | |
| "grad_norm": 1.8637932538986206, | |
| "learning_rate": 4.446078431372549e-05, | |
| "loss": 0.3247, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 1.203960396039604, | |
| "grad_norm": 1.7257781028747559, | |
| "learning_rate": 4.4411764705882356e-05, | |
| "loss": 0.1981, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 1.2092409240924091, | |
| "grad_norm": 1.9121214151382446, | |
| "learning_rate": 4.4362745098039216e-05, | |
| "loss": 0.2963, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 1.2145214521452146, | |
| "grad_norm": 1.5968807935714722, | |
| "learning_rate": 4.431372549019608e-05, | |
| "loss": 0.1802, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 1.2198019801980198, | |
| "grad_norm": 1.7256313562393188, | |
| "learning_rate": 4.4264705882352944e-05, | |
| "loss": 0.2442, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 1.225082508250825, | |
| "grad_norm": 1.7865731716156006, | |
| "learning_rate": 4.4215686274509805e-05, | |
| "loss": 0.2683, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 1.2303630363036304, | |
| "grad_norm": 1.7361854314804077, | |
| "learning_rate": 4.4166666666666665e-05, | |
| "loss": 0.2285, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 1.2356435643564356, | |
| "grad_norm": 1.8758944272994995, | |
| "learning_rate": 4.411764705882353e-05, | |
| "loss": 0.3081, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 1.2409240924092408, | |
| "grad_norm": 2.000033140182495, | |
| "learning_rate": 4.406862745098039e-05, | |
| "loss": 0.2405, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 1.2462046204620463, | |
| "grad_norm": 1.8750522136688232, | |
| "learning_rate": 4.401960784313725e-05, | |
| "loss": 0.2778, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 1.2514851485148515, | |
| "grad_norm": 1.7535063028335571, | |
| "learning_rate": 4.397058823529412e-05, | |
| "loss": 0.2045, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 1.2567656765676567, | |
| "grad_norm": 1.9849064350128174, | |
| "learning_rate": 4.392156862745098e-05, | |
| "loss": 0.3018, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 1.2620462046204621, | |
| "grad_norm": 1.8400393724441528, | |
| "learning_rate": 4.387254901960784e-05, | |
| "loss": 0.2555, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 1.2673267326732673, | |
| "grad_norm": 1.8575385808944702, | |
| "learning_rate": 4.382352941176471e-05, | |
| "loss": 0.2783, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 1.2726072607260726, | |
| "grad_norm": 1.8452024459838867, | |
| "learning_rate": 4.377450980392157e-05, | |
| "loss": 0.3091, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 1.277887788778878, | |
| "grad_norm": 1.9682793617248535, | |
| "learning_rate": 4.3725490196078437e-05, | |
| "loss": 0.2169, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 1.2831683168316832, | |
| "grad_norm": 1.7926579713821411, | |
| "learning_rate": 4.36764705882353e-05, | |
| "loss": 0.27, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 1.2884488448844884, | |
| "grad_norm": 1.690425157546997, | |
| "learning_rate": 4.362745098039216e-05, | |
| "loss": 0.2336, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 1.2937293729372938, | |
| "grad_norm": 1.775240421295166, | |
| "learning_rate": 4.3578431372549025e-05, | |
| "loss": 0.1981, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 1.299009900990099, | |
| "grad_norm": 1.7951467037200928, | |
| "learning_rate": 4.3529411764705885e-05, | |
| "loss": 0.2195, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 1.3042904290429043, | |
| "grad_norm": 1.5444797277450562, | |
| "learning_rate": 4.3480392156862746e-05, | |
| "loss": 0.2264, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 1.3095709570957097, | |
| "grad_norm": 1.731550931930542, | |
| "learning_rate": 4.343137254901961e-05, | |
| "loss": 0.1676, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 1.314851485148515, | |
| "grad_norm": 1.747083306312561, | |
| "learning_rate": 4.3382352941176474e-05, | |
| "loss": 0.28, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 1.3201320132013201, | |
| "grad_norm": 1.7961376905441284, | |
| "learning_rate": 4.3333333333333334e-05, | |
| "loss": 0.2003, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 1.3254125412541253, | |
| "grad_norm": 1.8366891145706177, | |
| "learning_rate": 4.32843137254902e-05, | |
| "loss": 0.2093, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 1.3306930693069308, | |
| "grad_norm": 1.4352390766143799, | |
| "learning_rate": 4.323529411764706e-05, | |
| "loss": 0.1597, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 1.335973597359736, | |
| "grad_norm": 1.9146888256072998, | |
| "learning_rate": 4.318627450980392e-05, | |
| "loss": 0.2425, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 1.3412541254125412, | |
| "grad_norm": 1.4383189678192139, | |
| "learning_rate": 4.313725490196079e-05, | |
| "loss": 0.1943, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 1.3465346534653464, | |
| "grad_norm": 1.5246001482009888, | |
| "learning_rate": 4.308823529411765e-05, | |
| "loss": 0.1543, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 1.3518151815181518, | |
| "grad_norm": 1.4863159656524658, | |
| "learning_rate": 4.303921568627452e-05, | |
| "loss": 0.2441, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 1.357095709570957, | |
| "grad_norm": 1.6645705699920654, | |
| "learning_rate": 4.299019607843138e-05, | |
| "loss": 0.2329, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 1.3623762376237623, | |
| "grad_norm": 1.946554183959961, | |
| "learning_rate": 4.294117647058823e-05, | |
| "loss": 0.1952, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 1.3676567656765677, | |
| "grad_norm": 2.0372443199157715, | |
| "learning_rate": 4.28921568627451e-05, | |
| "loss": 0.332, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 1.372937293729373, | |
| "grad_norm": 1.846138834953308, | |
| "learning_rate": 4.284313725490196e-05, | |
| "loss": 0.1408, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 1.378217821782178, | |
| "grad_norm": 1.5724695920944214, | |
| "learning_rate": 4.2794117647058827e-05, | |
| "loss": 0.1926, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 1.3834983498349835, | |
| "grad_norm": 2.1506614685058594, | |
| "learning_rate": 4.274509803921569e-05, | |
| "loss": 0.217, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 1.3887788778877888, | |
| "grad_norm": 2.0763325691223145, | |
| "learning_rate": 4.269607843137255e-05, | |
| "loss": 0.2871, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 1.394059405940594, | |
| "grad_norm": 1.9296153783798218, | |
| "learning_rate": 4.2647058823529415e-05, | |
| "loss": 0.2655, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 1.3993399339933994, | |
| "grad_norm": 1.7979801893234253, | |
| "learning_rate": 4.2598039215686275e-05, | |
| "loss": 0.2715, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 1.4046204620462046, | |
| "grad_norm": 1.4527943134307861, | |
| "learning_rate": 4.2549019607843136e-05, | |
| "loss": 0.203, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 1.4099009900990098, | |
| "grad_norm": 1.8454203605651855, | |
| "learning_rate": 4.25e-05, | |
| "loss": 0.198, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 1.4151815181518153, | |
| "grad_norm": 1.6438169479370117, | |
| "learning_rate": 4.2450980392156864e-05, | |
| "loss": 0.2056, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 1.4204620462046205, | |
| "grad_norm": 1.5819754600524902, | |
| "learning_rate": 4.2401960784313724e-05, | |
| "loss": 0.2154, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 1.4257425742574257, | |
| "grad_norm": 1.800973653793335, | |
| "learning_rate": 4.235294117647059e-05, | |
| "loss": 0.2536, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 1.431023102310231, | |
| "grad_norm": 1.6425402164459229, | |
| "learning_rate": 4.230392156862745e-05, | |
| "loss": 0.2111, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 1.4363036303630363, | |
| "grad_norm": 1.908632755279541, | |
| "learning_rate": 4.225490196078431e-05, | |
| "loss": 0.2822, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 1.4415841584158415, | |
| "grad_norm": 2.028026580810547, | |
| "learning_rate": 4.220588235294118e-05, | |
| "loss": 0.3118, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 1.446864686468647, | |
| "grad_norm": 1.9891923666000366, | |
| "learning_rate": 4.215686274509804e-05, | |
| "loss": 0.2469, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 1.4521452145214522, | |
| "grad_norm": 1.771210789680481, | |
| "learning_rate": 4.210784313725491e-05, | |
| "loss": 0.2057, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 1.4574257425742574, | |
| "grad_norm": 1.8335461616516113, | |
| "learning_rate": 4.205882352941177e-05, | |
| "loss": 0.203, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 1.4627062706270628, | |
| "grad_norm": 1.4403390884399414, | |
| "learning_rate": 4.200980392156863e-05, | |
| "loss": 0.1652, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 1.467986798679868, | |
| "grad_norm": 2.0193352699279785, | |
| "learning_rate": 4.1960784313725496e-05, | |
| "loss": 0.2914, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 1.4732673267326732, | |
| "grad_norm": 1.5873808860778809, | |
| "learning_rate": 4.1911764705882356e-05, | |
| "loss": 0.1871, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 1.4785478547854787, | |
| "grad_norm": 2.0619425773620605, | |
| "learning_rate": 4.1862745098039217e-05, | |
| "loss": 0.2449, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 1.4838283828382839, | |
| "grad_norm": 1.777978539466858, | |
| "learning_rate": 4.1813725490196084e-05, | |
| "loss": 0.2585, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 1.489108910891089, | |
| "grad_norm": 1.7166889905929565, | |
| "learning_rate": 4.1764705882352944e-05, | |
| "loss": 0.2648, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 1.4943894389438943, | |
| "grad_norm": 1.7009400129318237, | |
| "learning_rate": 4.1715686274509805e-05, | |
| "loss": 0.1847, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 1.4996699669966997, | |
| "grad_norm": 2.067512035369873, | |
| "learning_rate": 4.166666666666667e-05, | |
| "loss": 0.2803, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 1.504950495049505, | |
| "grad_norm": 1.6885766983032227, | |
| "learning_rate": 4.161764705882353e-05, | |
| "loss": 0.2037, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 1.5102310231023104, | |
| "grad_norm": 1.9202988147735596, | |
| "learning_rate": 4.156862745098039e-05, | |
| "loss": 0.2791, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 1.5155115511551154, | |
| "grad_norm": 1.6683584451675415, | |
| "learning_rate": 4.151960784313726e-05, | |
| "loss": 0.2228, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 1.5207920792079208, | |
| "grad_norm": 1.783361792564392, | |
| "learning_rate": 4.147058823529412e-05, | |
| "loss": 0.2524, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 1.526072607260726, | |
| "grad_norm": 1.9562329053878784, | |
| "learning_rate": 4.142156862745099e-05, | |
| "loss": 0.2773, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 1.5313531353135312, | |
| "grad_norm": 1.4846049547195435, | |
| "learning_rate": 4.137254901960784e-05, | |
| "loss": 0.2528, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 1.5366336633663367, | |
| "grad_norm": 1.6068270206451416, | |
| "learning_rate": 4.13235294117647e-05, | |
| "loss": 0.1513, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 1.5419141914191419, | |
| "grad_norm": 1.8563951253890991, | |
| "learning_rate": 4.127450980392157e-05, | |
| "loss": 0.2518, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 1.547194719471947, | |
| "grad_norm": 1.9122540950775146, | |
| "learning_rate": 4.122549019607843e-05, | |
| "loss": 0.2235, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 1.5524752475247525, | |
| "grad_norm": 1.619687557220459, | |
| "learning_rate": 4.11764705882353e-05, | |
| "loss": 0.236, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 1.5577557755775577, | |
| "grad_norm": 1.9006292819976807, | |
| "learning_rate": 4.112745098039216e-05, | |
| "loss": 0.2688, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 1.563036303630363, | |
| "grad_norm": 1.5912319421768188, | |
| "learning_rate": 4.107843137254902e-05, | |
| "loss": 0.2036, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 1.5683168316831684, | |
| "grad_norm": 1.9365366697311401, | |
| "learning_rate": 4.1029411764705886e-05, | |
| "loss": 0.2708, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 1.5735973597359736, | |
| "grad_norm": 1.535831332206726, | |
| "learning_rate": 4.0980392156862746e-05, | |
| "loss": 0.1875, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 1.5788778877887788, | |
| "grad_norm": 2.117027997970581, | |
| "learning_rate": 4.0931372549019607e-05, | |
| "loss": 0.3118, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 1.5841584158415842, | |
| "grad_norm": 1.8837215900421143, | |
| "learning_rate": 4.0882352941176474e-05, | |
| "loss": 0.2084, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 1.5894389438943894, | |
| "grad_norm": 1.502886176109314, | |
| "learning_rate": 4.0833333333333334e-05, | |
| "loss": 0.1348, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 1.5947194719471947, | |
| "grad_norm": 1.6588914394378662, | |
| "learning_rate": 4.0784313725490195e-05, | |
| "loss": 0.1775, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "grad_norm": 1.6253366470336914, | |
| "learning_rate": 4.073529411764706e-05, | |
| "loss": 0.2719, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 1.6052805280528053, | |
| "grad_norm": 1.9440994262695312, | |
| "learning_rate": 4.068627450980392e-05, | |
| "loss": 0.2278, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 1.6105610561056105, | |
| "grad_norm": 1.7787673473358154, | |
| "learning_rate": 4.063725490196078e-05, | |
| "loss": 0.2513, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 1.615841584158416, | |
| "grad_norm": 1.3408194780349731, | |
| "learning_rate": 4.058823529411765e-05, | |
| "loss": 0.1134, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 1.6211221122112212, | |
| "grad_norm": 1.739343523979187, | |
| "learning_rate": 4.053921568627451e-05, | |
| "loss": 0.194, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 1.6264026402640264, | |
| "grad_norm": 1.7391927242279053, | |
| "learning_rate": 4.049019607843138e-05, | |
| "loss": 0.2273, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 1.6316831683168318, | |
| "grad_norm": 2.134809732437134, | |
| "learning_rate": 4.044117647058824e-05, | |
| "loss": 0.349, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 1.636963696369637, | |
| "grad_norm": 1.8628054857254028, | |
| "learning_rate": 4.03921568627451e-05, | |
| "loss": 0.2687, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 1.6422442244224422, | |
| "grad_norm": 1.8895047903060913, | |
| "learning_rate": 4.0343137254901966e-05, | |
| "loss": 0.2684, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 1.6475247524752477, | |
| "grad_norm": 1.784513235092163, | |
| "learning_rate": 4.029411764705883e-05, | |
| "loss": 0.2728, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 1.6528052805280526, | |
| "grad_norm": 1.505031704902649, | |
| "learning_rate": 4.024509803921569e-05, | |
| "loss": 0.1806, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 1.658085808580858, | |
| "grad_norm": 1.9005438089370728, | |
| "learning_rate": 4.0196078431372555e-05, | |
| "loss": 0.3178, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 1.6633663366336635, | |
| "grad_norm": 1.769161581993103, | |
| "learning_rate": 4.0147058823529415e-05, | |
| "loss": 0.2392, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 1.6686468646864685, | |
| "grad_norm": 1.8065416812896729, | |
| "learning_rate": 4.0098039215686276e-05, | |
| "loss": 0.2972, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 1.673927392739274, | |
| "grad_norm": 1.6391319036483765, | |
| "learning_rate": 4.004901960784314e-05, | |
| "loss": 0.2132, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 1.6792079207920794, | |
| "grad_norm": 1.3760650157928467, | |
| "learning_rate": 4e-05, | |
| "loss": 0.1106, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 1.6844884488448844, | |
| "grad_norm": 1.5832698345184326, | |
| "learning_rate": 3.9950980392156864e-05, | |
| "loss": 0.1965, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 1.6897689768976898, | |
| "grad_norm": 1.830043911933899, | |
| "learning_rate": 3.990196078431373e-05, | |
| "loss": 0.2658, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 1.695049504950495, | |
| "grad_norm": 1.8273866176605225, | |
| "learning_rate": 3.985294117647059e-05, | |
| "loss": 0.2284, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 1.7003300330033002, | |
| "grad_norm": 1.6484299898147583, | |
| "learning_rate": 3.980392156862745e-05, | |
| "loss": 0.1503, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 1.7056105610561056, | |
| "grad_norm": 1.7297075986862183, | |
| "learning_rate": 3.975490196078431e-05, | |
| "loss": 0.1818, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 1.7108910891089109, | |
| "grad_norm": 1.9557067155838013, | |
| "learning_rate": 3.970588235294117e-05, | |
| "loss": 0.3179, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 1.716171617161716, | |
| "grad_norm": 1.8135654926300049, | |
| "learning_rate": 3.965686274509804e-05, | |
| "loss": 0.1871, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 1.7214521452145215, | |
| "grad_norm": 1.9970617294311523, | |
| "learning_rate": 3.96078431372549e-05, | |
| "loss": 0.2435, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 1.7267326732673267, | |
| "grad_norm": 1.584839105606079, | |
| "learning_rate": 3.955882352941177e-05, | |
| "loss": 0.1813, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 1.732013201320132, | |
| "grad_norm": 1.5239087343215942, | |
| "learning_rate": 3.950980392156863e-05, | |
| "loss": 0.1193, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 1.7372937293729374, | |
| "grad_norm": 1.514541506767273, | |
| "learning_rate": 3.946078431372549e-05, | |
| "loss": 0.1658, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 1.7425742574257426, | |
| "grad_norm": 1.8841454982757568, | |
| "learning_rate": 3.9411764705882356e-05, | |
| "loss": 0.3376, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 1.7478547854785478, | |
| "grad_norm": 1.5418506860733032, | |
| "learning_rate": 3.936274509803922e-05, | |
| "loss": 0.2179, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 1.7531353135313532, | |
| "grad_norm": 1.637239933013916, | |
| "learning_rate": 3.931372549019608e-05, | |
| "loss": 0.1976, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 1.7584158415841584, | |
| "grad_norm": 1.8015220165252686, | |
| "learning_rate": 3.9264705882352945e-05, | |
| "loss": 0.2629, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 1.7636963696369636, | |
| "grad_norm": 1.7710020542144775, | |
| "learning_rate": 3.9215686274509805e-05, | |
| "loss": 0.2208, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 1.768976897689769, | |
| "grad_norm": 1.9169963598251343, | |
| "learning_rate": 3.9166666666666665e-05, | |
| "loss": 0.2811, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 1.7742574257425743, | |
| "grad_norm": 1.738980770111084, | |
| "learning_rate": 3.911764705882353e-05, | |
| "loss": 0.2009, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 1.7795379537953795, | |
| "grad_norm": 1.888925313949585, | |
| "learning_rate": 3.906862745098039e-05, | |
| "loss": 0.2615, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 1.784818481848185, | |
| "grad_norm": 1.8248200416564941, | |
| "learning_rate": 3.9019607843137254e-05, | |
| "loss": 0.2284, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 1.7900990099009901, | |
| "grad_norm": 1.6162784099578857, | |
| "learning_rate": 3.897058823529412e-05, | |
| "loss": 0.2206, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 1.7953795379537953, | |
| "grad_norm": 1.441273808479309, | |
| "learning_rate": 3.892156862745098e-05, | |
| "loss": 0.1806, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 1.8006600660066008, | |
| "grad_norm": 1.815674066543579, | |
| "learning_rate": 3.887254901960785e-05, | |
| "loss": 0.2169, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 1.805940594059406, | |
| "grad_norm": 1.6441361904144287, | |
| "learning_rate": 3.882352941176471e-05, | |
| "loss": 0.2177, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 1.8112211221122112, | |
| "grad_norm": 1.923663854598999, | |
| "learning_rate": 3.877450980392157e-05, | |
| "loss": 0.2942, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 1.8165016501650166, | |
| "grad_norm": 1.978797197341919, | |
| "learning_rate": 3.872549019607844e-05, | |
| "loss": 0.2693, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 1.8217821782178216, | |
| "grad_norm": 1.8134146928787231, | |
| "learning_rate": 3.86764705882353e-05, | |
| "loss": 0.2713, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 1.827062706270627, | |
| "grad_norm": 1.7703922986984253, | |
| "learning_rate": 3.862745098039216e-05, | |
| "loss": 0.2413, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 1.8323432343234325, | |
| "grad_norm": 1.7030301094055176, | |
| "learning_rate": 3.8578431372549025e-05, | |
| "loss": 0.1866, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 1.8376237623762375, | |
| "grad_norm": 1.521941065788269, | |
| "learning_rate": 3.8529411764705886e-05, | |
| "loss": 0.174, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 1.842904290429043, | |
| "grad_norm": 1.8277267217636108, | |
| "learning_rate": 3.8480392156862746e-05, | |
| "loss": 0.2077, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 1.8481848184818483, | |
| "grad_norm": 2.028367042541504, | |
| "learning_rate": 3.8431372549019614e-05, | |
| "loss": 0.2073, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 1.8534653465346533, | |
| "grad_norm": 1.373708724975586, | |
| "learning_rate": 3.8382352941176474e-05, | |
| "loss": 0.1383, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 1.8587458745874588, | |
| "grad_norm": 1.6797735691070557, | |
| "learning_rate": 3.8333333333333334e-05, | |
| "loss": 0.1971, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 1.864026402640264, | |
| "grad_norm": 1.4328402280807495, | |
| "learning_rate": 3.82843137254902e-05, | |
| "loss": 0.1601, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 1.8693069306930692, | |
| "grad_norm": 1.8246557712554932, | |
| "learning_rate": 3.8235294117647055e-05, | |
| "loss": 0.2408, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 1.8745874587458746, | |
| "grad_norm": 1.6448115110397339, | |
| "learning_rate": 3.818627450980392e-05, | |
| "loss": 0.2677, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 1.8798679867986798, | |
| "grad_norm": 1.6840052604675293, | |
| "learning_rate": 3.813725490196078e-05, | |
| "loss": 0.1955, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 1.885148514851485, | |
| "grad_norm": 1.989355206489563, | |
| "learning_rate": 3.8088235294117644e-05, | |
| "loss": 0.2761, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 1.8904290429042905, | |
| "grad_norm": 1.5983843803405762, | |
| "learning_rate": 3.803921568627451e-05, | |
| "loss": 0.2057, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 1.8957095709570957, | |
| "grad_norm": 1.5310300588607788, | |
| "learning_rate": 3.799019607843137e-05, | |
| "loss": 0.1787, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 1.900990099009901, | |
| "grad_norm": 1.835742712020874, | |
| "learning_rate": 3.794117647058824e-05, | |
| "loss": 0.2972, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 1.9062706270627063, | |
| "grad_norm": 1.890306830406189, | |
| "learning_rate": 3.78921568627451e-05, | |
| "loss": 0.2972, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 1.9115511551155115, | |
| "grad_norm": 2.0529651641845703, | |
| "learning_rate": 3.784313725490196e-05, | |
| "loss": 0.3286, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 1.9168316831683168, | |
| "grad_norm": 1.742149829864502, | |
| "learning_rate": 3.779411764705883e-05, | |
| "loss": 0.2106, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 1.9221122112211222, | |
| "grad_norm": 1.8192955255508423, | |
| "learning_rate": 3.774509803921569e-05, | |
| "loss": 0.2289, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 1.9273927392739274, | |
| "grad_norm": 1.7159464359283447, | |
| "learning_rate": 3.769607843137255e-05, | |
| "loss": 0.2324, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 1.9326732673267326, | |
| "grad_norm": 1.7470611333847046, | |
| "learning_rate": 3.7647058823529415e-05, | |
| "loss": 0.1381, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 1.937953795379538, | |
| "grad_norm": 1.7340490818023682, | |
| "learning_rate": 3.7598039215686276e-05, | |
| "loss": 0.2136, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 1.9432343234323433, | |
| "grad_norm": 1.6391620635986328, | |
| "learning_rate": 3.7549019607843136e-05, | |
| "loss": 0.2178, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 1.9485148514851485, | |
| "grad_norm": 1.5655213594436646, | |
| "learning_rate": 3.7500000000000003e-05, | |
| "loss": 0.2091, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 1.953795379537954, | |
| "grad_norm": 1.7240495681762695, | |
| "learning_rate": 3.7450980392156864e-05, | |
| "loss": 0.1779, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 1.9590759075907591, | |
| "grad_norm": 1.5939252376556396, | |
| "learning_rate": 3.7401960784313724e-05, | |
| "loss": 0.1863, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 1.9643564356435643, | |
| "grad_norm": 1.5324146747589111, | |
| "learning_rate": 3.735294117647059e-05, | |
| "loss": 0.1767, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 1.9696369636963698, | |
| "grad_norm": 1.689475417137146, | |
| "learning_rate": 3.730392156862745e-05, | |
| "loss": 0.2135, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 1.974917491749175, | |
| "grad_norm": 1.9409862756729126, | |
| "learning_rate": 3.725490196078432e-05, | |
| "loss": 0.2633, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 1.9801980198019802, | |
| "grad_norm": 1.743085265159607, | |
| "learning_rate": 3.720588235294118e-05, | |
| "loss": 0.2098, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 1.9854785478547856, | |
| "grad_norm": 1.6699271202087402, | |
| "learning_rate": 3.715686274509804e-05, | |
| "loss": 0.2288, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 1.9907590759075906, | |
| "grad_norm": 1.662705421447754, | |
| "learning_rate": 3.710784313725491e-05, | |
| "loss": 0.2385, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 1.996039603960396, | |
| "grad_norm": 1.4365413188934326, | |
| "learning_rate": 3.705882352941177e-05, | |
| "loss": 0.1793, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 2.0013201320132015, | |
| "grad_norm": 1.5893417596817017, | |
| "learning_rate": 3.700980392156863e-05, | |
| "loss": 0.2293, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 2.0066006600660065, | |
| "grad_norm": 1.2844809293746948, | |
| "learning_rate": 3.6960784313725496e-05, | |
| "loss": 0.1413, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 2.011881188118812, | |
| "grad_norm": 1.3603469133377075, | |
| "learning_rate": 3.6911764705882356e-05, | |
| "loss": 0.1292, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 2.0171617161716173, | |
| "grad_norm": 1.5470837354660034, | |
| "learning_rate": 3.686274509803922e-05, | |
| "loss": 0.1572, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 2.0224422442244223, | |
| "grad_norm": 1.8591777086257935, | |
| "learning_rate": 3.6813725490196084e-05, | |
| "loss": 0.1545, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 2.0277227722772277, | |
| "grad_norm": 1.2977832555770874, | |
| "learning_rate": 3.6764705882352945e-05, | |
| "loss": 0.0742, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 2.033003300330033, | |
| "grad_norm": 1.9429755210876465, | |
| "learning_rate": 3.6715686274509805e-05, | |
| "loss": 0.1088, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 2.038283828382838, | |
| "grad_norm": 2.0210816860198975, | |
| "learning_rate": 3.6666666666666666e-05, | |
| "loss": 0.1492, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 2.0435643564356436, | |
| "grad_norm": 1.8192780017852783, | |
| "learning_rate": 3.6617647058823526e-05, | |
| "loss": 0.1585, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 2.048844884488449, | |
| "grad_norm": 1.5634256601333618, | |
| "learning_rate": 3.6568627450980393e-05, | |
| "loss": 0.1152, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 2.054125412541254, | |
| "grad_norm": 1.7004332542419434, | |
| "learning_rate": 3.6519607843137254e-05, | |
| "loss": 0.1629, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 2.0594059405940595, | |
| "grad_norm": 1.7340906858444214, | |
| "learning_rate": 3.6470588235294114e-05, | |
| "loss": 0.1523, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 2.064686468646865, | |
| "grad_norm": 1.4663294553756714, | |
| "learning_rate": 3.642156862745098e-05, | |
| "loss": 0.1138, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 2.06996699669967, | |
| "grad_norm": 1.6033658981323242, | |
| "learning_rate": 3.637254901960784e-05, | |
| "loss": 0.1485, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 2.0752475247524753, | |
| "grad_norm": 1.3706963062286377, | |
| "learning_rate": 3.632352941176471e-05, | |
| "loss": 0.0883, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 2.0805280528052803, | |
| "grad_norm": 1.0890157222747803, | |
| "learning_rate": 3.627450980392157e-05, | |
| "loss": 0.0526, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 2.0858085808580857, | |
| "grad_norm": 1.5241326093673706, | |
| "learning_rate": 3.622549019607843e-05, | |
| "loss": 0.1417, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 2.091089108910891, | |
| "grad_norm": 1.389540433883667, | |
| "learning_rate": 3.61764705882353e-05, | |
| "loss": 0.0972, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 2.096369636963696, | |
| "grad_norm": 1.8035510778427124, | |
| "learning_rate": 3.612745098039216e-05, | |
| "loss": 0.1539, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 2.1016501650165016, | |
| "grad_norm": 1.5949468612670898, | |
| "learning_rate": 3.607843137254902e-05, | |
| "loss": 0.1086, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 2.106930693069307, | |
| "grad_norm": 1.5921865701675415, | |
| "learning_rate": 3.6029411764705886e-05, | |
| "loss": 0.1227, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 2.112211221122112, | |
| "grad_norm": 1.6403027772903442, | |
| "learning_rate": 3.5980392156862746e-05, | |
| "loss": 0.1315, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 2.1174917491749174, | |
| "grad_norm": 1.7026506662368774, | |
| "learning_rate": 3.593137254901961e-05, | |
| "loss": 0.1178, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 2.122772277227723, | |
| "grad_norm": 1.5574462413787842, | |
| "learning_rate": 3.5882352941176474e-05, | |
| "loss": 0.1316, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 2.128052805280528, | |
| "grad_norm": 1.6316189765930176, | |
| "learning_rate": 3.5833333333333335e-05, | |
| "loss": 0.1351, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 2.1333333333333333, | |
| "grad_norm": 1.2187312841415405, | |
| "learning_rate": 3.5784313725490195e-05, | |
| "loss": 0.0708, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 2.1386138613861387, | |
| "grad_norm": 1.619545817375183, | |
| "learning_rate": 3.573529411764706e-05, | |
| "loss": 0.1369, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 2.1438943894389437, | |
| "grad_norm": 1.4654717445373535, | |
| "learning_rate": 3.568627450980392e-05, | |
| "loss": 0.1499, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 2.149174917491749, | |
| "grad_norm": 1.393074870109558, | |
| "learning_rate": 3.563725490196079e-05, | |
| "loss": 0.1277, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 2.1544554455445546, | |
| "grad_norm": 1.7896983623504639, | |
| "learning_rate": 3.558823529411765e-05, | |
| "loss": 0.166, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 2.1597359735973596, | |
| "grad_norm": 1.5691279172897339, | |
| "learning_rate": 3.553921568627451e-05, | |
| "loss": 0.0936, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 2.165016501650165, | |
| "grad_norm": 1.6163969039916992, | |
| "learning_rate": 3.549019607843138e-05, | |
| "loss": 0.1234, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 2.1702970297029704, | |
| "grad_norm": 1.7184455394744873, | |
| "learning_rate": 3.544117647058824e-05, | |
| "loss": 0.1287, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 2.1755775577557754, | |
| "grad_norm": 1.8023498058319092, | |
| "learning_rate": 3.53921568627451e-05, | |
| "loss": 0.1433, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 2.180858085808581, | |
| "grad_norm": 1.6063473224639893, | |
| "learning_rate": 3.534313725490197e-05, | |
| "loss": 0.1158, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 2.1861386138613863, | |
| "grad_norm": 1.6521129608154297, | |
| "learning_rate": 3.529411764705883e-05, | |
| "loss": 0.115, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 2.1914191419141913, | |
| "grad_norm": 1.5446815490722656, | |
| "learning_rate": 3.524509803921569e-05, | |
| "loss": 0.115, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 2.1966996699669967, | |
| "grad_norm": 1.4110487699508667, | |
| "learning_rate": 3.5196078431372555e-05, | |
| "loss": 0.132, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 2.201980198019802, | |
| "grad_norm": 1.732271671295166, | |
| "learning_rate": 3.514705882352941e-05, | |
| "loss": 0.116, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 2.207260726072607, | |
| "grad_norm": 1.8033490180969238, | |
| "learning_rate": 3.5098039215686276e-05, | |
| "loss": 0.1675, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 2.2125412541254126, | |
| "grad_norm": 1.6612602472305298, | |
| "learning_rate": 3.5049019607843136e-05, | |
| "loss": 0.1549, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 2.217821782178218, | |
| "grad_norm": 1.4840703010559082, | |
| "learning_rate": 3.5e-05, | |
| "loss": 0.116, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 2.223102310231023, | |
| "grad_norm": 1.5208748579025269, | |
| "learning_rate": 3.4950980392156864e-05, | |
| "loss": 0.1286, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 2.2283828382838284, | |
| "grad_norm": 1.6537951231002808, | |
| "learning_rate": 3.4901960784313725e-05, | |
| "loss": 0.0995, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 2.233663366336634, | |
| "grad_norm": 1.8060580492019653, | |
| "learning_rate": 3.4852941176470585e-05, | |
| "loss": 0.1264, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 2.238943894389439, | |
| "grad_norm": 1.8712666034698486, | |
| "learning_rate": 3.480392156862745e-05, | |
| "loss": 0.1722, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 2.2442244224422443, | |
| "grad_norm": 1.5156561136245728, | |
| "learning_rate": 3.475490196078431e-05, | |
| "loss": 0.117, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 2.2495049504950497, | |
| "grad_norm": 1.6250231266021729, | |
| "learning_rate": 3.470588235294118e-05, | |
| "loss": 0.1299, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 2.2547854785478547, | |
| "grad_norm": 2.060224771499634, | |
| "learning_rate": 3.465686274509804e-05, | |
| "loss": 0.1552, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 2.26006600660066, | |
| "grad_norm": 1.6682344675064087, | |
| "learning_rate": 3.46078431372549e-05, | |
| "loss": 0.137, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 2.2653465346534656, | |
| "grad_norm": 1.5238713026046753, | |
| "learning_rate": 3.455882352941177e-05, | |
| "loss": 0.0956, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 2.2706270627062706, | |
| "grad_norm": 1.7260777950286865, | |
| "learning_rate": 3.450980392156863e-05, | |
| "loss": 0.127, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 2.275907590759076, | |
| "grad_norm": 1.5617260932922363, | |
| "learning_rate": 3.446078431372549e-05, | |
| "loss": 0.1495, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 2.2811881188118814, | |
| "grad_norm": 1.6972553730010986, | |
| "learning_rate": 3.441176470588236e-05, | |
| "loss": 0.1521, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 2.2864686468646864, | |
| "grad_norm": 1.4397108554840088, | |
| "learning_rate": 3.436274509803922e-05, | |
| "loss": 0.1218, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 2.291749174917492, | |
| "grad_norm": 1.099647045135498, | |
| "learning_rate": 3.431372549019608e-05, | |
| "loss": 0.0629, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 2.297029702970297, | |
| "grad_norm": 1.1678043603897095, | |
| "learning_rate": 3.4264705882352945e-05, | |
| "loss": 0.0905, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 2.3023102310231023, | |
| "grad_norm": 1.3565727472305298, | |
| "learning_rate": 3.4215686274509805e-05, | |
| "loss": 0.1028, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 2.3075907590759077, | |
| "grad_norm": 1.846362590789795, | |
| "learning_rate": 3.4166666666666666e-05, | |
| "loss": 0.1715, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 2.3128712871287127, | |
| "grad_norm": 1.637338399887085, | |
| "learning_rate": 3.411764705882353e-05, | |
| "loss": 0.131, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 2.318151815181518, | |
| "grad_norm": 1.5920330286026, | |
| "learning_rate": 3.4068627450980394e-05, | |
| "loss": 0.1279, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 2.3234323432343236, | |
| "grad_norm": 1.6633886098861694, | |
| "learning_rate": 3.401960784313726e-05, | |
| "loss": 0.1284, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 2.3287128712871286, | |
| "grad_norm": 1.677240252494812, | |
| "learning_rate": 3.397058823529412e-05, | |
| "loss": 0.132, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 2.333993399339934, | |
| "grad_norm": 1.6058099269866943, | |
| "learning_rate": 3.392156862745098e-05, | |
| "loss": 0.116, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 2.3392739273927394, | |
| "grad_norm": 1.586955189704895, | |
| "learning_rate": 3.387254901960785e-05, | |
| "loss": 0.1201, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 2.3445544554455444, | |
| "grad_norm": 2.02844500541687, | |
| "learning_rate": 3.382352941176471e-05, | |
| "loss": 0.1732, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 2.34983498349835, | |
| "grad_norm": 1.858375072479248, | |
| "learning_rate": 3.377450980392157e-05, | |
| "loss": 0.1591, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 2.3551155115511553, | |
| "grad_norm": 1.9212247133255005, | |
| "learning_rate": 3.372549019607844e-05, | |
| "loss": 0.2203, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 2.3603960396039603, | |
| "grad_norm": 1.327735424041748, | |
| "learning_rate": 3.36764705882353e-05, | |
| "loss": 0.0983, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 2.3656765676567657, | |
| "grad_norm": 1.6712234020233154, | |
| "learning_rate": 3.362745098039216e-05, | |
| "loss": 0.1282, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 2.370957095709571, | |
| "grad_norm": 1.812563180923462, | |
| "learning_rate": 3.357843137254902e-05, | |
| "loss": 0.1417, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 2.376237623762376, | |
| "grad_norm": 1.4844976663589478, | |
| "learning_rate": 3.352941176470588e-05, | |
| "loss": 0.0892, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 2.3815181518151816, | |
| "grad_norm": 1.7959266901016235, | |
| "learning_rate": 3.348039215686275e-05, | |
| "loss": 0.1753, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 2.3867986798679866, | |
| "grad_norm": 1.4532088041305542, | |
| "learning_rate": 3.343137254901961e-05, | |
| "loss": 0.1154, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 2.392079207920792, | |
| "grad_norm": 1.618485927581787, | |
| "learning_rate": 3.338235294117647e-05, | |
| "loss": 0.1268, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 2.3973597359735974, | |
| "grad_norm": 1.537070870399475, | |
| "learning_rate": 3.3333333333333335e-05, | |
| "loss": 0.1304, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 2.4026402640264024, | |
| "grad_norm": 1.4429882764816284, | |
| "learning_rate": 3.3284313725490195e-05, | |
| "loss": 0.1521, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 2.407920792079208, | |
| "grad_norm": 1.5398533344268799, | |
| "learning_rate": 3.3235294117647056e-05, | |
| "loss": 0.1237, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 2.4132013201320133, | |
| "grad_norm": 1.7345346212387085, | |
| "learning_rate": 3.318627450980392e-05, | |
| "loss": 0.1765, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 2.4184818481848183, | |
| "grad_norm": 1.6664462089538574, | |
| "learning_rate": 3.3137254901960784e-05, | |
| "loss": 0.1363, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 2.4237623762376237, | |
| "grad_norm": 1.2778750658035278, | |
| "learning_rate": 3.308823529411765e-05, | |
| "loss": 0.0798, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 2.429042904290429, | |
| "grad_norm": 1.5715030431747437, | |
| "learning_rate": 3.303921568627451e-05, | |
| "loss": 0.1376, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 2.434323432343234, | |
| "grad_norm": 1.3661701679229736, | |
| "learning_rate": 3.299019607843137e-05, | |
| "loss": 0.1179, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 2.4396039603960396, | |
| "grad_norm": 1.6188709735870361, | |
| "learning_rate": 3.294117647058824e-05, | |
| "loss": 0.102, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 2.444884488448845, | |
| "grad_norm": 1.5044273138046265, | |
| "learning_rate": 3.28921568627451e-05, | |
| "loss": 0.1048, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 2.45016501650165, | |
| "grad_norm": 1.8329862356185913, | |
| "learning_rate": 3.284313725490196e-05, | |
| "loss": 0.1501, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 2.4554455445544554, | |
| "grad_norm": 1.4117523431777954, | |
| "learning_rate": 3.279411764705883e-05, | |
| "loss": 0.0899, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 2.460726072607261, | |
| "grad_norm": 1.4140467643737793, | |
| "learning_rate": 3.274509803921569e-05, | |
| "loss": 0.1143, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 2.466006600660066, | |
| "grad_norm": 1.5512601137161255, | |
| "learning_rate": 3.269607843137255e-05, | |
| "loss": 0.1294, | |
| "step": 467 | |
| }, | |
| { | |
| "epoch": 2.4712871287128713, | |
| "grad_norm": 1.366589069366455, | |
| "learning_rate": 3.2647058823529416e-05, | |
| "loss": 0.1044, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 2.4765676567656767, | |
| "grad_norm": 1.4923369884490967, | |
| "learning_rate": 3.2598039215686276e-05, | |
| "loss": 0.1209, | |
| "step": 469 | |
| }, | |
| { | |
| "epoch": 2.4818481848184817, | |
| "grad_norm": 1.4540033340454102, | |
| "learning_rate": 3.254901960784314e-05, | |
| "loss": 0.1094, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 2.487128712871287, | |
| "grad_norm": 1.7414342164993286, | |
| "learning_rate": 3.2500000000000004e-05, | |
| "loss": 0.2082, | |
| "step": 471 | |
| }, | |
| { | |
| "epoch": 2.4924092409240926, | |
| "grad_norm": 1.3162891864776611, | |
| "learning_rate": 3.2450980392156864e-05, | |
| "loss": 0.0972, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 2.4976897689768975, | |
| "grad_norm": 1.5640320777893066, | |
| "learning_rate": 3.240196078431373e-05, | |
| "loss": 0.1221, | |
| "step": 473 | |
| }, | |
| { | |
| "epoch": 2.502970297029703, | |
| "grad_norm": 1.3759536743164062, | |
| "learning_rate": 3.235294117647059e-05, | |
| "loss": 0.1026, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 2.5082508250825084, | |
| "grad_norm": 1.3398675918579102, | |
| "learning_rate": 3.230392156862745e-05, | |
| "loss": 0.0904, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 2.5135313531353134, | |
| "grad_norm": 1.1896995306015015, | |
| "learning_rate": 3.225490196078432e-05, | |
| "loss": 0.0847, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 2.518811881188119, | |
| "grad_norm": 1.3623279333114624, | |
| "learning_rate": 3.220588235294118e-05, | |
| "loss": 0.1159, | |
| "step": 477 | |
| }, | |
| { | |
| "epoch": 2.5240924092409243, | |
| "grad_norm": 1.3050968647003174, | |
| "learning_rate": 3.215686274509804e-05, | |
| "loss": 0.0967, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 2.5293729372937293, | |
| "grad_norm": 1.8732277154922485, | |
| "learning_rate": 3.210784313725491e-05, | |
| "loss": 0.1273, | |
| "step": 479 | |
| }, | |
| { | |
| "epoch": 2.5346534653465347, | |
| "grad_norm": 1.5116699934005737, | |
| "learning_rate": 3.205882352941177e-05, | |
| "loss": 0.1245, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 2.53993399339934, | |
| "grad_norm": 1.6075270175933838, | |
| "learning_rate": 3.200980392156863e-05, | |
| "loss": 0.1215, | |
| "step": 481 | |
| }, | |
| { | |
| "epoch": 2.545214521452145, | |
| "grad_norm": 1.6938683986663818, | |
| "learning_rate": 3.196078431372549e-05, | |
| "loss": 0.1226, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 2.5504950495049505, | |
| "grad_norm": 1.361075520515442, | |
| "learning_rate": 3.191176470588235e-05, | |
| "loss": 0.108, | |
| "step": 483 | |
| }, | |
| { | |
| "epoch": 2.555775577557756, | |
| "grad_norm": 1.388642430305481, | |
| "learning_rate": 3.186274509803922e-05, | |
| "loss": 0.1223, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 2.561056105610561, | |
| "grad_norm": 1.637742519378662, | |
| "learning_rate": 3.181372549019608e-05, | |
| "loss": 0.1395, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 2.5663366336633664, | |
| "grad_norm": 1.377528429031372, | |
| "learning_rate": 3.176470588235294e-05, | |
| "loss": 0.1064, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 2.571617161716172, | |
| "grad_norm": 1.3829468488693237, | |
| "learning_rate": 3.1715686274509806e-05, | |
| "loss": 0.1101, | |
| "step": 487 | |
| }, | |
| { | |
| "epoch": 2.576897689768977, | |
| "grad_norm": 1.142531394958496, | |
| "learning_rate": 3.1666666666666666e-05, | |
| "loss": 0.0885, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 2.5821782178217823, | |
| "grad_norm": 1.226916790008545, | |
| "learning_rate": 3.161764705882353e-05, | |
| "loss": 0.0713, | |
| "step": 489 | |
| }, | |
| { | |
| "epoch": 2.5874587458745877, | |
| "grad_norm": 1.3948824405670166, | |
| "learning_rate": 3.1568627450980394e-05, | |
| "loss": 0.0945, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 2.5927392739273927, | |
| "grad_norm": 1.679543375968933, | |
| "learning_rate": 3.1519607843137254e-05, | |
| "loss": 0.1122, | |
| "step": 491 | |
| }, | |
| { | |
| "epoch": 2.598019801980198, | |
| "grad_norm": 1.4996442794799805, | |
| "learning_rate": 3.147058823529412e-05, | |
| "loss": 0.1502, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 2.6033003300330035, | |
| "grad_norm": 1.6479462385177612, | |
| "learning_rate": 3.142156862745098e-05, | |
| "loss": 0.1335, | |
| "step": 493 | |
| }, | |
| { | |
| "epoch": 2.6085808580858085, | |
| "grad_norm": 1.9503371715545654, | |
| "learning_rate": 3.137254901960784e-05, | |
| "loss": 0.1378, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 2.613861386138614, | |
| "grad_norm": 1.6987338066101074, | |
| "learning_rate": 3.132352941176471e-05, | |
| "loss": 0.174, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 2.6191419141914194, | |
| "grad_norm": 1.4869428873062134, | |
| "learning_rate": 3.127450980392157e-05, | |
| "loss": 0.1444, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 2.6244224422442244, | |
| "grad_norm": 1.4220130443572998, | |
| "learning_rate": 3.122549019607843e-05, | |
| "loss": 0.1146, | |
| "step": 497 | |
| }, | |
| { | |
| "epoch": 2.62970297029703, | |
| "grad_norm": 1.8738462924957275, | |
| "learning_rate": 3.11764705882353e-05, | |
| "loss": 0.1999, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 2.6349834983498353, | |
| "grad_norm": 1.6982347965240479, | |
| "learning_rate": 3.112745098039216e-05, | |
| "loss": 0.1241, | |
| "step": 499 | |
| }, | |
| { | |
| "epoch": 2.6402640264026402, | |
| "grad_norm": 1.5183193683624268, | |
| "learning_rate": 3.107843137254902e-05, | |
| "loss": 0.1198, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 2.6455445544554457, | |
| "grad_norm": 1.4872636795043945, | |
| "learning_rate": 3.1029411764705886e-05, | |
| "loss": 0.0927, | |
| "step": 501 | |
| }, | |
| { | |
| "epoch": 2.6508250825082507, | |
| "grad_norm": 1.6051157712936401, | |
| "learning_rate": 3.098039215686275e-05, | |
| "loss": 0.1524, | |
| "step": 502 | |
| }, | |
| { | |
| "epoch": 2.656105610561056, | |
| "grad_norm": 1.5139557123184204, | |
| "learning_rate": 3.093137254901961e-05, | |
| "loss": 0.1237, | |
| "step": 503 | |
| }, | |
| { | |
| "epoch": 2.6613861386138615, | |
| "grad_norm": 1.5860090255737305, | |
| "learning_rate": 3.0882352941176475e-05, | |
| "loss": 0.1409, | |
| "step": 504 | |
| }, | |
| { | |
| "epoch": 2.6666666666666665, | |
| "grad_norm": 1.7378284931182861, | |
| "learning_rate": 3.0833333333333335e-05, | |
| "loss": 0.1551, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 2.671947194719472, | |
| "grad_norm": 1.5908591747283936, | |
| "learning_rate": 3.07843137254902e-05, | |
| "loss": 0.1169, | |
| "step": 506 | |
| }, | |
| { | |
| "epoch": 2.6772277227722774, | |
| "grad_norm": 1.4210619926452637, | |
| "learning_rate": 3.073529411764706e-05, | |
| "loss": 0.1217, | |
| "step": 507 | |
| }, | |
| { | |
| "epoch": 2.6825082508250824, | |
| "grad_norm": 1.4068129062652588, | |
| "learning_rate": 3.0686274509803923e-05, | |
| "loss": 0.1484, | |
| "step": 508 | |
| }, | |
| { | |
| "epoch": 2.687788778877888, | |
| "grad_norm": 1.4808945655822754, | |
| "learning_rate": 3.063725490196079e-05, | |
| "loss": 0.1402, | |
| "step": 509 | |
| }, | |
| { | |
| "epoch": 2.693069306930693, | |
| "grad_norm": 1.342729926109314, | |
| "learning_rate": 3.058823529411765e-05, | |
| "loss": 0.0826, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 2.6983498349834982, | |
| "grad_norm": 1.457270860671997, | |
| "learning_rate": 3.053921568627451e-05, | |
| "loss": 0.0809, | |
| "step": 511 | |
| }, | |
| { | |
| "epoch": 2.7036303630363037, | |
| "grad_norm": 1.3564640283584595, | |
| "learning_rate": 3.0490196078431376e-05, | |
| "loss": 0.1333, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 2.7089108910891087, | |
| "grad_norm": 1.5419988632202148, | |
| "learning_rate": 3.0441176470588233e-05, | |
| "loss": 0.1145, | |
| "step": 513 | |
| }, | |
| { | |
| "epoch": 2.714191419141914, | |
| "grad_norm": 1.4028230905532837, | |
| "learning_rate": 3.0392156862745097e-05, | |
| "loss": 0.1156, | |
| "step": 514 | |
| }, | |
| { | |
| "epoch": 2.7194719471947195, | |
| "grad_norm": 1.5217773914337158, | |
| "learning_rate": 3.034313725490196e-05, | |
| "loss": 0.0962, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 2.7247524752475245, | |
| "grad_norm": 1.2986435890197754, | |
| "learning_rate": 3.0294117647058824e-05, | |
| "loss": 0.1039, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 2.73003300330033, | |
| "grad_norm": 1.3331955671310425, | |
| "learning_rate": 3.0245098039215685e-05, | |
| "loss": 0.1077, | |
| "step": 517 | |
| }, | |
| { | |
| "epoch": 2.7353135313531354, | |
| "grad_norm": 1.7780873775482178, | |
| "learning_rate": 3.019607843137255e-05, | |
| "loss": 0.2033, | |
| "step": 518 | |
| }, | |
| { | |
| "epoch": 2.7405940594059404, | |
| "grad_norm": 1.1954641342163086, | |
| "learning_rate": 3.0147058823529413e-05, | |
| "loss": 0.1134, | |
| "step": 519 | |
| }, | |
| { | |
| "epoch": 2.745874587458746, | |
| "grad_norm": 1.465927004814148, | |
| "learning_rate": 3.0098039215686273e-05, | |
| "loss": 0.1329, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 2.7511551155115512, | |
| "grad_norm": 1.359671711921692, | |
| "learning_rate": 3.0049019607843137e-05, | |
| "loss": 0.0876, | |
| "step": 521 | |
| }, | |
| { | |
| "epoch": 2.756435643564356, | |
| "grad_norm": 1.570184350013733, | |
| "learning_rate": 3e-05, | |
| "loss": 0.1125, | |
| "step": 522 | |
| }, | |
| { | |
| "epoch": 2.7617161716171617, | |
| "grad_norm": 1.286363959312439, | |
| "learning_rate": 2.9950980392156865e-05, | |
| "loss": 0.0978, | |
| "step": 523 | |
| }, | |
| { | |
| "epoch": 2.766996699669967, | |
| "grad_norm": 1.6705877780914307, | |
| "learning_rate": 2.9901960784313725e-05, | |
| "loss": 0.1546, | |
| "step": 524 | |
| }, | |
| { | |
| "epoch": 2.772277227722772, | |
| "grad_norm": 1.475122332572937, | |
| "learning_rate": 2.985294117647059e-05, | |
| "loss": 0.1083, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 2.7775577557755775, | |
| "grad_norm": 1.4407463073730469, | |
| "learning_rate": 2.9803921568627453e-05, | |
| "loss": 0.1123, | |
| "step": 526 | |
| }, | |
| { | |
| "epoch": 2.782838283828383, | |
| "grad_norm": 1.4154635667800903, | |
| "learning_rate": 2.9754901960784313e-05, | |
| "loss": 0.1232, | |
| "step": 527 | |
| }, | |
| { | |
| "epoch": 2.788118811881188, | |
| "grad_norm": 1.6671222448349, | |
| "learning_rate": 2.9705882352941177e-05, | |
| "loss": 0.1767, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 2.7933993399339934, | |
| "grad_norm": 1.4719637632369995, | |
| "learning_rate": 2.965686274509804e-05, | |
| "loss": 0.1325, | |
| "step": 529 | |
| }, | |
| { | |
| "epoch": 2.798679867986799, | |
| "grad_norm": 0.979076087474823, | |
| "learning_rate": 2.9607843137254905e-05, | |
| "loss": 0.0692, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 2.803960396039604, | |
| "grad_norm": 1.507076621055603, | |
| "learning_rate": 2.9558823529411766e-05, | |
| "loss": 0.1275, | |
| "step": 531 | |
| }, | |
| { | |
| "epoch": 2.809240924092409, | |
| "grad_norm": 1.6182982921600342, | |
| "learning_rate": 2.950980392156863e-05, | |
| "loss": 0.1528, | |
| "step": 532 | |
| }, | |
| { | |
| "epoch": 2.8145214521452147, | |
| "grad_norm": 1.5272071361541748, | |
| "learning_rate": 2.9460784313725493e-05, | |
| "loss": 0.1344, | |
| "step": 533 | |
| }, | |
| { | |
| "epoch": 2.8198019801980196, | |
| "grad_norm": 1.3018370866775513, | |
| "learning_rate": 2.9411764705882354e-05, | |
| "loss": 0.101, | |
| "step": 534 | |
| }, | |
| { | |
| "epoch": 2.825082508250825, | |
| "grad_norm": 1.2394366264343262, | |
| "learning_rate": 2.9362745098039218e-05, | |
| "loss": 0.0893, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 2.8303630363036305, | |
| "grad_norm": 1.5351279973983765, | |
| "learning_rate": 2.931372549019608e-05, | |
| "loss": 0.1226, | |
| "step": 536 | |
| }, | |
| { | |
| "epoch": 2.8356435643564355, | |
| "grad_norm": 1.7062324285507202, | |
| "learning_rate": 2.9264705882352945e-05, | |
| "loss": 0.1363, | |
| "step": 537 | |
| }, | |
| { | |
| "epoch": 2.840924092409241, | |
| "grad_norm": 1.666043996810913, | |
| "learning_rate": 2.9215686274509806e-05, | |
| "loss": 0.137, | |
| "step": 538 | |
| }, | |
| { | |
| "epoch": 2.8462046204620464, | |
| "grad_norm": 1.3965767621994019, | |
| "learning_rate": 2.916666666666667e-05, | |
| "loss": 0.1018, | |
| "step": 539 | |
| }, | |
| { | |
| "epoch": 2.8514851485148514, | |
| "grad_norm": 1.4355634450912476, | |
| "learning_rate": 2.9117647058823534e-05, | |
| "loss": 0.1009, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 2.856765676567657, | |
| "grad_norm": 1.369983434677124, | |
| "learning_rate": 2.9068627450980394e-05, | |
| "loss": 0.0964, | |
| "step": 541 | |
| }, | |
| { | |
| "epoch": 2.862046204620462, | |
| "grad_norm": 1.5841052532196045, | |
| "learning_rate": 2.9019607843137258e-05, | |
| "loss": 0.148, | |
| "step": 542 | |
| }, | |
| { | |
| "epoch": 2.867326732673267, | |
| "grad_norm": 1.360392451286316, | |
| "learning_rate": 2.8970588235294122e-05, | |
| "loss": 0.0744, | |
| "step": 543 | |
| }, | |
| { | |
| "epoch": 2.8726072607260726, | |
| "grad_norm": 1.379198431968689, | |
| "learning_rate": 2.8921568627450986e-05, | |
| "loss": 0.1047, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 2.877887788778878, | |
| "grad_norm": 1.3699851036071777, | |
| "learning_rate": 2.8872549019607843e-05, | |
| "loss": 0.0813, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 2.883168316831683, | |
| "grad_norm": 1.162501335144043, | |
| "learning_rate": 2.8823529411764703e-05, | |
| "loss": 0.0848, | |
| "step": 546 | |
| }, | |
| { | |
| "epoch": 2.8884488448844885, | |
| "grad_norm": 1.6580568552017212, | |
| "learning_rate": 2.8774509803921567e-05, | |
| "loss": 0.1252, | |
| "step": 547 | |
| }, | |
| { | |
| "epoch": 2.893729372937294, | |
| "grad_norm": 1.8794211149215698, | |
| "learning_rate": 2.872549019607843e-05, | |
| "loss": 0.1581, | |
| "step": 548 | |
| }, | |
| { | |
| "epoch": 2.899009900990099, | |
| "grad_norm": 1.7686625719070435, | |
| "learning_rate": 2.8676470588235295e-05, | |
| "loss": 0.172, | |
| "step": 549 | |
| }, | |
| { | |
| "epoch": 2.9042904290429044, | |
| "grad_norm": 1.371221899986267, | |
| "learning_rate": 2.8627450980392155e-05, | |
| "loss": 0.0977, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 2.90957095709571, | |
| "grad_norm": 1.4571088552474976, | |
| "learning_rate": 2.857843137254902e-05, | |
| "loss": 0.1251, | |
| "step": 551 | |
| }, | |
| { | |
| "epoch": 2.9148514851485148, | |
| "grad_norm": 1.1055774688720703, | |
| "learning_rate": 2.8529411764705883e-05, | |
| "loss": 0.0761, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 2.92013201320132, | |
| "grad_norm": 1.4305754899978638, | |
| "learning_rate": 2.8480392156862744e-05, | |
| "loss": 0.143, | |
| "step": 553 | |
| }, | |
| { | |
| "epoch": 2.9254125412541256, | |
| "grad_norm": 1.3293156623840332, | |
| "learning_rate": 2.8431372549019608e-05, | |
| "loss": 0.1218, | |
| "step": 554 | |
| }, | |
| { | |
| "epoch": 2.9306930693069306, | |
| "grad_norm": 1.7073545455932617, | |
| "learning_rate": 2.838235294117647e-05, | |
| "loss": 0.1495, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 2.935973597359736, | |
| "grad_norm": 1.0066826343536377, | |
| "learning_rate": 2.8333333333333335e-05, | |
| "loss": 0.0698, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 2.9412541254125415, | |
| "grad_norm": 1.5967711210250854, | |
| "learning_rate": 2.8284313725490196e-05, | |
| "loss": 0.1436, | |
| "step": 557 | |
| }, | |
| { | |
| "epoch": 2.9465346534653465, | |
| "grad_norm": 1.7498648166656494, | |
| "learning_rate": 2.823529411764706e-05, | |
| "loss": 0.1793, | |
| "step": 558 | |
| }, | |
| { | |
| "epoch": 2.951815181518152, | |
| "grad_norm": 1.4582027196884155, | |
| "learning_rate": 2.8186274509803924e-05, | |
| "loss": 0.1152, | |
| "step": 559 | |
| }, | |
| { | |
| "epoch": 2.9570957095709574, | |
| "grad_norm": 1.1277716159820557, | |
| "learning_rate": 2.8137254901960784e-05, | |
| "loss": 0.0807, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 2.9623762376237623, | |
| "grad_norm": 1.4396110773086548, | |
| "learning_rate": 2.8088235294117648e-05, | |
| "loss": 0.1141, | |
| "step": 561 | |
| }, | |
| { | |
| "epoch": 2.9676567656765678, | |
| "grad_norm": 1.3722310066223145, | |
| "learning_rate": 2.8039215686274512e-05, | |
| "loss": 0.0952, | |
| "step": 562 | |
| }, | |
| { | |
| "epoch": 2.972937293729373, | |
| "grad_norm": 1.6217355728149414, | |
| "learning_rate": 2.7990196078431376e-05, | |
| "loss": 0.1329, | |
| "step": 563 | |
| }, | |
| { | |
| "epoch": 2.978217821782178, | |
| "grad_norm": 1.3327069282531738, | |
| "learning_rate": 2.7941176470588236e-05, | |
| "loss": 0.109, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 2.9834983498349836, | |
| "grad_norm": 1.6068836450576782, | |
| "learning_rate": 2.78921568627451e-05, | |
| "loss": 0.1136, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 2.9887788778877886, | |
| "grad_norm": 1.6134989261627197, | |
| "learning_rate": 2.7843137254901964e-05, | |
| "loss": 0.1452, | |
| "step": 566 | |
| }, | |
| { | |
| "epoch": 2.994059405940594, | |
| "grad_norm": 1.301640272140503, | |
| "learning_rate": 2.7794117647058824e-05, | |
| "loss": 0.1692, | |
| "step": 567 | |
| }, | |
| { | |
| "epoch": 2.9993399339933995, | |
| "grad_norm": 1.5887155532836914, | |
| "learning_rate": 2.774509803921569e-05, | |
| "loss": 0.1529, | |
| "step": 568 | |
| }, | |
| { | |
| "epoch": 3.0046204620462045, | |
| "grad_norm": 1.1964720487594604, | |
| "learning_rate": 2.7696078431372552e-05, | |
| "loss": 0.05, | |
| "step": 569 | |
| }, | |
| { | |
| "epoch": 3.00990099009901, | |
| "grad_norm": 1.26799476146698, | |
| "learning_rate": 2.7647058823529416e-05, | |
| "loss": 0.0582, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 3.0151815181518153, | |
| "grad_norm": 1.0527023077011108, | |
| "learning_rate": 2.7598039215686277e-05, | |
| "loss": 0.0661, | |
| "step": 571 | |
| }, | |
| { | |
| "epoch": 3.0204620462046203, | |
| "grad_norm": 1.0215531587600708, | |
| "learning_rate": 2.754901960784314e-05, | |
| "loss": 0.0487, | |
| "step": 572 | |
| }, | |
| { | |
| "epoch": 3.0257425742574258, | |
| "grad_norm": 0.9712955951690674, | |
| "learning_rate": 2.7500000000000004e-05, | |
| "loss": 0.0533, | |
| "step": 573 | |
| }, | |
| { | |
| "epoch": 3.031023102310231, | |
| "grad_norm": 0.7851030826568604, | |
| "learning_rate": 2.7450980392156865e-05, | |
| "loss": 0.0253, | |
| "step": 574 | |
| }, | |
| { | |
| "epoch": 3.036303630363036, | |
| "grad_norm": 1.6078637838363647, | |
| "learning_rate": 2.740196078431373e-05, | |
| "loss": 0.0841, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 3.0415841584158416, | |
| "grad_norm": 1.459649920463562, | |
| "learning_rate": 2.7352941176470593e-05, | |
| "loss": 0.0654, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 3.046864686468647, | |
| "grad_norm": 1.6587159633636475, | |
| "learning_rate": 2.730392156862745e-05, | |
| "loss": 0.0524, | |
| "step": 577 | |
| }, | |
| { | |
| "epoch": 3.052145214521452, | |
| "grad_norm": 1.8531984090805054, | |
| "learning_rate": 2.7254901960784314e-05, | |
| "loss": 0.0682, | |
| "step": 578 | |
| }, | |
| { | |
| "epoch": 3.0574257425742575, | |
| "grad_norm": 2.0687427520751953, | |
| "learning_rate": 2.7205882352941174e-05, | |
| "loss": 0.0809, | |
| "step": 579 | |
| }, | |
| { | |
| "epoch": 3.062706270627063, | |
| "grad_norm": 1.7597553730010986, | |
| "learning_rate": 2.7156862745098038e-05, | |
| "loss": 0.0906, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 3.067986798679868, | |
| "grad_norm": 1.593573808670044, | |
| "learning_rate": 2.7107843137254902e-05, | |
| "loss": 0.0798, | |
| "step": 581 | |
| }, | |
| { | |
| "epoch": 3.0732673267326733, | |
| "grad_norm": 1.6800479888916016, | |
| "learning_rate": 2.7058823529411766e-05, | |
| "loss": 0.0769, | |
| "step": 582 | |
| }, | |
| { | |
| "epoch": 3.0785478547854783, | |
| "grad_norm": 1.2938110828399658, | |
| "learning_rate": 2.7009803921568626e-05, | |
| "loss": 0.0588, | |
| "step": 583 | |
| }, | |
| { | |
| "epoch": 3.0838283828382838, | |
| "grad_norm": 0.9052571654319763, | |
| "learning_rate": 2.696078431372549e-05, | |
| "loss": 0.0343, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 3.089108910891089, | |
| "grad_norm": 1.4364440441131592, | |
| "learning_rate": 2.6911764705882354e-05, | |
| "loss": 0.0659, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 3.094389438943894, | |
| "grad_norm": 1.2938659191131592, | |
| "learning_rate": 2.6862745098039214e-05, | |
| "loss": 0.0653, | |
| "step": 586 | |
| }, | |
| { | |
| "epoch": 3.0996699669966996, | |
| "grad_norm": 1.2038064002990723, | |
| "learning_rate": 2.681372549019608e-05, | |
| "loss": 0.0731, | |
| "step": 587 | |
| }, | |
| { | |
| "epoch": 3.104950495049505, | |
| "grad_norm": 1.1313140392303467, | |
| "learning_rate": 2.6764705882352942e-05, | |
| "loss": 0.0488, | |
| "step": 588 | |
| }, | |
| { | |
| "epoch": 3.11023102310231, | |
| "grad_norm": 1.426941990852356, | |
| "learning_rate": 2.6715686274509806e-05, | |
| "loss": 0.0724, | |
| "step": 589 | |
| }, | |
| { | |
| "epoch": 3.1155115511551155, | |
| "grad_norm": 1.415529489517212, | |
| "learning_rate": 2.6666666666666667e-05, | |
| "loss": 0.0902, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 3.120792079207921, | |
| "grad_norm": 1.4976341724395752, | |
| "learning_rate": 2.661764705882353e-05, | |
| "loss": 0.0837, | |
| "step": 591 | |
| }, | |
| { | |
| "epoch": 3.126072607260726, | |
| "grad_norm": 1.3381704092025757, | |
| "learning_rate": 2.6568627450980394e-05, | |
| "loss": 0.0735, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 3.1313531353135313, | |
| "grad_norm": 1.0349838733673096, | |
| "learning_rate": 2.6519607843137255e-05, | |
| "loss": 0.0482, | |
| "step": 593 | |
| }, | |
| { | |
| "epoch": 3.1366336633663368, | |
| "grad_norm": 1.4465690851211548, | |
| "learning_rate": 2.647058823529412e-05, | |
| "loss": 0.0823, | |
| "step": 594 | |
| }, | |
| { | |
| "epoch": 3.1419141914191417, | |
| "grad_norm": 1.113172173500061, | |
| "learning_rate": 2.6421568627450983e-05, | |
| "loss": 0.0608, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 3.147194719471947, | |
| "grad_norm": 1.3824833631515503, | |
| "learning_rate": 2.6372549019607846e-05, | |
| "loss": 0.0566, | |
| "step": 596 | |
| }, | |
| { | |
| "epoch": 3.1524752475247526, | |
| "grad_norm": 1.2780243158340454, | |
| "learning_rate": 2.6323529411764707e-05, | |
| "loss": 0.0566, | |
| "step": 597 | |
| }, | |
| { | |
| "epoch": 3.1577557755775576, | |
| "grad_norm": 1.7640819549560547, | |
| "learning_rate": 2.627450980392157e-05, | |
| "loss": 0.0627, | |
| "step": 598 | |
| }, | |
| { | |
| "epoch": 3.163036303630363, | |
| "grad_norm": 1.8267886638641357, | |
| "learning_rate": 2.6225490196078435e-05, | |
| "loss": 0.1135, | |
| "step": 599 | |
| }, | |
| { | |
| "epoch": 3.1683168316831685, | |
| "grad_norm": 1.4951374530792236, | |
| "learning_rate": 2.6176470588235295e-05, | |
| "loss": 0.0824, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 3.1735973597359735, | |
| "grad_norm": 1.1276224851608276, | |
| "learning_rate": 2.612745098039216e-05, | |
| "loss": 0.0481, | |
| "step": 601 | |
| }, | |
| { | |
| "epoch": 3.178877887788779, | |
| "grad_norm": 1.3539289236068726, | |
| "learning_rate": 2.6078431372549023e-05, | |
| "loss": 0.0774, | |
| "step": 602 | |
| }, | |
| { | |
| "epoch": 3.1841584158415843, | |
| "grad_norm": 1.1666077375411987, | |
| "learning_rate": 2.6029411764705887e-05, | |
| "loss": 0.0484, | |
| "step": 603 | |
| }, | |
| { | |
| "epoch": 3.1894389438943893, | |
| "grad_norm": 1.4730373620986938, | |
| "learning_rate": 2.5980392156862747e-05, | |
| "loss": 0.0576, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 3.1947194719471947, | |
| "grad_norm": 1.3001021146774292, | |
| "learning_rate": 2.593137254901961e-05, | |
| "loss": 0.0657, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 3.2, | |
| "grad_norm": 1.2780375480651855, | |
| "learning_rate": 2.5882352941176475e-05, | |
| "loss": 0.0538, | |
| "step": 606 | |
| }, | |
| { | |
| "epoch": 3.205280528052805, | |
| "grad_norm": 1.1565543413162231, | |
| "learning_rate": 2.5833333333333336e-05, | |
| "loss": 0.0505, | |
| "step": 607 | |
| }, | |
| { | |
| "epoch": 3.2105610561056106, | |
| "grad_norm": 1.4991666078567505, | |
| "learning_rate": 2.57843137254902e-05, | |
| "loss": 0.058, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 3.215841584158416, | |
| "grad_norm": 1.1679130792617798, | |
| "learning_rate": 2.5735294117647057e-05, | |
| "loss": 0.0525, | |
| "step": 609 | |
| }, | |
| { | |
| "epoch": 3.221122112211221, | |
| "grad_norm": 1.251451849937439, | |
| "learning_rate": 2.568627450980392e-05, | |
| "loss": 0.0525, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 3.2264026402640265, | |
| "grad_norm": 1.4389278888702393, | |
| "learning_rate": 2.5637254901960784e-05, | |
| "loss": 0.0776, | |
| "step": 611 | |
| }, | |
| { | |
| "epoch": 3.231683168316832, | |
| "grad_norm": 1.2829740047454834, | |
| "learning_rate": 2.5588235294117645e-05, | |
| "loss": 0.0529, | |
| "step": 612 | |
| }, | |
| { | |
| "epoch": 3.236963696369637, | |
| "grad_norm": 1.0701279640197754, | |
| "learning_rate": 2.553921568627451e-05, | |
| "loss": 0.0409, | |
| "step": 613 | |
| }, | |
| { | |
| "epoch": 3.2422442244224423, | |
| "grad_norm": 1.363618016242981, | |
| "learning_rate": 2.5490196078431373e-05, | |
| "loss": 0.0797, | |
| "step": 614 | |
| }, | |
| { | |
| "epoch": 3.2475247524752477, | |
| "grad_norm": 1.2213047742843628, | |
| "learning_rate": 2.5441176470588236e-05, | |
| "loss": 0.0621, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 3.2528052805280527, | |
| "grad_norm": 1.4697519540786743, | |
| "learning_rate": 2.5392156862745097e-05, | |
| "loss": 0.0779, | |
| "step": 616 | |
| }, | |
| { | |
| "epoch": 3.258085808580858, | |
| "grad_norm": 1.4408804178237915, | |
| "learning_rate": 2.534313725490196e-05, | |
| "loss": 0.0642, | |
| "step": 617 | |
| }, | |
| { | |
| "epoch": 3.2633663366336636, | |
| "grad_norm": 1.207322120666504, | |
| "learning_rate": 2.5294117647058825e-05, | |
| "loss": 0.0588, | |
| "step": 618 | |
| }, | |
| { | |
| "epoch": 3.2686468646864686, | |
| "grad_norm": 1.399373173713684, | |
| "learning_rate": 2.5245098039215685e-05, | |
| "loss": 0.0789, | |
| "step": 619 | |
| }, | |
| { | |
| "epoch": 3.273927392739274, | |
| "grad_norm": 1.1810945272445679, | |
| "learning_rate": 2.519607843137255e-05, | |
| "loss": 0.0574, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 3.2792079207920795, | |
| "grad_norm": 1.0468858480453491, | |
| "learning_rate": 2.5147058823529413e-05, | |
| "loss": 0.0466, | |
| "step": 621 | |
| }, | |
| { | |
| "epoch": 3.2844884488448844, | |
| "grad_norm": 1.3326268196105957, | |
| "learning_rate": 2.5098039215686277e-05, | |
| "loss": 0.0674, | |
| "step": 622 | |
| }, | |
| { | |
| "epoch": 3.28976897689769, | |
| "grad_norm": 1.3134809732437134, | |
| "learning_rate": 2.5049019607843137e-05, | |
| "loss": 0.0588, | |
| "step": 623 | |
| }, | |
| { | |
| "epoch": 3.295049504950495, | |
| "grad_norm": 1.1663881540298462, | |
| "learning_rate": 2.5e-05, | |
| "loss": 0.0465, | |
| "step": 624 | |
| }, | |
| { | |
| "epoch": 3.3003300330033003, | |
| "grad_norm": 0.9541448950767517, | |
| "learning_rate": 2.4950980392156865e-05, | |
| "loss": 0.0434, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 3.3056105610561057, | |
| "grad_norm": 1.1596111059188843, | |
| "learning_rate": 2.4901960784313726e-05, | |
| "loss": 0.0573, | |
| "step": 626 | |
| }, | |
| { | |
| "epoch": 3.3108910891089107, | |
| "grad_norm": 1.3220415115356445, | |
| "learning_rate": 2.485294117647059e-05, | |
| "loss": 0.058, | |
| "step": 627 | |
| }, | |
| { | |
| "epoch": 3.316171617161716, | |
| "grad_norm": 1.2258095741271973, | |
| "learning_rate": 2.4803921568627453e-05, | |
| "loss": 0.0471, | |
| "step": 628 | |
| }, | |
| { | |
| "epoch": 3.3214521452145216, | |
| "grad_norm": 1.2251251935958862, | |
| "learning_rate": 2.4754901960784317e-05, | |
| "loss": 0.0647, | |
| "step": 629 | |
| }, | |
| { | |
| "epoch": 3.3267326732673266, | |
| "grad_norm": 1.299035906791687, | |
| "learning_rate": 2.4705882352941178e-05, | |
| "loss": 0.056, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 3.332013201320132, | |
| "grad_norm": 1.5529186725616455, | |
| "learning_rate": 2.465686274509804e-05, | |
| "loss": 0.084, | |
| "step": 631 | |
| }, | |
| { | |
| "epoch": 3.3372937293729374, | |
| "grad_norm": 1.0892446041107178, | |
| "learning_rate": 2.4607843137254902e-05, | |
| "loss": 0.0453, | |
| "step": 632 | |
| }, | |
| { | |
| "epoch": 3.3425742574257424, | |
| "grad_norm": 1.1661828756332397, | |
| "learning_rate": 2.4558823529411766e-05, | |
| "loss": 0.0485, | |
| "step": 633 | |
| }, | |
| { | |
| "epoch": 3.347854785478548, | |
| "grad_norm": 1.0938224792480469, | |
| "learning_rate": 2.4509803921568626e-05, | |
| "loss": 0.0469, | |
| "step": 634 | |
| }, | |
| { | |
| "epoch": 3.3531353135313533, | |
| "grad_norm": 1.2503447532653809, | |
| "learning_rate": 2.446078431372549e-05, | |
| "loss": 0.0777, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 3.3584158415841583, | |
| "grad_norm": 1.069814682006836, | |
| "learning_rate": 2.4411764705882354e-05, | |
| "loss": 0.049, | |
| "step": 636 | |
| }, | |
| { | |
| "epoch": 3.3636963696369637, | |
| "grad_norm": 1.321007251739502, | |
| "learning_rate": 2.4362745098039215e-05, | |
| "loss": 0.0596, | |
| "step": 637 | |
| }, | |
| { | |
| "epoch": 3.368976897689769, | |
| "grad_norm": 1.5562186241149902, | |
| "learning_rate": 2.431372549019608e-05, | |
| "loss": 0.0883, | |
| "step": 638 | |
| }, | |
| { | |
| "epoch": 3.374257425742574, | |
| "grad_norm": 1.3883391618728638, | |
| "learning_rate": 2.4264705882352942e-05, | |
| "loss": 0.0777, | |
| "step": 639 | |
| }, | |
| { | |
| "epoch": 3.3795379537953796, | |
| "grad_norm": 1.6028392314910889, | |
| "learning_rate": 2.4215686274509806e-05, | |
| "loss": 0.0769, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 3.384818481848185, | |
| "grad_norm": 1.777230978012085, | |
| "learning_rate": 2.4166666666666667e-05, | |
| "loss": 0.0943, | |
| "step": 641 | |
| }, | |
| { | |
| "epoch": 3.39009900990099, | |
| "grad_norm": 1.0344568490982056, | |
| "learning_rate": 2.411764705882353e-05, | |
| "loss": 0.0452, | |
| "step": 642 | |
| }, | |
| { | |
| "epoch": 3.3953795379537954, | |
| "grad_norm": 1.2247084379196167, | |
| "learning_rate": 2.4068627450980395e-05, | |
| "loss": 0.0516, | |
| "step": 643 | |
| }, | |
| { | |
| "epoch": 3.4006600660066004, | |
| "grad_norm": 1.2037266492843628, | |
| "learning_rate": 2.401960784313726e-05, | |
| "loss": 0.0486, | |
| "step": 644 | |
| }, | |
| { | |
| "epoch": 3.405940594059406, | |
| "grad_norm": 1.3176463842391968, | |
| "learning_rate": 2.397058823529412e-05, | |
| "loss": 0.0522, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 3.4112211221122113, | |
| "grad_norm": 1.6888436079025269, | |
| "learning_rate": 2.3921568627450983e-05, | |
| "loss": 0.1001, | |
| "step": 646 | |
| }, | |
| { | |
| "epoch": 3.4165016501650163, | |
| "grad_norm": 1.296751856803894, | |
| "learning_rate": 2.3872549019607847e-05, | |
| "loss": 0.0656, | |
| "step": 647 | |
| }, | |
| { | |
| "epoch": 3.4217821782178217, | |
| "grad_norm": 1.126638650894165, | |
| "learning_rate": 2.3823529411764707e-05, | |
| "loss": 0.0477, | |
| "step": 648 | |
| }, | |
| { | |
| "epoch": 3.427062706270627, | |
| "grad_norm": 1.394085168838501, | |
| "learning_rate": 2.3774509803921568e-05, | |
| "loss": 0.0761, | |
| "step": 649 | |
| }, | |
| { | |
| "epoch": 3.432343234323432, | |
| "grad_norm": 1.4828298091888428, | |
| "learning_rate": 2.372549019607843e-05, | |
| "loss": 0.0683, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 3.4376237623762376, | |
| "grad_norm": 1.4716849327087402, | |
| "learning_rate": 2.3676470588235295e-05, | |
| "loss": 0.0855, | |
| "step": 651 | |
| }, | |
| { | |
| "epoch": 3.442904290429043, | |
| "grad_norm": 1.3064403533935547, | |
| "learning_rate": 2.3627450980392156e-05, | |
| "loss": 0.0644, | |
| "step": 652 | |
| }, | |
| { | |
| "epoch": 3.448184818481848, | |
| "grad_norm": 1.4186874628067017, | |
| "learning_rate": 2.357843137254902e-05, | |
| "loss": 0.0556, | |
| "step": 653 | |
| }, | |
| { | |
| "epoch": 3.4534653465346534, | |
| "grad_norm": 1.3081494569778442, | |
| "learning_rate": 2.3529411764705884e-05, | |
| "loss": 0.0663, | |
| "step": 654 | |
| }, | |
| { | |
| "epoch": 3.458745874587459, | |
| "grad_norm": 1.174717664718628, | |
| "learning_rate": 2.3480392156862748e-05, | |
| "loss": 0.0704, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 3.464026402640264, | |
| "grad_norm": 1.310571551322937, | |
| "learning_rate": 2.3431372549019608e-05, | |
| "loss": 0.0555, | |
| "step": 656 | |
| }, | |
| { | |
| "epoch": 3.4693069306930693, | |
| "grad_norm": 1.3356480598449707, | |
| "learning_rate": 2.3382352941176472e-05, | |
| "loss": 0.0565, | |
| "step": 657 | |
| }, | |
| { | |
| "epoch": 3.4745874587458747, | |
| "grad_norm": 1.5645304918289185, | |
| "learning_rate": 2.3333333333333336e-05, | |
| "loss": 0.0885, | |
| "step": 658 | |
| }, | |
| { | |
| "epoch": 3.4798679867986797, | |
| "grad_norm": 1.6264077425003052, | |
| "learning_rate": 2.3284313725490196e-05, | |
| "loss": 0.0936, | |
| "step": 659 | |
| }, | |
| { | |
| "epoch": 3.485148514851485, | |
| "grad_norm": 1.1637386083602905, | |
| "learning_rate": 2.323529411764706e-05, | |
| "loss": 0.0499, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 3.4904290429042906, | |
| "grad_norm": 1.3853108882904053, | |
| "learning_rate": 2.3186274509803924e-05, | |
| "loss": 0.0781, | |
| "step": 661 | |
| }, | |
| { | |
| "epoch": 3.4957095709570956, | |
| "grad_norm": 1.1978479623794556, | |
| "learning_rate": 2.3137254901960788e-05, | |
| "loss": 0.0524, | |
| "step": 662 | |
| }, | |
| { | |
| "epoch": 3.500990099009901, | |
| "grad_norm": 1.1041313409805298, | |
| "learning_rate": 2.308823529411765e-05, | |
| "loss": 0.054, | |
| "step": 663 | |
| }, | |
| { | |
| "epoch": 3.5062706270627064, | |
| "grad_norm": 0.9631710052490234, | |
| "learning_rate": 2.303921568627451e-05, | |
| "loss": 0.0393, | |
| "step": 664 | |
| }, | |
| { | |
| "epoch": 3.5115511551155114, | |
| "grad_norm": 1.166167974472046, | |
| "learning_rate": 2.2990196078431373e-05, | |
| "loss": 0.0401, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 3.516831683168317, | |
| "grad_norm": 1.7196093797683716, | |
| "learning_rate": 2.2941176470588237e-05, | |
| "loss": 0.0686, | |
| "step": 666 | |
| }, | |
| { | |
| "epoch": 3.5221122112211223, | |
| "grad_norm": 1.4491196870803833, | |
| "learning_rate": 2.2892156862745097e-05, | |
| "loss": 0.0721, | |
| "step": 667 | |
| }, | |
| { | |
| "epoch": 3.5273927392739273, | |
| "grad_norm": 1.724544882774353, | |
| "learning_rate": 2.284313725490196e-05, | |
| "loss": 0.0789, | |
| "step": 668 | |
| }, | |
| { | |
| "epoch": 3.5326732673267327, | |
| "grad_norm": 1.2871899604797363, | |
| "learning_rate": 2.2794117647058825e-05, | |
| "loss": 0.0502, | |
| "step": 669 | |
| }, | |
| { | |
| "epoch": 3.537953795379538, | |
| "grad_norm": 1.7234700918197632, | |
| "learning_rate": 2.2745098039215685e-05, | |
| "loss": 0.0896, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 3.543234323432343, | |
| "grad_norm": 1.6124573945999146, | |
| "learning_rate": 2.269607843137255e-05, | |
| "loss": 0.0899, | |
| "step": 671 | |
| }, | |
| { | |
| "epoch": 3.5485148514851486, | |
| "grad_norm": 1.1992857456207275, | |
| "learning_rate": 2.2647058823529413e-05, | |
| "loss": 0.0514, | |
| "step": 672 | |
| }, | |
| { | |
| "epoch": 3.553795379537954, | |
| "grad_norm": 1.2639023065567017, | |
| "learning_rate": 2.2598039215686277e-05, | |
| "loss": 0.0584, | |
| "step": 673 | |
| }, | |
| { | |
| "epoch": 3.559075907590759, | |
| "grad_norm": 1.1662899255752563, | |
| "learning_rate": 2.2549019607843138e-05, | |
| "loss": 0.0551, | |
| "step": 674 | |
| }, | |
| { | |
| "epoch": 3.5643564356435644, | |
| "grad_norm": 1.27886962890625, | |
| "learning_rate": 2.25e-05, | |
| "loss": 0.0706, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 3.56963696369637, | |
| "grad_norm": 1.487029790878296, | |
| "learning_rate": 2.2450980392156865e-05, | |
| "loss": 0.0973, | |
| "step": 676 | |
| }, | |
| { | |
| "epoch": 3.574917491749175, | |
| "grad_norm": 1.0966662168502808, | |
| "learning_rate": 2.2401960784313726e-05, | |
| "loss": 0.0483, | |
| "step": 677 | |
| }, | |
| { | |
| "epoch": 3.5801980198019803, | |
| "grad_norm": 1.592883825302124, | |
| "learning_rate": 2.235294117647059e-05, | |
| "loss": 0.097, | |
| "step": 678 | |
| }, | |
| { | |
| "epoch": 3.5854785478547857, | |
| "grad_norm": 1.1297260522842407, | |
| "learning_rate": 2.2303921568627454e-05, | |
| "loss": 0.052, | |
| "step": 679 | |
| }, | |
| { | |
| "epoch": 3.5907590759075907, | |
| "grad_norm": 1.3549542427062988, | |
| "learning_rate": 2.2254901960784314e-05, | |
| "loss": 0.0636, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 3.596039603960396, | |
| "grad_norm": 1.4858061075210571, | |
| "learning_rate": 2.2205882352941178e-05, | |
| "loss": 0.0823, | |
| "step": 681 | |
| }, | |
| { | |
| "epoch": 3.6013201320132016, | |
| "grad_norm": 1.2688394784927368, | |
| "learning_rate": 2.215686274509804e-05, | |
| "loss": 0.0621, | |
| "step": 682 | |
| }, | |
| { | |
| "epoch": 3.6066006600660065, | |
| "grad_norm": 1.2265433073043823, | |
| "learning_rate": 2.2107843137254902e-05, | |
| "loss": 0.0521, | |
| "step": 683 | |
| }, | |
| { | |
| "epoch": 3.611881188118812, | |
| "grad_norm": 1.5389267206192017, | |
| "learning_rate": 2.2058823529411766e-05, | |
| "loss": 0.0644, | |
| "step": 684 | |
| }, | |
| { | |
| "epoch": 3.6171617161716174, | |
| "grad_norm": 1.0617897510528564, | |
| "learning_rate": 2.2009803921568627e-05, | |
| "loss": 0.0487, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 3.6224422442244224, | |
| "grad_norm": 1.5505889654159546, | |
| "learning_rate": 2.196078431372549e-05, | |
| "loss": 0.0727, | |
| "step": 686 | |
| }, | |
| { | |
| "epoch": 3.627722772277228, | |
| "grad_norm": 1.6774746179580688, | |
| "learning_rate": 2.1911764705882354e-05, | |
| "loss": 0.0763, | |
| "step": 687 | |
| }, | |
| { | |
| "epoch": 3.6330033003300333, | |
| "grad_norm": 1.253771185874939, | |
| "learning_rate": 2.1862745098039218e-05, | |
| "loss": 0.0629, | |
| "step": 688 | |
| }, | |
| { | |
| "epoch": 3.6382838283828383, | |
| "grad_norm": 1.324569821357727, | |
| "learning_rate": 2.181372549019608e-05, | |
| "loss": 0.0673, | |
| "step": 689 | |
| }, | |
| { | |
| "epoch": 3.6435643564356437, | |
| "grad_norm": 1.7756513357162476, | |
| "learning_rate": 2.1764705882352943e-05, | |
| "loss": 0.1118, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 3.6488448844884487, | |
| "grad_norm": 1.2172956466674805, | |
| "learning_rate": 2.1715686274509807e-05, | |
| "loss": 0.0553, | |
| "step": 691 | |
| }, | |
| { | |
| "epoch": 3.654125412541254, | |
| "grad_norm": 1.201130747795105, | |
| "learning_rate": 2.1666666666666667e-05, | |
| "loss": 0.0654, | |
| "step": 692 | |
| }, | |
| { | |
| "epoch": 3.6594059405940595, | |
| "grad_norm": 1.3230106830596924, | |
| "learning_rate": 2.161764705882353e-05, | |
| "loss": 0.0647, | |
| "step": 693 | |
| }, | |
| { | |
| "epoch": 3.6646864686468645, | |
| "grad_norm": 1.3373692035675049, | |
| "learning_rate": 2.1568627450980395e-05, | |
| "loss": 0.0691, | |
| "step": 694 | |
| }, | |
| { | |
| "epoch": 3.66996699669967, | |
| "grad_norm": 1.423130750656128, | |
| "learning_rate": 2.151960784313726e-05, | |
| "loss": 0.0683, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 3.6752475247524754, | |
| "grad_norm": 1.2763397693634033, | |
| "learning_rate": 2.1470588235294116e-05, | |
| "loss": 0.0589, | |
| "step": 696 | |
| }, | |
| { | |
| "epoch": 3.6805280528052804, | |
| "grad_norm": 1.749027967453003, | |
| "learning_rate": 2.142156862745098e-05, | |
| "loss": 0.0918, | |
| "step": 697 | |
| }, | |
| { | |
| "epoch": 3.685808580858086, | |
| "grad_norm": 1.3163336515426636, | |
| "learning_rate": 2.1372549019607844e-05, | |
| "loss": 0.0585, | |
| "step": 698 | |
| }, | |
| { | |
| "epoch": 3.691089108910891, | |
| "grad_norm": 1.6769047975540161, | |
| "learning_rate": 2.1323529411764707e-05, | |
| "loss": 0.0967, | |
| "step": 699 | |
| }, | |
| { | |
| "epoch": 3.6963696369636962, | |
| "grad_norm": 1.2727911472320557, | |
| "learning_rate": 2.1274509803921568e-05, | |
| "loss": 0.0713, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 3.7016501650165017, | |
| "grad_norm": 1.4433225393295288, | |
| "learning_rate": 2.1225490196078432e-05, | |
| "loss": 0.0783, | |
| "step": 701 | |
| }, | |
| { | |
| "epoch": 3.7069306930693067, | |
| "grad_norm": 1.3980076313018799, | |
| "learning_rate": 2.1176470588235296e-05, | |
| "loss": 0.0677, | |
| "step": 702 | |
| }, | |
| { | |
| "epoch": 3.712211221122112, | |
| "grad_norm": 1.3446606397628784, | |
| "learning_rate": 2.1127450980392156e-05, | |
| "loss": 0.0746, | |
| "step": 703 | |
| }, | |
| { | |
| "epoch": 3.7174917491749175, | |
| "grad_norm": 1.2781853675842285, | |
| "learning_rate": 2.107843137254902e-05, | |
| "loss": 0.0545, | |
| "step": 704 | |
| }, | |
| { | |
| "epoch": 3.7227722772277225, | |
| "grad_norm": 1.4710532426834106, | |
| "learning_rate": 2.1029411764705884e-05, | |
| "loss": 0.0612, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 3.728052805280528, | |
| "grad_norm": 1.2814794778823853, | |
| "learning_rate": 2.0980392156862748e-05, | |
| "loss": 0.0718, | |
| "step": 706 | |
| }, | |
| { | |
| "epoch": 3.7333333333333334, | |
| "grad_norm": 1.4661047458648682, | |
| "learning_rate": 2.0931372549019608e-05, | |
| "loss": 0.0791, | |
| "step": 707 | |
| }, | |
| { | |
| "epoch": 3.7386138613861384, | |
| "grad_norm": 1.37482750415802, | |
| "learning_rate": 2.0882352941176472e-05, | |
| "loss": 0.0737, | |
| "step": 708 | |
| }, | |
| { | |
| "epoch": 3.743894389438944, | |
| "grad_norm": 1.3684672117233276, | |
| "learning_rate": 2.0833333333333336e-05, | |
| "loss": 0.0743, | |
| "step": 709 | |
| }, | |
| { | |
| "epoch": 3.7491749174917492, | |
| "grad_norm": 0.859553873538971, | |
| "learning_rate": 2.0784313725490197e-05, | |
| "loss": 0.0356, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 3.7544554455445542, | |
| "grad_norm": 1.2247698307037354, | |
| "learning_rate": 2.073529411764706e-05, | |
| "loss": 0.0671, | |
| "step": 711 | |
| }, | |
| { | |
| "epoch": 3.7597359735973597, | |
| "grad_norm": 1.2793521881103516, | |
| "learning_rate": 2.068627450980392e-05, | |
| "loss": 0.0692, | |
| "step": 712 | |
| }, | |
| { | |
| "epoch": 3.765016501650165, | |
| "grad_norm": 1.26449453830719, | |
| "learning_rate": 2.0637254901960785e-05, | |
| "loss": 0.0622, | |
| "step": 713 | |
| }, | |
| { | |
| "epoch": 3.77029702970297, | |
| "grad_norm": 1.1848664283752441, | |
| "learning_rate": 2.058823529411765e-05, | |
| "loss": 0.0535, | |
| "step": 714 | |
| }, | |
| { | |
| "epoch": 3.7755775577557755, | |
| "grad_norm": 1.034454107284546, | |
| "learning_rate": 2.053921568627451e-05, | |
| "loss": 0.0403, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 3.780858085808581, | |
| "grad_norm": 1.005811333656311, | |
| "learning_rate": 2.0490196078431373e-05, | |
| "loss": 0.0489, | |
| "step": 716 | |
| }, | |
| { | |
| "epoch": 3.786138613861386, | |
| "grad_norm": 1.294252872467041, | |
| "learning_rate": 2.0441176470588237e-05, | |
| "loss": 0.0599, | |
| "step": 717 | |
| }, | |
| { | |
| "epoch": 3.7914191419141914, | |
| "grad_norm": 1.2683149576187134, | |
| "learning_rate": 2.0392156862745097e-05, | |
| "loss": 0.0527, | |
| "step": 718 | |
| }, | |
| { | |
| "epoch": 3.796699669966997, | |
| "grad_norm": 1.4859176874160767, | |
| "learning_rate": 2.034313725490196e-05, | |
| "loss": 0.0748, | |
| "step": 719 | |
| }, | |
| { | |
| "epoch": 3.801980198019802, | |
| "grad_norm": 1.5484555959701538, | |
| "learning_rate": 2.0294117647058825e-05, | |
| "loss": 0.0694, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 3.8072607260726072, | |
| "grad_norm": 1.0861327648162842, | |
| "learning_rate": 2.024509803921569e-05, | |
| "loss": 0.0489, | |
| "step": 721 | |
| }, | |
| { | |
| "epoch": 3.8125412541254127, | |
| "grad_norm": 2.1083528995513916, | |
| "learning_rate": 2.019607843137255e-05, | |
| "loss": 0.0996, | |
| "step": 722 | |
| }, | |
| { | |
| "epoch": 3.8178217821782177, | |
| "grad_norm": 1.1636090278625488, | |
| "learning_rate": 2.0147058823529413e-05, | |
| "loss": 0.0439, | |
| "step": 723 | |
| }, | |
| { | |
| "epoch": 3.823102310231023, | |
| "grad_norm": 1.4956183433532715, | |
| "learning_rate": 2.0098039215686277e-05, | |
| "loss": 0.0771, | |
| "step": 724 | |
| }, | |
| { | |
| "epoch": 3.8283828382838285, | |
| "grad_norm": 1.4790761470794678, | |
| "learning_rate": 2.0049019607843138e-05, | |
| "loss": 0.0561, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 3.8336633663366335, | |
| "grad_norm": 1.4751077890396118, | |
| "learning_rate": 2e-05, | |
| "loss": 0.0788, | |
| "step": 726 | |
| }, | |
| { | |
| "epoch": 3.838943894389439, | |
| "grad_norm": 1.2702194452285767, | |
| "learning_rate": 1.9950980392156866e-05, | |
| "loss": 0.07, | |
| "step": 727 | |
| }, | |
| { | |
| "epoch": 3.8442244224422444, | |
| "grad_norm": 0.9263429045677185, | |
| "learning_rate": 1.9901960784313726e-05, | |
| "loss": 0.0359, | |
| "step": 728 | |
| }, | |
| { | |
| "epoch": 3.8495049504950494, | |
| "grad_norm": 1.4288864135742188, | |
| "learning_rate": 1.9852941176470586e-05, | |
| "loss": 0.0869, | |
| "step": 729 | |
| }, | |
| { | |
| "epoch": 3.854785478547855, | |
| "grad_norm": 1.3054994344711304, | |
| "learning_rate": 1.980392156862745e-05, | |
| "loss": 0.0755, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 3.8600660066006602, | |
| "grad_norm": 1.422020435333252, | |
| "learning_rate": 1.9754901960784314e-05, | |
| "loss": 0.072, | |
| "step": 731 | |
| }, | |
| { | |
| "epoch": 3.8653465346534652, | |
| "grad_norm": 1.2863701581954956, | |
| "learning_rate": 1.9705882352941178e-05, | |
| "loss": 0.0676, | |
| "step": 732 | |
| }, | |
| { | |
| "epoch": 3.8706270627062707, | |
| "grad_norm": 1.4083125591278076, | |
| "learning_rate": 1.965686274509804e-05, | |
| "loss": 0.0805, | |
| "step": 733 | |
| }, | |
| { | |
| "epoch": 3.875907590759076, | |
| "grad_norm": 1.275930404663086, | |
| "learning_rate": 1.9607843137254903e-05, | |
| "loss": 0.0737, | |
| "step": 734 | |
| }, | |
| { | |
| "epoch": 3.881188118811881, | |
| "grad_norm": 1.2777963876724243, | |
| "learning_rate": 1.9558823529411766e-05, | |
| "loss": 0.0674, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 3.8864686468646865, | |
| "grad_norm": 1.145695686340332, | |
| "learning_rate": 1.9509803921568627e-05, | |
| "loss": 0.0593, | |
| "step": 736 | |
| }, | |
| { | |
| "epoch": 3.891749174917492, | |
| "grad_norm": 1.13607919216156, | |
| "learning_rate": 1.946078431372549e-05, | |
| "loss": 0.0481, | |
| "step": 737 | |
| }, | |
| { | |
| "epoch": 3.897029702970297, | |
| "grad_norm": 1.5105829238891602, | |
| "learning_rate": 1.9411764705882355e-05, | |
| "loss": 0.1056, | |
| "step": 738 | |
| }, | |
| { | |
| "epoch": 3.9023102310231024, | |
| "grad_norm": 1.080869197845459, | |
| "learning_rate": 1.936274509803922e-05, | |
| "loss": 0.0469, | |
| "step": 739 | |
| }, | |
| { | |
| "epoch": 3.907590759075908, | |
| "grad_norm": 1.6012327671051025, | |
| "learning_rate": 1.931372549019608e-05, | |
| "loss": 0.0883, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 3.912871287128713, | |
| "grad_norm": 1.3881632089614868, | |
| "learning_rate": 1.9264705882352943e-05, | |
| "loss": 0.0736, | |
| "step": 741 | |
| }, | |
| { | |
| "epoch": 3.9181518151815182, | |
| "grad_norm": 1.2698092460632324, | |
| "learning_rate": 1.9215686274509807e-05, | |
| "loss": 0.061, | |
| "step": 742 | |
| }, | |
| { | |
| "epoch": 3.9234323432343237, | |
| "grad_norm": 1.7859208583831787, | |
| "learning_rate": 1.9166666666666667e-05, | |
| "loss": 0.0981, | |
| "step": 743 | |
| }, | |
| { | |
| "epoch": 3.9287128712871286, | |
| "grad_norm": 1.1902238130569458, | |
| "learning_rate": 1.9117647058823528e-05, | |
| "loss": 0.0489, | |
| "step": 744 | |
| }, | |
| { | |
| "epoch": 3.933993399339934, | |
| "grad_norm": 1.3295652866363525, | |
| "learning_rate": 1.906862745098039e-05, | |
| "loss": 0.0634, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 3.9392739273927395, | |
| "grad_norm": 1.2844423055648804, | |
| "learning_rate": 1.9019607843137255e-05, | |
| "loss": 0.0759, | |
| "step": 746 | |
| }, | |
| { | |
| "epoch": 3.9445544554455445, | |
| "grad_norm": 1.062388300895691, | |
| "learning_rate": 1.897058823529412e-05, | |
| "loss": 0.0449, | |
| "step": 747 | |
| }, | |
| { | |
| "epoch": 3.94983498349835, | |
| "grad_norm": 1.51730477809906, | |
| "learning_rate": 1.892156862745098e-05, | |
| "loss": 0.1383, | |
| "step": 748 | |
| }, | |
| { | |
| "epoch": 3.9551155115511554, | |
| "grad_norm": 1.5683430433273315, | |
| "learning_rate": 1.8872549019607844e-05, | |
| "loss": 0.0915, | |
| "step": 749 | |
| }, | |
| { | |
| "epoch": 3.9603960396039604, | |
| "grad_norm": 1.20012366771698, | |
| "learning_rate": 1.8823529411764708e-05, | |
| "loss": 0.057, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 3.965676567656766, | |
| "grad_norm": 1.351365089416504, | |
| "learning_rate": 1.8774509803921568e-05, | |
| "loss": 0.0663, | |
| "step": 751 | |
| }, | |
| { | |
| "epoch": 3.9709570957095712, | |
| "grad_norm": 1.5210295915603638, | |
| "learning_rate": 1.8725490196078432e-05, | |
| "loss": 0.0832, | |
| "step": 752 | |
| }, | |
| { | |
| "epoch": 3.976237623762376, | |
| "grad_norm": 1.5740433931350708, | |
| "learning_rate": 1.8676470588235296e-05, | |
| "loss": 0.0951, | |
| "step": 753 | |
| }, | |
| { | |
| "epoch": 3.9815181518151816, | |
| "grad_norm": 1.2916804552078247, | |
| "learning_rate": 1.862745098039216e-05, | |
| "loss": 0.0638, | |
| "step": 754 | |
| }, | |
| { | |
| "epoch": 3.9867986798679866, | |
| "grad_norm": 1.4912750720977783, | |
| "learning_rate": 1.857843137254902e-05, | |
| "loss": 0.0718, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 3.992079207920792, | |
| "grad_norm": 1.2898180484771729, | |
| "learning_rate": 1.8529411764705884e-05, | |
| "loss": 0.0672, | |
| "step": 756 | |
| }, | |
| { | |
| "epoch": 3.9973597359735975, | |
| "grad_norm": 0.9569465517997742, | |
| "learning_rate": 1.8480392156862748e-05, | |
| "loss": 0.0425, | |
| "step": 757 | |
| }, | |
| { | |
| "epoch": 4.002640264026403, | |
| "grad_norm": 1.1869324445724487, | |
| "learning_rate": 1.843137254901961e-05, | |
| "loss": 0.0527, | |
| "step": 758 | |
| }, | |
| { | |
| "epoch": 4.007920792079208, | |
| "grad_norm": 0.7702716588973999, | |
| "learning_rate": 1.8382352941176472e-05, | |
| "loss": 0.0358, | |
| "step": 759 | |
| }, | |
| { | |
| "epoch": 4.013201320132013, | |
| "grad_norm": 0.6784669160842896, | |
| "learning_rate": 1.8333333333333333e-05, | |
| "loss": 0.0268, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 4.018481848184819, | |
| "grad_norm": 0.8795797228813171, | |
| "learning_rate": 1.8284313725490197e-05, | |
| "loss": 0.0363, | |
| "step": 761 | |
| }, | |
| { | |
| "epoch": 4.023762376237624, | |
| "grad_norm": 0.9152675271034241, | |
| "learning_rate": 1.8235294117647057e-05, | |
| "loss": 0.0303, | |
| "step": 762 | |
| }, | |
| { | |
| "epoch": 4.029042904290429, | |
| "grad_norm": 0.685607373714447, | |
| "learning_rate": 1.818627450980392e-05, | |
| "loss": 0.0249, | |
| "step": 763 | |
| }, | |
| { | |
| "epoch": 4.034323432343235, | |
| "grad_norm": 0.8284955620765686, | |
| "learning_rate": 1.8137254901960785e-05, | |
| "loss": 0.0257, | |
| "step": 764 | |
| }, | |
| { | |
| "epoch": 4.03960396039604, | |
| "grad_norm": 0.8242619037628174, | |
| "learning_rate": 1.808823529411765e-05, | |
| "loss": 0.027, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 4.044884488448845, | |
| "grad_norm": 1.0114595890045166, | |
| "learning_rate": 1.803921568627451e-05, | |
| "loss": 0.0222, | |
| "step": 766 | |
| }, | |
| { | |
| "epoch": 4.0501650165016505, | |
| "grad_norm": 0.6319472789764404, | |
| "learning_rate": 1.7990196078431373e-05, | |
| "loss": 0.0182, | |
| "step": 767 | |
| }, | |
| { | |
| "epoch": 4.0554455445544555, | |
| "grad_norm": 0.9475066661834717, | |
| "learning_rate": 1.7941176470588237e-05, | |
| "loss": 0.0269, | |
| "step": 768 | |
| }, | |
| { | |
| "epoch": 4.0607260726072605, | |
| "grad_norm": 1.0184468030929565, | |
| "learning_rate": 1.7892156862745098e-05, | |
| "loss": 0.0258, | |
| "step": 769 | |
| }, | |
| { | |
| "epoch": 4.066006600660066, | |
| "grad_norm": 1.226219654083252, | |
| "learning_rate": 1.784313725490196e-05, | |
| "loss": 0.036, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 4.071287128712871, | |
| "grad_norm": 0.839829683303833, | |
| "learning_rate": 1.7794117647058825e-05, | |
| "loss": 0.027, | |
| "step": 771 | |
| }, | |
| { | |
| "epoch": 4.076567656765676, | |
| "grad_norm": 1.2068742513656616, | |
| "learning_rate": 1.774509803921569e-05, | |
| "loss": 0.0345, | |
| "step": 772 | |
| }, | |
| { | |
| "epoch": 4.081848184818482, | |
| "grad_norm": 1.1739568710327148, | |
| "learning_rate": 1.769607843137255e-05, | |
| "loss": 0.0334, | |
| "step": 773 | |
| }, | |
| { | |
| "epoch": 4.087128712871287, | |
| "grad_norm": 1.167466640472412, | |
| "learning_rate": 1.7647058823529414e-05, | |
| "loss": 0.0308, | |
| "step": 774 | |
| }, | |
| { | |
| "epoch": 4.092409240924092, | |
| "grad_norm": 1.0854191780090332, | |
| "learning_rate": 1.7598039215686277e-05, | |
| "loss": 0.0288, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 4.097689768976898, | |
| "grad_norm": 0.8785208463668823, | |
| "learning_rate": 1.7549019607843138e-05, | |
| "loss": 0.0298, | |
| "step": 776 | |
| }, | |
| { | |
| "epoch": 4.102970297029703, | |
| "grad_norm": 1.252079725265503, | |
| "learning_rate": 1.75e-05, | |
| "loss": 0.044, | |
| "step": 777 | |
| }, | |
| { | |
| "epoch": 4.108250825082508, | |
| "grad_norm": 1.1048779487609863, | |
| "learning_rate": 1.7450980392156862e-05, | |
| "loss": 0.0285, | |
| "step": 778 | |
| }, | |
| { | |
| "epoch": 4.113531353135314, | |
| "grad_norm": 1.3396317958831787, | |
| "learning_rate": 1.7401960784313726e-05, | |
| "loss": 0.0457, | |
| "step": 779 | |
| }, | |
| { | |
| "epoch": 4.118811881188119, | |
| "grad_norm": 1.165590763092041, | |
| "learning_rate": 1.735294117647059e-05, | |
| "loss": 0.0319, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 4.124092409240924, | |
| "grad_norm": 0.9108593463897705, | |
| "learning_rate": 1.730392156862745e-05, | |
| "loss": 0.0286, | |
| "step": 781 | |
| }, | |
| { | |
| "epoch": 4.12937293729373, | |
| "grad_norm": 0.9101956486701965, | |
| "learning_rate": 1.7254901960784314e-05, | |
| "loss": 0.0291, | |
| "step": 782 | |
| }, | |
| { | |
| "epoch": 4.134653465346535, | |
| "grad_norm": 1.055116891860962, | |
| "learning_rate": 1.720588235294118e-05, | |
| "loss": 0.0326, | |
| "step": 783 | |
| }, | |
| { | |
| "epoch": 4.13993399339934, | |
| "grad_norm": 0.9137353897094727, | |
| "learning_rate": 1.715686274509804e-05, | |
| "loss": 0.0312, | |
| "step": 784 | |
| }, | |
| { | |
| "epoch": 4.145214521452146, | |
| "grad_norm": 0.9962388277053833, | |
| "learning_rate": 1.7107843137254903e-05, | |
| "loss": 0.0317, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 4.150495049504951, | |
| "grad_norm": 0.891434907913208, | |
| "learning_rate": 1.7058823529411767e-05, | |
| "loss": 0.0357, | |
| "step": 786 | |
| }, | |
| { | |
| "epoch": 4.155775577557756, | |
| "grad_norm": 0.9022417068481445, | |
| "learning_rate": 1.700980392156863e-05, | |
| "loss": 0.0346, | |
| "step": 787 | |
| }, | |
| { | |
| "epoch": 4.161056105610561, | |
| "grad_norm": 0.886080265045166, | |
| "learning_rate": 1.696078431372549e-05, | |
| "loss": 0.0308, | |
| "step": 788 | |
| }, | |
| { | |
| "epoch": 4.1663366336633665, | |
| "grad_norm": 1.155617356300354, | |
| "learning_rate": 1.6911764705882355e-05, | |
| "loss": 0.0904, | |
| "step": 789 | |
| }, | |
| { | |
| "epoch": 4.1716171617161715, | |
| "grad_norm": 1.101927399635315, | |
| "learning_rate": 1.686274509803922e-05, | |
| "loss": 0.0289, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 4.1768976897689765, | |
| "grad_norm": 1.2381272315979004, | |
| "learning_rate": 1.681372549019608e-05, | |
| "loss": 0.0473, | |
| "step": 791 | |
| }, | |
| { | |
| "epoch": 4.182178217821782, | |
| "grad_norm": 1.2841603755950928, | |
| "learning_rate": 1.676470588235294e-05, | |
| "loss": 0.0392, | |
| "step": 792 | |
| }, | |
| { | |
| "epoch": 4.187458745874587, | |
| "grad_norm": 0.9901617765426636, | |
| "learning_rate": 1.6715686274509804e-05, | |
| "loss": 0.0274, | |
| "step": 793 | |
| }, | |
| { | |
| "epoch": 4.192739273927392, | |
| "grad_norm": 1.011318325996399, | |
| "learning_rate": 1.6666666666666667e-05, | |
| "loss": 0.0329, | |
| "step": 794 | |
| }, | |
| { | |
| "epoch": 4.198019801980198, | |
| "grad_norm": 0.8173012733459473, | |
| "learning_rate": 1.6617647058823528e-05, | |
| "loss": 0.0272, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 4.203300330033003, | |
| "grad_norm": 1.3152134418487549, | |
| "learning_rate": 1.6568627450980392e-05, | |
| "loss": 0.0309, | |
| "step": 796 | |
| }, | |
| { | |
| "epoch": 4.208580858085808, | |
| "grad_norm": 0.9203322529792786, | |
| "learning_rate": 1.6519607843137256e-05, | |
| "loss": 0.0287, | |
| "step": 797 | |
| }, | |
| { | |
| "epoch": 4.213861386138614, | |
| "grad_norm": 0.8306743502616882, | |
| "learning_rate": 1.647058823529412e-05, | |
| "loss": 0.0291, | |
| "step": 798 | |
| }, | |
| { | |
| "epoch": 4.219141914191419, | |
| "grad_norm": 1.1384245157241821, | |
| "learning_rate": 1.642156862745098e-05, | |
| "loss": 0.0295, | |
| "step": 799 | |
| }, | |
| { | |
| "epoch": 4.224422442244224, | |
| "grad_norm": 1.0017738342285156, | |
| "learning_rate": 1.6372549019607844e-05, | |
| "loss": 0.0347, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 4.22970297029703, | |
| "grad_norm": 0.7871435284614563, | |
| "learning_rate": 1.6323529411764708e-05, | |
| "loss": 0.0277, | |
| "step": 801 | |
| }, | |
| { | |
| "epoch": 4.234983498349835, | |
| "grad_norm": 0.9366597533226013, | |
| "learning_rate": 1.627450980392157e-05, | |
| "loss": 0.0323, | |
| "step": 802 | |
| }, | |
| { | |
| "epoch": 4.24026402640264, | |
| "grad_norm": 0.9586142301559448, | |
| "learning_rate": 1.6225490196078432e-05, | |
| "loss": 0.0212, | |
| "step": 803 | |
| }, | |
| { | |
| "epoch": 4.245544554455446, | |
| "grad_norm": 1.21640145778656, | |
| "learning_rate": 1.6176470588235296e-05, | |
| "loss": 0.0357, | |
| "step": 804 | |
| }, | |
| { | |
| "epoch": 4.250825082508251, | |
| "grad_norm": 0.9384037852287292, | |
| "learning_rate": 1.612745098039216e-05, | |
| "loss": 0.0315, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 4.256105610561056, | |
| "grad_norm": 1.0717084407806396, | |
| "learning_rate": 1.607843137254902e-05, | |
| "loss": 0.0337, | |
| "step": 806 | |
| }, | |
| { | |
| "epoch": 4.261386138613862, | |
| "grad_norm": 0.9708730578422546, | |
| "learning_rate": 1.6029411764705884e-05, | |
| "loss": 0.0319, | |
| "step": 807 | |
| }, | |
| { | |
| "epoch": 4.266666666666667, | |
| "grad_norm": 0.7535268068313599, | |
| "learning_rate": 1.5980392156862745e-05, | |
| "loss": 0.0294, | |
| "step": 808 | |
| }, | |
| { | |
| "epoch": 4.271947194719472, | |
| "grad_norm": 1.0513904094696045, | |
| "learning_rate": 1.593137254901961e-05, | |
| "loss": 0.0262, | |
| "step": 809 | |
| }, | |
| { | |
| "epoch": 4.2772277227722775, | |
| "grad_norm": 1.4030704498291016, | |
| "learning_rate": 1.588235294117647e-05, | |
| "loss": 0.0436, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 4.2825082508250825, | |
| "grad_norm": 0.5738725662231445, | |
| "learning_rate": 1.5833333333333333e-05, | |
| "loss": 0.0201, | |
| "step": 811 | |
| }, | |
| { | |
| "epoch": 4.2877887788778875, | |
| "grad_norm": 1.305862307548523, | |
| "learning_rate": 1.5784313725490197e-05, | |
| "loss": 0.043, | |
| "step": 812 | |
| }, | |
| { | |
| "epoch": 4.293069306930693, | |
| "grad_norm": 0.9697719812393188, | |
| "learning_rate": 1.573529411764706e-05, | |
| "loss": 0.0306, | |
| "step": 813 | |
| }, | |
| { | |
| "epoch": 4.298349834983498, | |
| "grad_norm": 0.8352718353271484, | |
| "learning_rate": 1.568627450980392e-05, | |
| "loss": 0.0269, | |
| "step": 814 | |
| }, | |
| { | |
| "epoch": 4.303630363036303, | |
| "grad_norm": 1.000013828277588, | |
| "learning_rate": 1.5637254901960785e-05, | |
| "loss": 0.0272, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 4.308910891089109, | |
| "grad_norm": 0.6394527554512024, | |
| "learning_rate": 1.558823529411765e-05, | |
| "loss": 0.0224, | |
| "step": 816 | |
| }, | |
| { | |
| "epoch": 4.314191419141914, | |
| "grad_norm": 0.9506754875183105, | |
| "learning_rate": 1.553921568627451e-05, | |
| "loss": 0.0293, | |
| "step": 817 | |
| }, | |
| { | |
| "epoch": 4.319471947194719, | |
| "grad_norm": 0.9406654834747314, | |
| "learning_rate": 1.5490196078431373e-05, | |
| "loss": 0.0261, | |
| "step": 818 | |
| }, | |
| { | |
| "epoch": 4.324752475247525, | |
| "grad_norm": 1.1264249086380005, | |
| "learning_rate": 1.5441176470588237e-05, | |
| "loss": 0.0347, | |
| "step": 819 | |
| }, | |
| { | |
| "epoch": 4.33003300330033, | |
| "grad_norm": 0.9461225271224976, | |
| "learning_rate": 1.53921568627451e-05, | |
| "loss": 0.0261, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 4.335313531353135, | |
| "grad_norm": 1.1620982885360718, | |
| "learning_rate": 1.5343137254901962e-05, | |
| "loss": 0.0309, | |
| "step": 821 | |
| }, | |
| { | |
| "epoch": 4.340594059405941, | |
| "grad_norm": 0.8539828658103943, | |
| "learning_rate": 1.5294117647058826e-05, | |
| "loss": 0.028, | |
| "step": 822 | |
| }, | |
| { | |
| "epoch": 4.345874587458746, | |
| "grad_norm": 1.1822220087051392, | |
| "learning_rate": 1.5245098039215688e-05, | |
| "loss": 0.0352, | |
| "step": 823 | |
| }, | |
| { | |
| "epoch": 4.351155115511551, | |
| "grad_norm": 1.000964641571045, | |
| "learning_rate": 1.5196078431372548e-05, | |
| "loss": 0.0325, | |
| "step": 824 | |
| }, | |
| { | |
| "epoch": 4.356435643564357, | |
| "grad_norm": 1.211593508720398, | |
| "learning_rate": 1.5147058823529412e-05, | |
| "loss": 0.0394, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 4.361716171617162, | |
| "grad_norm": 1.0759446620941162, | |
| "learning_rate": 1.5098039215686274e-05, | |
| "loss": 0.0299, | |
| "step": 826 | |
| }, | |
| { | |
| "epoch": 4.366996699669967, | |
| "grad_norm": 0.7947413921356201, | |
| "learning_rate": 1.5049019607843137e-05, | |
| "loss": 0.0269, | |
| "step": 827 | |
| }, | |
| { | |
| "epoch": 4.372277227722773, | |
| "grad_norm": 1.283584713935852, | |
| "learning_rate": 1.5e-05, | |
| "loss": 0.0396, | |
| "step": 828 | |
| }, | |
| { | |
| "epoch": 4.377557755775578, | |
| "grad_norm": 1.0424344539642334, | |
| "learning_rate": 1.4950980392156863e-05, | |
| "loss": 0.0303, | |
| "step": 829 | |
| }, | |
| { | |
| "epoch": 4.382838283828383, | |
| "grad_norm": 1.1539040803909302, | |
| "learning_rate": 1.4901960784313726e-05, | |
| "loss": 0.0369, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 4.3881188118811885, | |
| "grad_norm": 0.9597378373146057, | |
| "learning_rate": 1.4852941176470589e-05, | |
| "loss": 0.0341, | |
| "step": 831 | |
| }, | |
| { | |
| "epoch": 4.3933993399339935, | |
| "grad_norm": 1.133381724357605, | |
| "learning_rate": 1.4803921568627453e-05, | |
| "loss": 0.032, | |
| "step": 832 | |
| }, | |
| { | |
| "epoch": 4.398679867986798, | |
| "grad_norm": 0.8584814071655273, | |
| "learning_rate": 1.4754901960784315e-05, | |
| "loss": 0.0274, | |
| "step": 833 | |
| }, | |
| { | |
| "epoch": 4.403960396039604, | |
| "grad_norm": 0.7152499556541443, | |
| "learning_rate": 1.4705882352941177e-05, | |
| "loss": 0.0228, | |
| "step": 834 | |
| }, | |
| { | |
| "epoch": 4.409240924092409, | |
| "grad_norm": 1.0020297765731812, | |
| "learning_rate": 1.465686274509804e-05, | |
| "loss": 0.0405, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 4.414521452145214, | |
| "grad_norm": 1.236913800239563, | |
| "learning_rate": 1.4607843137254903e-05, | |
| "loss": 0.0399, | |
| "step": 836 | |
| }, | |
| { | |
| "epoch": 4.41980198019802, | |
| "grad_norm": 1.0684503316879272, | |
| "learning_rate": 1.4558823529411767e-05, | |
| "loss": 0.0375, | |
| "step": 837 | |
| }, | |
| { | |
| "epoch": 4.425082508250825, | |
| "grad_norm": 1.2516707181930542, | |
| "learning_rate": 1.4509803921568629e-05, | |
| "loss": 0.0379, | |
| "step": 838 | |
| }, | |
| { | |
| "epoch": 4.43036303630363, | |
| "grad_norm": 1.063377857208252, | |
| "learning_rate": 1.4460784313725493e-05, | |
| "loss": 0.0389, | |
| "step": 839 | |
| }, | |
| { | |
| "epoch": 4.435643564356436, | |
| "grad_norm": 0.740829586982727, | |
| "learning_rate": 1.4411764705882352e-05, | |
| "loss": 0.0258, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 4.440924092409241, | |
| "grad_norm": 0.9484118223190308, | |
| "learning_rate": 1.4362745098039216e-05, | |
| "loss": 0.0339, | |
| "step": 841 | |
| }, | |
| { | |
| "epoch": 4.446204620462046, | |
| "grad_norm": 1.024625539779663, | |
| "learning_rate": 1.4313725490196078e-05, | |
| "loss": 0.0329, | |
| "step": 842 | |
| }, | |
| { | |
| "epoch": 4.451485148514852, | |
| "grad_norm": 0.8846271634101868, | |
| "learning_rate": 1.4264705882352942e-05, | |
| "loss": 0.0293, | |
| "step": 843 | |
| }, | |
| { | |
| "epoch": 4.456765676567657, | |
| "grad_norm": 0.6733059883117676, | |
| "learning_rate": 1.4215686274509804e-05, | |
| "loss": 0.0224, | |
| "step": 844 | |
| }, | |
| { | |
| "epoch": 4.462046204620462, | |
| "grad_norm": 1.042506456375122, | |
| "learning_rate": 1.4166666666666668e-05, | |
| "loss": 0.0273, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 4.467326732673268, | |
| "grad_norm": 0.9040712118148804, | |
| "learning_rate": 1.411764705882353e-05, | |
| "loss": 0.0293, | |
| "step": 846 | |
| }, | |
| { | |
| "epoch": 4.472607260726073, | |
| "grad_norm": 0.8816120028495789, | |
| "learning_rate": 1.4068627450980392e-05, | |
| "loss": 0.0264, | |
| "step": 847 | |
| }, | |
| { | |
| "epoch": 4.477887788778878, | |
| "grad_norm": 0.9503970146179199, | |
| "learning_rate": 1.4019607843137256e-05, | |
| "loss": 0.0386, | |
| "step": 848 | |
| }, | |
| { | |
| "epoch": 4.483168316831684, | |
| "grad_norm": 0.6163874268531799, | |
| "learning_rate": 1.3970588235294118e-05, | |
| "loss": 0.0179, | |
| "step": 849 | |
| }, | |
| { | |
| "epoch": 4.488448844884489, | |
| "grad_norm": 1.139955759048462, | |
| "learning_rate": 1.3921568627450982e-05, | |
| "loss": 0.0364, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 4.493729372937294, | |
| "grad_norm": 0.8715453147888184, | |
| "learning_rate": 1.3872549019607844e-05, | |
| "loss": 0.0254, | |
| "step": 851 | |
| }, | |
| { | |
| "epoch": 4.4990099009900995, | |
| "grad_norm": 0.9393417835235596, | |
| "learning_rate": 1.3823529411764708e-05, | |
| "loss": 0.033, | |
| "step": 852 | |
| }, | |
| { | |
| "epoch": 4.504290429042904, | |
| "grad_norm": 1.1215251684188843, | |
| "learning_rate": 1.377450980392157e-05, | |
| "loss": 0.0357, | |
| "step": 853 | |
| }, | |
| { | |
| "epoch": 4.509570957095709, | |
| "grad_norm": 0.9020299911499023, | |
| "learning_rate": 1.3725490196078432e-05, | |
| "loss": 0.0257, | |
| "step": 854 | |
| }, | |
| { | |
| "epoch": 4.514851485148515, | |
| "grad_norm": 1.0256519317626953, | |
| "learning_rate": 1.3676470588235296e-05, | |
| "loss": 0.0271, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 4.52013201320132, | |
| "grad_norm": 0.7362510561943054, | |
| "learning_rate": 1.3627450980392157e-05, | |
| "loss": 0.0214, | |
| "step": 856 | |
| }, | |
| { | |
| "epoch": 4.525412541254125, | |
| "grad_norm": 1.1379399299621582, | |
| "learning_rate": 1.3578431372549019e-05, | |
| "loss": 0.0337, | |
| "step": 857 | |
| }, | |
| { | |
| "epoch": 4.530693069306931, | |
| "grad_norm": 1.2522259950637817, | |
| "learning_rate": 1.3529411764705883e-05, | |
| "loss": 0.0401, | |
| "step": 858 | |
| }, | |
| { | |
| "epoch": 4.535973597359736, | |
| "grad_norm": 1.050032138824463, | |
| "learning_rate": 1.3480392156862745e-05, | |
| "loss": 0.0346, | |
| "step": 859 | |
| }, | |
| { | |
| "epoch": 4.541254125412541, | |
| "grad_norm": 0.9979643821716309, | |
| "learning_rate": 1.3431372549019607e-05, | |
| "loss": 0.0283, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 4.546534653465347, | |
| "grad_norm": 0.9979017972946167, | |
| "learning_rate": 1.3382352941176471e-05, | |
| "loss": 0.0302, | |
| "step": 861 | |
| }, | |
| { | |
| "epoch": 4.551815181518152, | |
| "grad_norm": 0.9042947292327881, | |
| "learning_rate": 1.3333333333333333e-05, | |
| "loss": 0.0299, | |
| "step": 862 | |
| }, | |
| { | |
| "epoch": 4.557095709570957, | |
| "grad_norm": 1.0963839292526245, | |
| "learning_rate": 1.3284313725490197e-05, | |
| "loss": 0.0331, | |
| "step": 863 | |
| }, | |
| { | |
| "epoch": 4.562376237623763, | |
| "grad_norm": 0.9561034440994263, | |
| "learning_rate": 1.323529411764706e-05, | |
| "loss": 0.025, | |
| "step": 864 | |
| }, | |
| { | |
| "epoch": 4.567656765676568, | |
| "grad_norm": 1.147443175315857, | |
| "learning_rate": 1.3186274509803923e-05, | |
| "loss": 0.0363, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 4.572937293729373, | |
| "grad_norm": 1.173801064491272, | |
| "learning_rate": 1.3137254901960785e-05, | |
| "loss": 0.0384, | |
| "step": 866 | |
| }, | |
| { | |
| "epoch": 4.578217821782178, | |
| "grad_norm": 1.3326085805892944, | |
| "learning_rate": 1.3088235294117648e-05, | |
| "loss": 0.0428, | |
| "step": 867 | |
| }, | |
| { | |
| "epoch": 4.583498349834984, | |
| "grad_norm": 0.8449905514717102, | |
| "learning_rate": 1.3039215686274511e-05, | |
| "loss": 0.02, | |
| "step": 868 | |
| }, | |
| { | |
| "epoch": 4.588778877887789, | |
| "grad_norm": 0.9651418924331665, | |
| "learning_rate": 1.2990196078431374e-05, | |
| "loss": 0.03, | |
| "step": 869 | |
| }, | |
| { | |
| "epoch": 4.594059405940594, | |
| "grad_norm": 0.7175216674804688, | |
| "learning_rate": 1.2941176470588238e-05, | |
| "loss": 0.0243, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 4.5993399339934, | |
| "grad_norm": 0.6259102821350098, | |
| "learning_rate": 1.28921568627451e-05, | |
| "loss": 0.0212, | |
| "step": 871 | |
| }, | |
| { | |
| "epoch": 4.604620462046205, | |
| "grad_norm": 0.7783799171447754, | |
| "learning_rate": 1.284313725490196e-05, | |
| "loss": 0.0261, | |
| "step": 872 | |
| }, | |
| { | |
| "epoch": 4.6099009900990096, | |
| "grad_norm": 1.1052485704421997, | |
| "learning_rate": 1.2794117647058822e-05, | |
| "loss": 0.0413, | |
| "step": 873 | |
| }, | |
| { | |
| "epoch": 4.615181518151815, | |
| "grad_norm": 1.0588634014129639, | |
| "learning_rate": 1.2745098039215686e-05, | |
| "loss": 0.0324, | |
| "step": 874 | |
| }, | |
| { | |
| "epoch": 4.62046204620462, | |
| "grad_norm": 0.8361983299255371, | |
| "learning_rate": 1.2696078431372548e-05, | |
| "loss": 0.0263, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 4.625742574257425, | |
| "grad_norm": 0.9796653985977173, | |
| "learning_rate": 1.2647058823529412e-05, | |
| "loss": 0.0357, | |
| "step": 876 | |
| }, | |
| { | |
| "epoch": 4.631023102310231, | |
| "grad_norm": 1.612053394317627, | |
| "learning_rate": 1.2598039215686275e-05, | |
| "loss": 0.0404, | |
| "step": 877 | |
| }, | |
| { | |
| "epoch": 4.636303630363036, | |
| "grad_norm": 0.853406548500061, | |
| "learning_rate": 1.2549019607843138e-05, | |
| "loss": 0.0266, | |
| "step": 878 | |
| }, | |
| { | |
| "epoch": 4.641584158415841, | |
| "grad_norm": 1.0649423599243164, | |
| "learning_rate": 1.25e-05, | |
| "loss": 0.0373, | |
| "step": 879 | |
| }, | |
| { | |
| "epoch": 4.646864686468647, | |
| "grad_norm": 1.1256874799728394, | |
| "learning_rate": 1.2450980392156863e-05, | |
| "loss": 0.0251, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 4.652145214521452, | |
| "grad_norm": 0.8546039462089539, | |
| "learning_rate": 1.2401960784313727e-05, | |
| "loss": 0.0272, | |
| "step": 881 | |
| }, | |
| { | |
| "epoch": 4.657425742574257, | |
| "grad_norm": 0.7499862313270569, | |
| "learning_rate": 1.2352941176470589e-05, | |
| "loss": 0.0227, | |
| "step": 882 | |
| }, | |
| { | |
| "epoch": 4.662706270627063, | |
| "grad_norm": 1.2291607856750488, | |
| "learning_rate": 1.2303921568627451e-05, | |
| "loss": 0.0416, | |
| "step": 883 | |
| }, | |
| { | |
| "epoch": 4.667986798679868, | |
| "grad_norm": 0.919152021408081, | |
| "learning_rate": 1.2254901960784313e-05, | |
| "loss": 0.0325, | |
| "step": 884 | |
| }, | |
| { | |
| "epoch": 4.673267326732673, | |
| "grad_norm": 0.8878404498100281, | |
| "learning_rate": 1.2205882352941177e-05, | |
| "loss": 0.0296, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 4.678547854785479, | |
| "grad_norm": 1.1350431442260742, | |
| "learning_rate": 1.215686274509804e-05, | |
| "loss": 0.0372, | |
| "step": 886 | |
| }, | |
| { | |
| "epoch": 4.683828382838284, | |
| "grad_norm": 0.9311625957489014, | |
| "learning_rate": 1.2107843137254903e-05, | |
| "loss": 0.0324, | |
| "step": 887 | |
| }, | |
| { | |
| "epoch": 4.689108910891089, | |
| "grad_norm": 1.2680948972702026, | |
| "learning_rate": 1.2058823529411765e-05, | |
| "loss": 0.0427, | |
| "step": 888 | |
| }, | |
| { | |
| "epoch": 4.694389438943895, | |
| "grad_norm": 1.0892651081085205, | |
| "learning_rate": 1.200980392156863e-05, | |
| "loss": 0.0289, | |
| "step": 889 | |
| }, | |
| { | |
| "epoch": 4.6996699669967, | |
| "grad_norm": 0.9546079635620117, | |
| "learning_rate": 1.1960784313725491e-05, | |
| "loss": 0.0355, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 4.704950495049505, | |
| "grad_norm": 1.0489920377731323, | |
| "learning_rate": 1.1911764705882354e-05, | |
| "loss": 0.0287, | |
| "step": 891 | |
| }, | |
| { | |
| "epoch": 4.710231023102311, | |
| "grad_norm": 0.8627532720565796, | |
| "learning_rate": 1.1862745098039216e-05, | |
| "loss": 0.03, | |
| "step": 892 | |
| }, | |
| { | |
| "epoch": 4.7155115511551156, | |
| "grad_norm": 1.1037335395812988, | |
| "learning_rate": 1.1813725490196078e-05, | |
| "loss": 0.0369, | |
| "step": 893 | |
| }, | |
| { | |
| "epoch": 4.7207920792079205, | |
| "grad_norm": 0.8661030530929565, | |
| "learning_rate": 1.1764705882352942e-05, | |
| "loss": 0.0267, | |
| "step": 894 | |
| }, | |
| { | |
| "epoch": 4.726072607260726, | |
| "grad_norm": 0.7932984828948975, | |
| "learning_rate": 1.1715686274509804e-05, | |
| "loss": 0.0254, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 4.731353135313531, | |
| "grad_norm": 1.1760293245315552, | |
| "learning_rate": 1.1666666666666668e-05, | |
| "loss": 0.0408, | |
| "step": 896 | |
| }, | |
| { | |
| "epoch": 4.736633663366336, | |
| "grad_norm": 1.098027229309082, | |
| "learning_rate": 1.161764705882353e-05, | |
| "loss": 0.0304, | |
| "step": 897 | |
| }, | |
| { | |
| "epoch": 4.741914191419142, | |
| "grad_norm": 1.1287803649902344, | |
| "learning_rate": 1.1568627450980394e-05, | |
| "loss": 0.0305, | |
| "step": 898 | |
| }, | |
| { | |
| "epoch": 4.747194719471947, | |
| "grad_norm": 1.1614177227020264, | |
| "learning_rate": 1.1519607843137254e-05, | |
| "loss": 0.0436, | |
| "step": 899 | |
| }, | |
| { | |
| "epoch": 4.752475247524752, | |
| "grad_norm": 0.8995096683502197, | |
| "learning_rate": 1.1470588235294118e-05, | |
| "loss": 0.027, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 4.757755775577558, | |
| "grad_norm": 1.0893324613571167, | |
| "learning_rate": 1.142156862745098e-05, | |
| "loss": 0.0304, | |
| "step": 901 | |
| }, | |
| { | |
| "epoch": 4.763036303630363, | |
| "grad_norm": 0.8086807131767273, | |
| "learning_rate": 1.1372549019607843e-05, | |
| "loss": 0.0304, | |
| "step": 902 | |
| }, | |
| { | |
| "epoch": 4.768316831683168, | |
| "grad_norm": 0.6953551769256592, | |
| "learning_rate": 1.1323529411764707e-05, | |
| "loss": 0.0248, | |
| "step": 903 | |
| }, | |
| { | |
| "epoch": 4.773597359735973, | |
| "grad_norm": 0.707534670829773, | |
| "learning_rate": 1.1274509803921569e-05, | |
| "loss": 0.0219, | |
| "step": 904 | |
| }, | |
| { | |
| "epoch": 4.778877887788779, | |
| "grad_norm": 0.9743644595146179, | |
| "learning_rate": 1.1225490196078433e-05, | |
| "loss": 0.0327, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 4.784158415841584, | |
| "grad_norm": 0.9171704053878784, | |
| "learning_rate": 1.1176470588235295e-05, | |
| "loss": 0.0277, | |
| "step": 906 | |
| }, | |
| { | |
| "epoch": 4.789438943894389, | |
| "grad_norm": 1.019524097442627, | |
| "learning_rate": 1.1127450980392157e-05, | |
| "loss": 0.0308, | |
| "step": 907 | |
| }, | |
| { | |
| "epoch": 4.794719471947195, | |
| "grad_norm": 0.8753800392150879, | |
| "learning_rate": 1.107843137254902e-05, | |
| "loss": 0.0282, | |
| "step": 908 | |
| }, | |
| { | |
| "epoch": 4.8, | |
| "grad_norm": 1.0594887733459473, | |
| "learning_rate": 1.1029411764705883e-05, | |
| "loss": 0.0283, | |
| "step": 909 | |
| }, | |
| { | |
| "epoch": 4.805280528052805, | |
| "grad_norm": 0.9229967594146729, | |
| "learning_rate": 1.0980392156862745e-05, | |
| "loss": 0.0298, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 4.810561056105611, | |
| "grad_norm": 0.8133540153503418, | |
| "learning_rate": 1.0931372549019609e-05, | |
| "loss": 0.0276, | |
| "step": 911 | |
| }, | |
| { | |
| "epoch": 4.815841584158416, | |
| "grad_norm": 0.8808728456497192, | |
| "learning_rate": 1.0882352941176471e-05, | |
| "loss": 0.0304, | |
| "step": 912 | |
| }, | |
| { | |
| "epoch": 4.821122112211221, | |
| "grad_norm": 1.0457746982574463, | |
| "learning_rate": 1.0833333333333334e-05, | |
| "loss": 0.0313, | |
| "step": 913 | |
| }, | |
| { | |
| "epoch": 4.8264026402640265, | |
| "grad_norm": 1.004420518875122, | |
| "learning_rate": 1.0784313725490197e-05, | |
| "loss": 0.0341, | |
| "step": 914 | |
| }, | |
| { | |
| "epoch": 4.8316831683168315, | |
| "grad_norm": 0.8167664408683777, | |
| "learning_rate": 1.0735294117647058e-05, | |
| "loss": 0.0254, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 4.8369636963696365, | |
| "grad_norm": 0.591806948184967, | |
| "learning_rate": 1.0686274509803922e-05, | |
| "loss": 0.0191, | |
| "step": 916 | |
| }, | |
| { | |
| "epoch": 4.842244224422442, | |
| "grad_norm": 1.1454893350601196, | |
| "learning_rate": 1.0637254901960784e-05, | |
| "loss": 0.0394, | |
| "step": 917 | |
| }, | |
| { | |
| "epoch": 4.847524752475247, | |
| "grad_norm": 0.9010074734687805, | |
| "learning_rate": 1.0588235294117648e-05, | |
| "loss": 0.0287, | |
| "step": 918 | |
| }, | |
| { | |
| "epoch": 4.852805280528052, | |
| "grad_norm": 0.9918347597122192, | |
| "learning_rate": 1.053921568627451e-05, | |
| "loss": 0.0288, | |
| "step": 919 | |
| }, | |
| { | |
| "epoch": 4.858085808580858, | |
| "grad_norm": 0.9720319509506226, | |
| "learning_rate": 1.0490196078431374e-05, | |
| "loss": 0.0356, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 4.863366336633663, | |
| "grad_norm": 0.8841050267219543, | |
| "learning_rate": 1.0441176470588236e-05, | |
| "loss": 0.0301, | |
| "step": 921 | |
| }, | |
| { | |
| "epoch": 4.868646864686468, | |
| "grad_norm": 1.0269768238067627, | |
| "learning_rate": 1.0392156862745098e-05, | |
| "loss": 0.0315, | |
| "step": 922 | |
| }, | |
| { | |
| "epoch": 4.873927392739274, | |
| "grad_norm": 0.7187484502792358, | |
| "learning_rate": 1.034313725490196e-05, | |
| "loss": 0.0198, | |
| "step": 923 | |
| }, | |
| { | |
| "epoch": 4.879207920792079, | |
| "grad_norm": 0.8568077087402344, | |
| "learning_rate": 1.0294117647058824e-05, | |
| "loss": 0.0298, | |
| "step": 924 | |
| }, | |
| { | |
| "epoch": 4.884488448844884, | |
| "grad_norm": 1.0901175737380981, | |
| "learning_rate": 1.0245098039215687e-05, | |
| "loss": 0.0321, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 4.88976897689769, | |
| "grad_norm": 0.7445207834243774, | |
| "learning_rate": 1.0196078431372549e-05, | |
| "loss": 0.0328, | |
| "step": 926 | |
| }, | |
| { | |
| "epoch": 4.895049504950495, | |
| "grad_norm": 1.0356990098953247, | |
| "learning_rate": 1.0147058823529413e-05, | |
| "loss": 0.0301, | |
| "step": 927 | |
| }, | |
| { | |
| "epoch": 4.9003300330033, | |
| "grad_norm": 1.0626378059387207, | |
| "learning_rate": 1.0098039215686275e-05, | |
| "loss": 0.0317, | |
| "step": 928 | |
| }, | |
| { | |
| "epoch": 4.905610561056106, | |
| "grad_norm": 0.9375091195106506, | |
| "learning_rate": 1.0049019607843139e-05, | |
| "loss": 0.0306, | |
| "step": 929 | |
| }, | |
| { | |
| "epoch": 4.910891089108911, | |
| "grad_norm": 0.7572767734527588, | |
| "learning_rate": 1e-05, | |
| "loss": 0.0233, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 4.916171617161716, | |
| "grad_norm": 0.7957236170768738, | |
| "learning_rate": 9.950980392156863e-06, | |
| "loss": 0.0286, | |
| "step": 931 | |
| }, | |
| { | |
| "epoch": 4.921452145214522, | |
| "grad_norm": 0.723996639251709, | |
| "learning_rate": 9.901960784313725e-06, | |
| "loss": 0.0221, | |
| "step": 932 | |
| }, | |
| { | |
| "epoch": 4.926732673267327, | |
| "grad_norm": 0.7637007236480713, | |
| "learning_rate": 9.852941176470589e-06, | |
| "loss": 0.0222, | |
| "step": 933 | |
| }, | |
| { | |
| "epoch": 4.932013201320132, | |
| "grad_norm": 1.145386815071106, | |
| "learning_rate": 9.803921568627451e-06, | |
| "loss": 0.0401, | |
| "step": 934 | |
| }, | |
| { | |
| "epoch": 4.9372937293729375, | |
| "grad_norm": 1.1786422729492188, | |
| "learning_rate": 9.754901960784313e-06, | |
| "loss": 0.04, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 4.9425742574257425, | |
| "grad_norm": 0.881608784198761, | |
| "learning_rate": 9.705882352941177e-06, | |
| "loss": 0.0292, | |
| "step": 936 | |
| }, | |
| { | |
| "epoch": 4.9478547854785475, | |
| "grad_norm": 1.056412935256958, | |
| "learning_rate": 9.65686274509804e-06, | |
| "loss": 0.0325, | |
| "step": 937 | |
| }, | |
| { | |
| "epoch": 4.953135313531353, | |
| "grad_norm": 1.0825217962265015, | |
| "learning_rate": 9.607843137254903e-06, | |
| "loss": 0.039, | |
| "step": 938 | |
| }, | |
| { | |
| "epoch": 4.958415841584158, | |
| "grad_norm": 0.6380865573883057, | |
| "learning_rate": 9.558823529411764e-06, | |
| "loss": 0.0231, | |
| "step": 939 | |
| }, | |
| { | |
| "epoch": 4.963696369636963, | |
| "grad_norm": 0.8444051742553711, | |
| "learning_rate": 9.509803921568628e-06, | |
| "loss": 0.0284, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 4.968976897689769, | |
| "grad_norm": 0.8642618656158447, | |
| "learning_rate": 9.46078431372549e-06, | |
| "loss": 0.0265, | |
| "step": 941 | |
| }, | |
| { | |
| "epoch": 4.974257425742574, | |
| "grad_norm": 1.0572503805160522, | |
| "learning_rate": 9.411764705882354e-06, | |
| "loss": 0.0341, | |
| "step": 942 | |
| }, | |
| { | |
| "epoch": 4.979537953795379, | |
| "grad_norm": 0.8778902888298035, | |
| "learning_rate": 9.362745098039216e-06, | |
| "loss": 0.0304, | |
| "step": 943 | |
| }, | |
| { | |
| "epoch": 4.984818481848185, | |
| "grad_norm": 0.8647822737693787, | |
| "learning_rate": 9.31372549019608e-06, | |
| "loss": 0.0289, | |
| "step": 944 | |
| }, | |
| { | |
| "epoch": 4.99009900990099, | |
| "grad_norm": 1.0456162691116333, | |
| "learning_rate": 9.264705882352942e-06, | |
| "loss": 0.0322, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 4.995379537953795, | |
| "grad_norm": 1.451906681060791, | |
| "learning_rate": 9.215686274509804e-06, | |
| "loss": 0.0414, | |
| "step": 946 | |
| }, | |
| { | |
| "epoch": 5.000660066006601, | |
| "grad_norm": 1.2711623907089233, | |
| "learning_rate": 9.166666666666666e-06, | |
| "loss": 0.0378, | |
| "step": 947 | |
| }, | |
| { | |
| "epoch": 5.005940594059406, | |
| "grad_norm": 0.5490008592605591, | |
| "learning_rate": 9.117647058823529e-06, | |
| "loss": 0.0236, | |
| "step": 948 | |
| }, | |
| { | |
| "epoch": 5.011221122112211, | |
| "grad_norm": 0.4675862789154053, | |
| "learning_rate": 9.068627450980392e-06, | |
| "loss": 0.0193, | |
| "step": 949 | |
| }, | |
| { | |
| "epoch": 5.016501650165017, | |
| "grad_norm": 0.5535048246383667, | |
| "learning_rate": 9.019607843137255e-06, | |
| "loss": 0.0198, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 5.021782178217822, | |
| "grad_norm": 0.759076714515686, | |
| "learning_rate": 8.970588235294119e-06, | |
| "loss": 0.024, | |
| "step": 951 | |
| }, | |
| { | |
| "epoch": 5.027062706270627, | |
| "grad_norm": 0.551156222820282, | |
| "learning_rate": 8.92156862745098e-06, | |
| "loss": 0.0182, | |
| "step": 952 | |
| }, | |
| { | |
| "epoch": 5.032343234323433, | |
| "grad_norm": 0.561464786529541, | |
| "learning_rate": 8.872549019607845e-06, | |
| "loss": 0.0211, | |
| "step": 953 | |
| }, | |
| { | |
| "epoch": 5.037623762376238, | |
| "grad_norm": 0.3056110441684723, | |
| "learning_rate": 8.823529411764707e-06, | |
| "loss": 0.0146, | |
| "step": 954 | |
| }, | |
| { | |
| "epoch": 5.042904290429043, | |
| "grad_norm": 0.6087129712104797, | |
| "learning_rate": 8.774509803921569e-06, | |
| "loss": 0.0186, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 5.0481848184818485, | |
| "grad_norm": 0.37718823552131653, | |
| "learning_rate": 8.725490196078431e-06, | |
| "loss": 0.0157, | |
| "step": 956 | |
| }, | |
| { | |
| "epoch": 5.0534653465346535, | |
| "grad_norm": 0.5187584161758423, | |
| "learning_rate": 8.676470588235295e-06, | |
| "loss": 0.0206, | |
| "step": 957 | |
| }, | |
| { | |
| "epoch": 5.0587458745874585, | |
| "grad_norm": 0.4911420941352844, | |
| "learning_rate": 8.627450980392157e-06, | |
| "loss": 0.0198, | |
| "step": 958 | |
| }, | |
| { | |
| "epoch": 5.064026402640264, | |
| "grad_norm": 0.5944388508796692, | |
| "learning_rate": 8.57843137254902e-06, | |
| "loss": 0.0199, | |
| "step": 959 | |
| }, | |
| { | |
| "epoch": 5.069306930693069, | |
| "grad_norm": 0.4171479344367981, | |
| "learning_rate": 8.529411764705883e-06, | |
| "loss": 0.018, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 5.074587458745874, | |
| "grad_norm": 0.4937235414981842, | |
| "learning_rate": 8.480392156862745e-06, | |
| "loss": 0.0182, | |
| "step": 961 | |
| }, | |
| { | |
| "epoch": 5.07986798679868, | |
| "grad_norm": 0.44039833545684814, | |
| "learning_rate": 8.43137254901961e-06, | |
| "loss": 0.0166, | |
| "step": 962 | |
| }, | |
| { | |
| "epoch": 5.085148514851485, | |
| "grad_norm": 0.5266954302787781, | |
| "learning_rate": 8.38235294117647e-06, | |
| "loss": 0.0199, | |
| "step": 963 | |
| }, | |
| { | |
| "epoch": 5.09042904290429, | |
| "grad_norm": 0.5795379877090454, | |
| "learning_rate": 8.333333333333334e-06, | |
| "loss": 0.0168, | |
| "step": 964 | |
| }, | |
| { | |
| "epoch": 5.095709570957096, | |
| "grad_norm": 0.7336843013763428, | |
| "learning_rate": 8.284313725490196e-06, | |
| "loss": 0.0265, | |
| "step": 965 | |
| }, | |
| { | |
| "epoch": 5.100990099009901, | |
| "grad_norm": 0.4279429018497467, | |
| "learning_rate": 8.23529411764706e-06, | |
| "loss": 0.0177, | |
| "step": 966 | |
| }, | |
| { | |
| "epoch": 5.106270627062706, | |
| "grad_norm": 0.6605322957038879, | |
| "learning_rate": 8.186274509803922e-06, | |
| "loss": 0.0211, | |
| "step": 967 | |
| }, | |
| { | |
| "epoch": 5.111551155115512, | |
| "grad_norm": 0.6620057225227356, | |
| "learning_rate": 8.137254901960784e-06, | |
| "loss": 0.0186, | |
| "step": 968 | |
| }, | |
| { | |
| "epoch": 5.116831683168317, | |
| "grad_norm": 0.36124753952026367, | |
| "learning_rate": 8.088235294117648e-06, | |
| "loss": 0.015, | |
| "step": 969 | |
| }, | |
| { | |
| "epoch": 5.122112211221122, | |
| "grad_norm": 0.64441978931427, | |
| "learning_rate": 8.03921568627451e-06, | |
| "loss": 0.0176, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 5.127392739273928, | |
| "grad_norm": 0.45318105816841125, | |
| "learning_rate": 7.990196078431372e-06, | |
| "loss": 0.0155, | |
| "step": 971 | |
| }, | |
| { | |
| "epoch": 5.132673267326733, | |
| "grad_norm": 0.528228223323822, | |
| "learning_rate": 7.941176470588235e-06, | |
| "loss": 0.0169, | |
| "step": 972 | |
| }, | |
| { | |
| "epoch": 5.137953795379538, | |
| "grad_norm": 0.5010206699371338, | |
| "learning_rate": 7.892156862745098e-06, | |
| "loss": 0.018, | |
| "step": 973 | |
| }, | |
| { | |
| "epoch": 5.143234323432344, | |
| "grad_norm": 0.4574146568775177, | |
| "learning_rate": 7.84313725490196e-06, | |
| "loss": 0.0172, | |
| "step": 974 | |
| }, | |
| { | |
| "epoch": 5.148514851485149, | |
| "grad_norm": 0.5487357974052429, | |
| "learning_rate": 7.794117647058825e-06, | |
| "loss": 0.0173, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 5.153795379537954, | |
| "grad_norm": 0.7028687596321106, | |
| "learning_rate": 7.745098039215687e-06, | |
| "loss": 0.0219, | |
| "step": 976 | |
| }, | |
| { | |
| "epoch": 5.1590759075907595, | |
| "grad_norm": 0.5689717531204224, | |
| "learning_rate": 7.69607843137255e-06, | |
| "loss": 0.0167, | |
| "step": 977 | |
| }, | |
| { | |
| "epoch": 5.1643564356435645, | |
| "grad_norm": 0.4024108052253723, | |
| "learning_rate": 7.647058823529413e-06, | |
| "loss": 0.0139, | |
| "step": 978 | |
| }, | |
| { | |
| "epoch": 5.1696369636963695, | |
| "grad_norm": 0.81581050157547, | |
| "learning_rate": 7.598039215686274e-06, | |
| "loss": 0.0156, | |
| "step": 979 | |
| }, | |
| { | |
| "epoch": 5.174917491749175, | |
| "grad_norm": 0.4765579104423523, | |
| "learning_rate": 7.549019607843137e-06, | |
| "loss": 0.0155, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 5.18019801980198, | |
| "grad_norm": 0.43915316462516785, | |
| "learning_rate": 7.5e-06, | |
| "loss": 0.0188, | |
| "step": 981 | |
| }, | |
| { | |
| "epoch": 5.185478547854785, | |
| "grad_norm": 0.5227847099304199, | |
| "learning_rate": 7.450980392156863e-06, | |
| "loss": 0.0206, | |
| "step": 982 | |
| }, | |
| { | |
| "epoch": 5.19075907590759, | |
| "grad_norm": 0.6932447552680969, | |
| "learning_rate": 7.401960784313726e-06, | |
| "loss": 0.0215, | |
| "step": 983 | |
| }, | |
| { | |
| "epoch": 5.196039603960396, | |
| "grad_norm": 0.6111219525337219, | |
| "learning_rate": 7.3529411764705884e-06, | |
| "loss": 0.0212, | |
| "step": 984 | |
| }, | |
| { | |
| "epoch": 5.201320132013201, | |
| "grad_norm": 0.4277690649032593, | |
| "learning_rate": 7.3039215686274515e-06, | |
| "loss": 0.0159, | |
| "step": 985 | |
| }, | |
| { | |
| "epoch": 5.206600660066006, | |
| "grad_norm": 0.47906622290611267, | |
| "learning_rate": 7.2549019607843145e-06, | |
| "loss": 0.0148, | |
| "step": 986 | |
| }, | |
| { | |
| "epoch": 5.211881188118812, | |
| "grad_norm": 0.565126359462738, | |
| "learning_rate": 7.205882352941176e-06, | |
| "loss": 0.0195, | |
| "step": 987 | |
| }, | |
| { | |
| "epoch": 5.217161716171617, | |
| "grad_norm": 0.46082907915115356, | |
| "learning_rate": 7.156862745098039e-06, | |
| "loss": 0.0183, | |
| "step": 988 | |
| }, | |
| { | |
| "epoch": 5.222442244224422, | |
| "grad_norm": 0.6407860517501831, | |
| "learning_rate": 7.107843137254902e-06, | |
| "loss": 0.0186, | |
| "step": 989 | |
| }, | |
| { | |
| "epoch": 5.227722772277228, | |
| "grad_norm": 0.4975033104419708, | |
| "learning_rate": 7.058823529411765e-06, | |
| "loss": 0.018, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 5.233003300330033, | |
| "grad_norm": 0.8421338200569153, | |
| "learning_rate": 7.009803921568628e-06, | |
| "loss": 0.0236, | |
| "step": 991 | |
| }, | |
| { | |
| "epoch": 5.238283828382838, | |
| "grad_norm": 0.5029832124710083, | |
| "learning_rate": 6.960784313725491e-06, | |
| "loss": 0.017, | |
| "step": 992 | |
| }, | |
| { | |
| "epoch": 5.243564356435644, | |
| "grad_norm": 0.6490949988365173, | |
| "learning_rate": 6.911764705882354e-06, | |
| "loss": 0.0225, | |
| "step": 993 | |
| }, | |
| { | |
| "epoch": 5.248844884488449, | |
| "grad_norm": 0.4616677165031433, | |
| "learning_rate": 6.862745098039216e-06, | |
| "loss": 0.0141, | |
| "step": 994 | |
| }, | |
| { | |
| "epoch": 5.254125412541254, | |
| "grad_norm": 0.5079948306083679, | |
| "learning_rate": 6.813725490196078e-06, | |
| "loss": 0.0201, | |
| "step": 995 | |
| }, | |
| { | |
| "epoch": 5.25940594059406, | |
| "grad_norm": 0.6577730178833008, | |
| "learning_rate": 6.7647058823529414e-06, | |
| "loss": 0.0166, | |
| "step": 996 | |
| }, | |
| { | |
| "epoch": 5.264686468646865, | |
| "grad_norm": 0.4667012691497803, | |
| "learning_rate": 6.715686274509804e-06, | |
| "loss": 0.0142, | |
| "step": 997 | |
| }, | |
| { | |
| "epoch": 5.26996699669967, | |
| "grad_norm": 0.4957946836948395, | |
| "learning_rate": 6.666666666666667e-06, | |
| "loss": 0.0183, | |
| "step": 998 | |
| }, | |
| { | |
| "epoch": 5.2752475247524755, | |
| "grad_norm": 0.5439987182617188, | |
| "learning_rate": 6.61764705882353e-06, | |
| "loss": 0.0203, | |
| "step": 999 | |
| }, | |
| { | |
| "epoch": 5.2805280528052805, | |
| "grad_norm": 0.6871292591094971, | |
| "learning_rate": 6.568627450980393e-06, | |
| "loss": 0.0187, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 5.2858085808580855, | |
| "grad_norm": 1.2314088344573975, | |
| "learning_rate": 6.519607843137256e-06, | |
| "loss": 0.0218, | |
| "step": 1001 | |
| }, | |
| { | |
| "epoch": 5.291089108910891, | |
| "grad_norm": 0.9941577911376953, | |
| "learning_rate": 6.470588235294119e-06, | |
| "loss": 0.0231, | |
| "step": 1002 | |
| }, | |
| { | |
| "epoch": 5.296369636963696, | |
| "grad_norm": 0.5083587765693665, | |
| "learning_rate": 6.42156862745098e-06, | |
| "loss": 0.0175, | |
| "step": 1003 | |
| }, | |
| { | |
| "epoch": 5.301650165016501, | |
| "grad_norm": 0.548037052154541, | |
| "learning_rate": 6.372549019607843e-06, | |
| "loss": 0.019, | |
| "step": 1004 | |
| }, | |
| { | |
| "epoch": 5.306930693069307, | |
| "grad_norm": 0.5824829339981079, | |
| "learning_rate": 6.323529411764706e-06, | |
| "loss": 0.0229, | |
| "step": 1005 | |
| }, | |
| { | |
| "epoch": 5.312211221122112, | |
| "grad_norm": 0.40190955996513367, | |
| "learning_rate": 6.274509803921569e-06, | |
| "loss": 0.0177, | |
| "step": 1006 | |
| }, | |
| { | |
| "epoch": 5.317491749174917, | |
| "grad_norm": 0.5402151346206665, | |
| "learning_rate": 6.225490196078431e-06, | |
| "loss": 0.0204, | |
| "step": 1007 | |
| }, | |
| { | |
| "epoch": 5.322772277227723, | |
| "grad_norm": 0.5250375866889954, | |
| "learning_rate": 6.1764705882352944e-06, | |
| "loss": 0.0179, | |
| "step": 1008 | |
| }, | |
| { | |
| "epoch": 5.328052805280528, | |
| "grad_norm": 0.823615550994873, | |
| "learning_rate": 6.127450980392157e-06, | |
| "loss": 0.018, | |
| "step": 1009 | |
| }, | |
| { | |
| "epoch": 5.333333333333333, | |
| "grad_norm": 0.5749222636222839, | |
| "learning_rate": 6.07843137254902e-06, | |
| "loss": 0.0179, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 5.338613861386139, | |
| "grad_norm": 0.44801655411720276, | |
| "learning_rate": 6.029411764705883e-06, | |
| "loss": 0.0196, | |
| "step": 1011 | |
| }, | |
| { | |
| "epoch": 5.343894389438944, | |
| "grad_norm": 0.41333383321762085, | |
| "learning_rate": 5.980392156862746e-06, | |
| "loss": 0.0174, | |
| "step": 1012 | |
| }, | |
| { | |
| "epoch": 5.349174917491749, | |
| "grad_norm": 0.6496670842170715, | |
| "learning_rate": 5.931372549019608e-06, | |
| "loss": 0.0185, | |
| "step": 1013 | |
| }, | |
| { | |
| "epoch": 5.354455445544555, | |
| "grad_norm": 0.5372949838638306, | |
| "learning_rate": 5.882352941176471e-06, | |
| "loss": 0.0196, | |
| "step": 1014 | |
| }, | |
| { | |
| "epoch": 5.35973597359736, | |
| "grad_norm": 0.6921015381813049, | |
| "learning_rate": 5.833333333333334e-06, | |
| "loss": 0.0208, | |
| "step": 1015 | |
| }, | |
| { | |
| "epoch": 5.365016501650165, | |
| "grad_norm": 0.40831905603408813, | |
| "learning_rate": 5.784313725490197e-06, | |
| "loss": 0.0156, | |
| "step": 1016 | |
| }, | |
| { | |
| "epoch": 5.370297029702971, | |
| "grad_norm": 0.746163010597229, | |
| "learning_rate": 5.735294117647059e-06, | |
| "loss": 0.0247, | |
| "step": 1017 | |
| }, | |
| { | |
| "epoch": 5.375577557755776, | |
| "grad_norm": 0.3688511252403259, | |
| "learning_rate": 5.686274509803921e-06, | |
| "loss": 0.0149, | |
| "step": 1018 | |
| }, | |
| { | |
| "epoch": 5.380858085808581, | |
| "grad_norm": 0.6566423773765564, | |
| "learning_rate": 5.637254901960784e-06, | |
| "loss": 0.0207, | |
| "step": 1019 | |
| }, | |
| { | |
| "epoch": 5.3861386138613865, | |
| "grad_norm": 0.5603022575378418, | |
| "learning_rate": 5.588235294117647e-06, | |
| "loss": 0.015, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 5.3914191419141915, | |
| "grad_norm": 0.48195961117744446, | |
| "learning_rate": 5.53921568627451e-06, | |
| "loss": 0.0196, | |
| "step": 1021 | |
| }, | |
| { | |
| "epoch": 5.3966996699669965, | |
| "grad_norm": 0.8574143052101135, | |
| "learning_rate": 5.490196078431373e-06, | |
| "loss": 0.0213, | |
| "step": 1022 | |
| }, | |
| { | |
| "epoch": 5.401980198019802, | |
| "grad_norm": 0.6029947996139526, | |
| "learning_rate": 5.441176470588236e-06, | |
| "loss": 0.0157, | |
| "step": 1023 | |
| }, | |
| { | |
| "epoch": 5.407260726072607, | |
| "grad_norm": 0.6817245483398438, | |
| "learning_rate": 5.392156862745099e-06, | |
| "loss": 0.021, | |
| "step": 1024 | |
| }, | |
| { | |
| "epoch": 5.412541254125412, | |
| "grad_norm": 0.5209661722183228, | |
| "learning_rate": 5.343137254901961e-06, | |
| "loss": 0.0174, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 5.417821782178218, | |
| "grad_norm": 0.637236475944519, | |
| "learning_rate": 5.294117647058824e-06, | |
| "loss": 0.0227, | |
| "step": 1026 | |
| }, | |
| { | |
| "epoch": 5.423102310231023, | |
| "grad_norm": 0.4161434471607208, | |
| "learning_rate": 5.245098039215687e-06, | |
| "loss": 0.0186, | |
| "step": 1027 | |
| }, | |
| { | |
| "epoch": 5.428382838283828, | |
| "grad_norm": 0.4232980012893677, | |
| "learning_rate": 5.196078431372549e-06, | |
| "loss": 0.0157, | |
| "step": 1028 | |
| }, | |
| { | |
| "epoch": 5.433663366336634, | |
| "grad_norm": 0.8480959534645081, | |
| "learning_rate": 5.147058823529412e-06, | |
| "loss": 0.0236, | |
| "step": 1029 | |
| }, | |
| { | |
| "epoch": 5.438943894389439, | |
| "grad_norm": 0.7071532011032104, | |
| "learning_rate": 5.098039215686274e-06, | |
| "loss": 0.0232, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 5.444224422442244, | |
| "grad_norm": 0.46706661581993103, | |
| "learning_rate": 5.049019607843137e-06, | |
| "loss": 0.0154, | |
| "step": 1031 | |
| }, | |
| { | |
| "epoch": 5.44950495049505, | |
| "grad_norm": 0.38799214363098145, | |
| "learning_rate": 5e-06, | |
| "loss": 0.0153, | |
| "step": 1032 | |
| }, | |
| { | |
| "epoch": 5.454785478547855, | |
| "grad_norm": 0.6109775900840759, | |
| "learning_rate": 4.950980392156863e-06, | |
| "loss": 0.0185, | |
| "step": 1033 | |
| }, | |
| { | |
| "epoch": 5.46006600660066, | |
| "grad_norm": 0.4341484308242798, | |
| "learning_rate": 4.901960784313726e-06, | |
| "loss": 0.0167, | |
| "step": 1034 | |
| }, | |
| { | |
| "epoch": 5.465346534653466, | |
| "grad_norm": 0.7449372410774231, | |
| "learning_rate": 4.852941176470589e-06, | |
| "loss": 0.0198, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 5.470627062706271, | |
| "grad_norm": 0.36592140793800354, | |
| "learning_rate": 4.803921568627452e-06, | |
| "loss": 0.0157, | |
| "step": 1036 | |
| }, | |
| { | |
| "epoch": 5.475907590759076, | |
| "grad_norm": 0.5300789475440979, | |
| "learning_rate": 4.754901960784314e-06, | |
| "loss": 0.018, | |
| "step": 1037 | |
| }, | |
| { | |
| "epoch": 5.481188118811881, | |
| "grad_norm": 0.5365732908248901, | |
| "learning_rate": 4.705882352941177e-06, | |
| "loss": 0.0192, | |
| "step": 1038 | |
| }, | |
| { | |
| "epoch": 5.486468646864687, | |
| "grad_norm": 0.5415321588516235, | |
| "learning_rate": 4.65686274509804e-06, | |
| "loss": 0.0176, | |
| "step": 1039 | |
| }, | |
| { | |
| "epoch": 5.491749174917492, | |
| "grad_norm": 0.4487341344356537, | |
| "learning_rate": 4.607843137254902e-06, | |
| "loss": 0.0149, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 5.497029702970297, | |
| "grad_norm": 0.5181805491447449, | |
| "learning_rate": 4.558823529411764e-06, | |
| "loss": 0.0167, | |
| "step": 1041 | |
| }, | |
| { | |
| "epoch": 5.5023102310231025, | |
| "grad_norm": 0.7573541402816772, | |
| "learning_rate": 4.509803921568627e-06, | |
| "loss": 0.0192, | |
| "step": 1042 | |
| }, | |
| { | |
| "epoch": 5.5075907590759074, | |
| "grad_norm": 2.3085265159606934, | |
| "learning_rate": 4.46078431372549e-06, | |
| "loss": 0.0912, | |
| "step": 1043 | |
| }, | |
| { | |
| "epoch": 5.512871287128712, | |
| "grad_norm": 0.5817523002624512, | |
| "learning_rate": 4.411764705882353e-06, | |
| "loss": 0.015, | |
| "step": 1044 | |
| }, | |
| { | |
| "epoch": 5.518151815181518, | |
| "grad_norm": 0.6805261969566345, | |
| "learning_rate": 4.362745098039216e-06, | |
| "loss": 0.0215, | |
| "step": 1045 | |
| }, | |
| { | |
| "epoch": 5.523432343234323, | |
| "grad_norm": 0.597396969795227, | |
| "learning_rate": 4.313725490196079e-06, | |
| "loss": 0.0206, | |
| "step": 1046 | |
| }, | |
| { | |
| "epoch": 5.528712871287128, | |
| "grad_norm": 0.934479296207428, | |
| "learning_rate": 4.264705882352942e-06, | |
| "loss": 0.0207, | |
| "step": 1047 | |
| }, | |
| { | |
| "epoch": 5.533993399339934, | |
| "grad_norm": 0.4807792603969574, | |
| "learning_rate": 4.215686274509805e-06, | |
| "loss": 0.0158, | |
| "step": 1048 | |
| }, | |
| { | |
| "epoch": 5.539273927392739, | |
| "grad_norm": 0.6328734159469604, | |
| "learning_rate": 4.166666666666667e-06, | |
| "loss": 0.0202, | |
| "step": 1049 | |
| }, | |
| { | |
| "epoch": 5.544554455445544, | |
| "grad_norm": 0.4848245084285736, | |
| "learning_rate": 4.11764705882353e-06, | |
| "loss": 0.0199, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 5.54983498349835, | |
| "grad_norm": 0.4785003066062927, | |
| "learning_rate": 4.068627450980392e-06, | |
| "loss": 0.0188, | |
| "step": 1051 | |
| }, | |
| { | |
| "epoch": 5.555115511551155, | |
| "grad_norm": 0.4844956696033478, | |
| "learning_rate": 4.019607843137255e-06, | |
| "loss": 0.0175, | |
| "step": 1052 | |
| }, | |
| { | |
| "epoch": 5.56039603960396, | |
| "grad_norm": 0.40522444248199463, | |
| "learning_rate": 3.970588235294117e-06, | |
| "loss": 0.0159, | |
| "step": 1053 | |
| }, | |
| { | |
| "epoch": 5.565676567656766, | |
| "grad_norm": 0.40739139914512634, | |
| "learning_rate": 3.92156862745098e-06, | |
| "loss": 0.0165, | |
| "step": 1054 | |
| }, | |
| { | |
| "epoch": 5.570957095709571, | |
| "grad_norm": 0.42678526043891907, | |
| "learning_rate": 3.872549019607843e-06, | |
| "loss": 0.0152, | |
| "step": 1055 | |
| }, | |
| { | |
| "epoch": 5.576237623762376, | |
| "grad_norm": 0.5168190598487854, | |
| "learning_rate": 3.823529411764706e-06, | |
| "loss": 0.0164, | |
| "step": 1056 | |
| }, | |
| { | |
| "epoch": 5.581518151815182, | |
| "grad_norm": 0.4231308400630951, | |
| "learning_rate": 3.7745098039215686e-06, | |
| "loss": 0.016, | |
| "step": 1057 | |
| }, | |
| { | |
| "epoch": 5.586798679867987, | |
| "grad_norm": 0.604710578918457, | |
| "learning_rate": 3.7254901960784316e-06, | |
| "loss": 0.0225, | |
| "step": 1058 | |
| }, | |
| { | |
| "epoch": 5.592079207920792, | |
| "grad_norm": 0.6866090893745422, | |
| "learning_rate": 3.6764705882352942e-06, | |
| "loss": 0.0195, | |
| "step": 1059 | |
| }, | |
| { | |
| "epoch": 5.597359735973598, | |
| "grad_norm": 0.34970760345458984, | |
| "learning_rate": 3.6274509803921573e-06, | |
| "loss": 0.014, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 5.602640264026403, | |
| "grad_norm": 0.4601968228816986, | |
| "learning_rate": 3.5784313725490194e-06, | |
| "loss": 0.018, | |
| "step": 1061 | |
| }, | |
| { | |
| "epoch": 5.607920792079208, | |
| "grad_norm": 0.3815479874610901, | |
| "learning_rate": 3.5294117647058825e-06, | |
| "loss": 0.0152, | |
| "step": 1062 | |
| }, | |
| { | |
| "epoch": 5.6132013201320134, | |
| "grad_norm": 0.36140522360801697, | |
| "learning_rate": 3.4803921568627455e-06, | |
| "loss": 0.0149, | |
| "step": 1063 | |
| }, | |
| { | |
| "epoch": 5.618481848184818, | |
| "grad_norm": 0.485866904258728, | |
| "learning_rate": 3.431372549019608e-06, | |
| "loss": 0.0171, | |
| "step": 1064 | |
| }, | |
| { | |
| "epoch": 5.623762376237623, | |
| "grad_norm": 0.5746606588363647, | |
| "learning_rate": 3.3823529411764707e-06, | |
| "loss": 0.0168, | |
| "step": 1065 | |
| }, | |
| { | |
| "epoch": 5.629042904290429, | |
| "grad_norm": 0.5221585035324097, | |
| "learning_rate": 3.3333333333333333e-06, | |
| "loss": 0.0199, | |
| "step": 1066 | |
| }, | |
| { | |
| "epoch": 5.634323432343234, | |
| "grad_norm": 0.538971483707428, | |
| "learning_rate": 3.2843137254901964e-06, | |
| "loss": 0.0225, | |
| "step": 1067 | |
| }, | |
| { | |
| "epoch": 5.639603960396039, | |
| "grad_norm": 0.4141756594181061, | |
| "learning_rate": 3.2352941176470594e-06, | |
| "loss": 0.0171, | |
| "step": 1068 | |
| }, | |
| { | |
| "epoch": 5.644884488448845, | |
| "grad_norm": 0.3699265420436859, | |
| "learning_rate": 3.1862745098039216e-06, | |
| "loss": 0.0127, | |
| "step": 1069 | |
| }, | |
| { | |
| "epoch": 5.65016501650165, | |
| "grad_norm": 0.8097654581069946, | |
| "learning_rate": 3.1372549019607846e-06, | |
| "loss": 0.0168, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 5.655445544554455, | |
| "grad_norm": 0.8609626889228821, | |
| "learning_rate": 3.0882352941176472e-06, | |
| "loss": 0.0202, | |
| "step": 1071 | |
| }, | |
| { | |
| "epoch": 5.660726072607261, | |
| "grad_norm": 0.7239216566085815, | |
| "learning_rate": 3.03921568627451e-06, | |
| "loss": 0.0173, | |
| "step": 1072 | |
| }, | |
| { | |
| "epoch": 5.666006600660066, | |
| "grad_norm": 0.5894525647163391, | |
| "learning_rate": 2.990196078431373e-06, | |
| "loss": 0.0248, | |
| "step": 1073 | |
| }, | |
| { | |
| "epoch": 5.671287128712871, | |
| "grad_norm": 0.39227187633514404, | |
| "learning_rate": 2.9411764705882355e-06, | |
| "loss": 0.0173, | |
| "step": 1074 | |
| }, | |
| { | |
| "epoch": 5.676567656765677, | |
| "grad_norm": 0.5740650296211243, | |
| "learning_rate": 2.8921568627450985e-06, | |
| "loss": 0.0196, | |
| "step": 1075 | |
| }, | |
| { | |
| "epoch": 5.681848184818482, | |
| "grad_norm": 0.5081339478492737, | |
| "learning_rate": 2.8431372549019607e-06, | |
| "loss": 0.0178, | |
| "step": 1076 | |
| }, | |
| { | |
| "epoch": 5.687128712871287, | |
| "grad_norm": 0.45653530955314636, | |
| "learning_rate": 2.7941176470588237e-06, | |
| "loss": 0.017, | |
| "step": 1077 | |
| }, | |
| { | |
| "epoch": 5.692409240924093, | |
| "grad_norm": 0.5196064710617065, | |
| "learning_rate": 2.7450980392156863e-06, | |
| "loss": 0.0158, | |
| "step": 1078 | |
| }, | |
| { | |
| "epoch": 5.697689768976898, | |
| "grad_norm": 0.5605891942977905, | |
| "learning_rate": 2.6960784313725493e-06, | |
| "loss": 0.0172, | |
| "step": 1079 | |
| }, | |
| { | |
| "epoch": 5.702970297029703, | |
| "grad_norm": 0.6851004958152771, | |
| "learning_rate": 2.647058823529412e-06, | |
| "loss": 0.0172, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 5.708250825082509, | |
| "grad_norm": 0.4508073031902313, | |
| "learning_rate": 2.5980392156862746e-06, | |
| "loss": 0.0171, | |
| "step": 1081 | |
| }, | |
| { | |
| "epoch": 5.713531353135314, | |
| "grad_norm": 0.4393492639064789, | |
| "learning_rate": 2.549019607843137e-06, | |
| "loss": 0.0167, | |
| "step": 1082 | |
| }, | |
| { | |
| "epoch": 5.718811881188119, | |
| "grad_norm": 0.5042679309844971, | |
| "learning_rate": 2.5e-06, | |
| "loss": 0.0197, | |
| "step": 1083 | |
| }, | |
| { | |
| "epoch": 5.724092409240924, | |
| "grad_norm": 0.40561428666114807, | |
| "learning_rate": 2.450980392156863e-06, | |
| "loss": 0.0175, | |
| "step": 1084 | |
| }, | |
| { | |
| "epoch": 5.729372937293729, | |
| "grad_norm": 0.390462189912796, | |
| "learning_rate": 2.401960784313726e-06, | |
| "loss": 0.0166, | |
| "step": 1085 | |
| }, | |
| { | |
| "epoch": 5.734653465346534, | |
| "grad_norm": 0.44142773747444153, | |
| "learning_rate": 2.3529411764705885e-06, | |
| "loss": 0.0161, | |
| "step": 1086 | |
| }, | |
| { | |
| "epoch": 5.73993399339934, | |
| "grad_norm": 0.6280815601348877, | |
| "learning_rate": 2.303921568627451e-06, | |
| "loss": 0.0191, | |
| "step": 1087 | |
| }, | |
| { | |
| "epoch": 5.745214521452145, | |
| "grad_norm": 0.6288333535194397, | |
| "learning_rate": 2.2549019607843137e-06, | |
| "loss": 0.0169, | |
| "step": 1088 | |
| }, | |
| { | |
| "epoch": 5.75049504950495, | |
| "grad_norm": 0.39587247371673584, | |
| "learning_rate": 2.2058823529411767e-06, | |
| "loss": 0.0154, | |
| "step": 1089 | |
| }, | |
| { | |
| "epoch": 5.755775577557756, | |
| "grad_norm": 0.8085327744483948, | |
| "learning_rate": 2.1568627450980393e-06, | |
| "loss": 0.0214, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 5.761056105610561, | |
| "grad_norm": 0.31861352920532227, | |
| "learning_rate": 2.1078431372549023e-06, | |
| "loss": 0.0113, | |
| "step": 1091 | |
| }, | |
| { | |
| "epoch": 5.766336633663366, | |
| "grad_norm": 0.6090717315673828, | |
| "learning_rate": 2.058823529411765e-06, | |
| "loss": 0.0176, | |
| "step": 1092 | |
| }, | |
| { | |
| "epoch": 5.771617161716172, | |
| "grad_norm": 0.37864407896995544, | |
| "learning_rate": 2.0098039215686276e-06, | |
| "loss": 0.013, | |
| "step": 1093 | |
| }, | |
| { | |
| "epoch": 5.776897689768977, | |
| "grad_norm": 0.39943984150886536, | |
| "learning_rate": 1.96078431372549e-06, | |
| "loss": 0.0183, | |
| "step": 1094 | |
| }, | |
| { | |
| "epoch": 5.782178217821782, | |
| "grad_norm": 0.3896962106227875, | |
| "learning_rate": 1.911764705882353e-06, | |
| "loss": 0.0147, | |
| "step": 1095 | |
| }, | |
| { | |
| "epoch": 5.787458745874588, | |
| "grad_norm": 0.8300476670265198, | |
| "learning_rate": 1.8627450980392158e-06, | |
| "loss": 0.0219, | |
| "step": 1096 | |
| }, | |
| { | |
| "epoch": 5.792739273927393, | |
| "grad_norm": 0.7217696905136108, | |
| "learning_rate": 1.8137254901960786e-06, | |
| "loss": 0.0193, | |
| "step": 1097 | |
| }, | |
| { | |
| "epoch": 5.798019801980198, | |
| "grad_norm": 0.7046459317207336, | |
| "learning_rate": 1.7647058823529412e-06, | |
| "loss": 0.019, | |
| "step": 1098 | |
| }, | |
| { | |
| "epoch": 5.803300330033004, | |
| "grad_norm": 0.5788043141365051, | |
| "learning_rate": 1.715686274509804e-06, | |
| "loss": 0.0197, | |
| "step": 1099 | |
| }, | |
| { | |
| "epoch": 5.808580858085809, | |
| "grad_norm": 0.3641752004623413, | |
| "learning_rate": 1.6666666666666667e-06, | |
| "loss": 0.0139, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 5.813861386138614, | |
| "grad_norm": 0.5467818379402161, | |
| "learning_rate": 1.6176470588235297e-06, | |
| "loss": 0.0185, | |
| "step": 1101 | |
| }, | |
| { | |
| "epoch": 5.81914191419142, | |
| "grad_norm": 0.7358901500701904, | |
| "learning_rate": 1.5686274509803923e-06, | |
| "loss": 0.0218, | |
| "step": 1102 | |
| }, | |
| { | |
| "epoch": 5.824422442244225, | |
| "grad_norm": 0.39868056774139404, | |
| "learning_rate": 1.519607843137255e-06, | |
| "loss": 0.0168, | |
| "step": 1103 | |
| }, | |
| { | |
| "epoch": 5.8297029702970296, | |
| "grad_norm": 0.5790492296218872, | |
| "learning_rate": 1.4705882352941177e-06, | |
| "loss": 0.0154, | |
| "step": 1104 | |
| }, | |
| { | |
| "epoch": 5.834983498349835, | |
| "grad_norm": 0.6561235189437866, | |
| "learning_rate": 1.4215686274509803e-06, | |
| "loss": 0.0184, | |
| "step": 1105 | |
| }, | |
| { | |
| "epoch": 5.84026402640264, | |
| "grad_norm": 0.4628670811653137, | |
| "learning_rate": 1.3725490196078432e-06, | |
| "loss": 0.0178, | |
| "step": 1106 | |
| }, | |
| { | |
| "epoch": 5.845544554455445, | |
| "grad_norm": 0.5217213034629822, | |
| "learning_rate": 1.323529411764706e-06, | |
| "loss": 0.0179, | |
| "step": 1107 | |
| }, | |
| { | |
| "epoch": 5.850825082508251, | |
| "grad_norm": 0.6204583644866943, | |
| "learning_rate": 1.2745098039215686e-06, | |
| "loss": 0.0197, | |
| "step": 1108 | |
| }, | |
| { | |
| "epoch": 5.856105610561056, | |
| "grad_norm": 0.48575830459594727, | |
| "learning_rate": 1.2254901960784314e-06, | |
| "loss": 0.0191, | |
| "step": 1109 | |
| }, | |
| { | |
| "epoch": 5.861386138613861, | |
| "grad_norm": 0.8481599688529968, | |
| "learning_rate": 1.1764705882352942e-06, | |
| "loss": 0.023, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 5.866666666666667, | |
| "grad_norm": 0.5180693864822388, | |
| "learning_rate": 1.1274509803921568e-06, | |
| "loss": 0.016, | |
| "step": 1111 | |
| }, | |
| { | |
| "epoch": 5.871947194719472, | |
| "grad_norm": 0.47926902770996094, | |
| "learning_rate": 1.0784313725490197e-06, | |
| "loss": 0.018, | |
| "step": 1112 | |
| }, | |
| { | |
| "epoch": 5.877227722772277, | |
| "grad_norm": 0.4460168182849884, | |
| "learning_rate": 1.0294117647058825e-06, | |
| "loss": 0.0155, | |
| "step": 1113 | |
| }, | |
| { | |
| "epoch": 5.882508250825083, | |
| "grad_norm": 0.6543490886688232, | |
| "learning_rate": 9.80392156862745e-07, | |
| "loss": 0.0193, | |
| "step": 1114 | |
| }, | |
| { | |
| "epoch": 5.887788778877888, | |
| "grad_norm": 0.5319867134094238, | |
| "learning_rate": 9.313725490196079e-07, | |
| "loss": 0.0164, | |
| "step": 1115 | |
| }, | |
| { | |
| "epoch": 5.893069306930693, | |
| "grad_norm": 0.5237565636634827, | |
| "learning_rate": 8.823529411764706e-07, | |
| "loss": 0.0143, | |
| "step": 1116 | |
| }, | |
| { | |
| "epoch": 5.898349834983498, | |
| "grad_norm": 0.43923959136009216, | |
| "learning_rate": 8.333333333333333e-07, | |
| "loss": 0.0172, | |
| "step": 1117 | |
| }, | |
| { | |
| "epoch": 5.903630363036304, | |
| "grad_norm": 0.5296475291252136, | |
| "learning_rate": 7.843137254901962e-07, | |
| "loss": 0.0182, | |
| "step": 1118 | |
| }, | |
| { | |
| "epoch": 5.908910891089109, | |
| "grad_norm": 0.38380494713783264, | |
| "learning_rate": 7.352941176470589e-07, | |
| "loss": 0.0159, | |
| "step": 1119 | |
| }, | |
| { | |
| "epoch": 5.914191419141914, | |
| "grad_norm": 0.6031242609024048, | |
| "learning_rate": 6.862745098039216e-07, | |
| "loss": 0.0152, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 5.91947194719472, | |
| "grad_norm": 0.5882181525230408, | |
| "learning_rate": 6.372549019607843e-07, | |
| "loss": 0.0171, | |
| "step": 1121 | |
| }, | |
| { | |
| "epoch": 5.924752475247525, | |
| "grad_norm": 0.41396641731262207, | |
| "learning_rate": 5.882352941176471e-07, | |
| "loss": 0.0162, | |
| "step": 1122 | |
| }, | |
| { | |
| "epoch": 5.93003300330033, | |
| "grad_norm": 0.595112144947052, | |
| "learning_rate": 5.392156862745098e-07, | |
| "loss": 0.0174, | |
| "step": 1123 | |
| }, | |
| { | |
| "epoch": 5.9353135313531356, | |
| "grad_norm": 0.6171261072158813, | |
| "learning_rate": 4.901960784313725e-07, | |
| "loss": 0.022, | |
| "step": 1124 | |
| }, | |
| { | |
| "epoch": 5.9405940594059405, | |
| "grad_norm": 0.40438172221183777, | |
| "learning_rate": 4.411764705882353e-07, | |
| "loss": 0.0144, | |
| "step": 1125 | |
| }, | |
| { | |
| "epoch": 5.9458745874587455, | |
| "grad_norm": 0.6478520035743713, | |
| "learning_rate": 3.921568627450981e-07, | |
| "loss": 0.0169, | |
| "step": 1126 | |
| }, | |
| { | |
| "epoch": 5.951155115511551, | |
| "grad_norm": 0.3631436824798584, | |
| "learning_rate": 3.431372549019608e-07, | |
| "loss": 0.0149, | |
| "step": 1127 | |
| }, | |
| { | |
| "epoch": 5.956435643564356, | |
| "grad_norm": 0.7141023278236389, | |
| "learning_rate": 2.9411764705882356e-07, | |
| "loss": 0.0169, | |
| "step": 1128 | |
| }, | |
| { | |
| "epoch": 5.961716171617161, | |
| "grad_norm": 0.6038995981216431, | |
| "learning_rate": 2.4509803921568627e-07, | |
| "loss": 0.0197, | |
| "step": 1129 | |
| }, | |
| { | |
| "epoch": 5.966996699669967, | |
| "grad_norm": 0.5263422727584839, | |
| "learning_rate": 1.9607843137254904e-07, | |
| "loss": 0.0201, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 5.972277227722772, | |
| "grad_norm": 0.604172945022583, | |
| "learning_rate": 1.4705882352941178e-07, | |
| "loss": 0.0181, | |
| "step": 1131 | |
| }, | |
| { | |
| "epoch": 5.977557755775577, | |
| "grad_norm": 0.6950435638427734, | |
| "learning_rate": 9.803921568627452e-08, | |
| "loss": 0.0211, | |
| "step": 1132 | |
| }, | |
| { | |
| "epoch": 5.982838283828383, | |
| "grad_norm": 0.5159468650817871, | |
| "learning_rate": 4.901960784313726e-08, | |
| "loss": 0.0153, | |
| "step": 1133 | |
| }, | |
| { | |
| "epoch": 5.988118811881188, | |
| "grad_norm": 0.7329381108283997, | |
| "learning_rate": 0.0, | |
| "loss": 0.0224, | |
| "step": 1134 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 1134, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 6, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 6.176433819923251e+17, | |
| "train_batch_size": 2, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |