| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 4.0, | |
| "eval_steps": 500, | |
| "global_step": 500, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.008, | |
| "grad_norm": 145.86453247070312, | |
| "learning_rate": 1e-05, | |
| "loss": 18.4693, | |
| "mean_token_accuracy": 0.4418533444404602, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.016, | |
| "grad_norm": 131.47439575195312, | |
| "learning_rate": 2e-05, | |
| "loss": 17.1904, | |
| "mean_token_accuracy": 0.4580386132001877, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.024, | |
| "grad_norm": 93.59867095947266, | |
| "learning_rate": 3e-05, | |
| "loss": 13.3976, | |
| "mean_token_accuracy": 0.520381361246109, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.032, | |
| "grad_norm": 89.1477279663086, | |
| "learning_rate": 4e-05, | |
| "loss": 11.5329, | |
| "mean_token_accuracy": 0.5683971494436264, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 109.09867858886719, | |
| "learning_rate": 5e-05, | |
| "loss": 9.6112, | |
| "mean_token_accuracy": 0.6148007661104202, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.048, | |
| "grad_norm": 70.35514068603516, | |
| "learning_rate": 4.98989898989899e-05, | |
| "loss": 8.5729, | |
| "mean_token_accuracy": 0.6641161292791367, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.056, | |
| "grad_norm": 69.57074737548828, | |
| "learning_rate": 4.97979797979798e-05, | |
| "loss": 7.9422, | |
| "mean_token_accuracy": 0.6828223764896393, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.064, | |
| "grad_norm": 55.16558837890625, | |
| "learning_rate": 4.9696969696969694e-05, | |
| "loss": 7.5634, | |
| "mean_token_accuracy": 0.6935229450464249, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.072, | |
| "grad_norm": 55.306785583496094, | |
| "learning_rate": 4.9595959595959594e-05, | |
| "loss": 6.9885, | |
| "mean_token_accuracy": 0.6911955326795578, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 60.01897048950195, | |
| "learning_rate": 4.94949494949495e-05, | |
| "loss": 7.8794, | |
| "mean_token_accuracy": 0.7122491598129272, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.088, | |
| "grad_norm": 44.305118560791016, | |
| "learning_rate": 4.93939393939394e-05, | |
| "loss": 6.9955, | |
| "mean_token_accuracy": 0.7298235297203064, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.096, | |
| "grad_norm": 40.814090728759766, | |
| "learning_rate": 4.92929292929293e-05, | |
| "loss": 6.7332, | |
| "mean_token_accuracy": 0.7177170217037201, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.104, | |
| "grad_norm": 31.777416229248047, | |
| "learning_rate": 4.919191919191919e-05, | |
| "loss": 6.5969, | |
| "mean_token_accuracy": 0.7185130566358566, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.112, | |
| "grad_norm": 32.8369140625, | |
| "learning_rate": 4.909090909090909e-05, | |
| "loss": 6.518, | |
| "mean_token_accuracy": 0.7157497704029083, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 27.942502975463867, | |
| "learning_rate": 4.898989898989899e-05, | |
| "loss": 6.0411, | |
| "mean_token_accuracy": 0.7299728095531464, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.128, | |
| "grad_norm": 26.94302749633789, | |
| "learning_rate": 4.888888888888889e-05, | |
| "loss": 6.0288, | |
| "mean_token_accuracy": 0.7378736436367035, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.136, | |
| "grad_norm": 29.398008346557617, | |
| "learning_rate": 4.878787878787879e-05, | |
| "loss": 5.7906, | |
| "mean_token_accuracy": 0.7615540027618408, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.144, | |
| "grad_norm": 33.691036224365234, | |
| "learning_rate": 4.868686868686869e-05, | |
| "loss": 6.148, | |
| "mean_token_accuracy": 0.7420072853565216, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.152, | |
| "grad_norm": 28.954744338989258, | |
| "learning_rate": 4.858585858585859e-05, | |
| "loss": 6.074, | |
| "mean_token_accuracy": 0.7376030832529068, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 30.093732833862305, | |
| "learning_rate": 4.848484848484849e-05, | |
| "loss": 6.0524, | |
| "mean_token_accuracy": 0.7355515509843826, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.168, | |
| "grad_norm": 32.198020935058594, | |
| "learning_rate": 4.838383838383839e-05, | |
| "loss": 6.3595, | |
| "mean_token_accuracy": 0.7233454436063766, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.176, | |
| "grad_norm": 27.62493896484375, | |
| "learning_rate": 4.828282828282829e-05, | |
| "loss": 4.9527, | |
| "mean_token_accuracy": 0.7749838680028915, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.184, | |
| "grad_norm": 29.359989166259766, | |
| "learning_rate": 4.8181818181818186e-05, | |
| "loss": 6.0378, | |
| "mean_token_accuracy": 0.7387848794460297, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.192, | |
| "grad_norm": 28.882671356201172, | |
| "learning_rate": 4.808080808080808e-05, | |
| "loss": 5.889, | |
| "mean_token_accuracy": 0.7291557788848877, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 26.889739990234375, | |
| "learning_rate": 4.797979797979798e-05, | |
| "loss": 5.5897, | |
| "mean_token_accuracy": 0.7566297799348831, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.208, | |
| "grad_norm": 27.18845558166504, | |
| "learning_rate": 4.787878787878788e-05, | |
| "loss": 5.4735, | |
| "mean_token_accuracy": 0.7656728178262711, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.216, | |
| "grad_norm": 23.547197341918945, | |
| "learning_rate": 4.7777777777777784e-05, | |
| "loss": 5.6021, | |
| "mean_token_accuracy": 0.740881085395813, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.224, | |
| "grad_norm": 26.538345336914062, | |
| "learning_rate": 4.7676767676767684e-05, | |
| "loss": 5.22, | |
| "mean_token_accuracy": 0.7639279812574387, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.232, | |
| "grad_norm": 24.266155242919922, | |
| "learning_rate": 4.7575757575757576e-05, | |
| "loss": 5.449, | |
| "mean_token_accuracy": 0.7561690509319305, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 26.028945922851562, | |
| "learning_rate": 4.7474747474747476e-05, | |
| "loss": 5.6177, | |
| "mean_token_accuracy": 0.7446402460336685, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.248, | |
| "grad_norm": 29.959096908569336, | |
| "learning_rate": 4.7373737373737375e-05, | |
| "loss": 5.9764, | |
| "mean_token_accuracy": 0.7454901784658432, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.256, | |
| "grad_norm": 33.07018280029297, | |
| "learning_rate": 4.7272727272727275e-05, | |
| "loss": 4.9627, | |
| "mean_token_accuracy": 0.787445604801178, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.264, | |
| "grad_norm": 26.5466365814209, | |
| "learning_rate": 4.7171717171717174e-05, | |
| "loss": 5.0905, | |
| "mean_token_accuracy": 0.772098571062088, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.272, | |
| "grad_norm": 29.91367530822754, | |
| "learning_rate": 4.7070707070707074e-05, | |
| "loss": 5.3797, | |
| "mean_token_accuracy": 0.7676996439695358, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 27.248432159423828, | |
| "learning_rate": 4.696969696969697e-05, | |
| "loss": 5.671, | |
| "mean_token_accuracy": 0.755952462553978, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.288, | |
| "grad_norm": 27.235933303833008, | |
| "learning_rate": 4.686868686868687e-05, | |
| "loss": 5.2058, | |
| "mean_token_accuracy": 0.787963479757309, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.296, | |
| "grad_norm": 28.442148208618164, | |
| "learning_rate": 4.676767676767677e-05, | |
| "loss": 4.8147, | |
| "mean_token_accuracy": 0.7970237880945206, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.304, | |
| "grad_norm": 28.56350326538086, | |
| "learning_rate": 4.666666666666667e-05, | |
| "loss": 5.7801, | |
| "mean_token_accuracy": 0.7452163100242615, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.312, | |
| "grad_norm": 24.276662826538086, | |
| "learning_rate": 4.656565656565657e-05, | |
| "loss": 5.3188, | |
| "mean_token_accuracy": 0.7638489454984665, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 25.21291732788086, | |
| "learning_rate": 4.6464646464646464e-05, | |
| "loss": 5.4533, | |
| "mean_token_accuracy": 0.7537444680929184, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.328, | |
| "grad_norm": 26.179948806762695, | |
| "learning_rate": 4.636363636363636e-05, | |
| "loss": 5.2261, | |
| "mean_token_accuracy": 0.7736174613237381, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.336, | |
| "grad_norm": 23.527063369750977, | |
| "learning_rate": 4.626262626262626e-05, | |
| "loss": 4.9225, | |
| "mean_token_accuracy": 0.7722691893577576, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.344, | |
| "grad_norm": 24.048324584960938, | |
| "learning_rate": 4.616161616161616e-05, | |
| "loss": 5.0608, | |
| "mean_token_accuracy": 0.7645718157291412, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.352, | |
| "grad_norm": 24.779226303100586, | |
| "learning_rate": 4.606060606060607e-05, | |
| "loss": 4.8156, | |
| "mean_token_accuracy": 0.7857754081487656, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 23.97334861755371, | |
| "learning_rate": 4.595959595959596e-05, | |
| "loss": 5.0281, | |
| "mean_token_accuracy": 0.7758517414331436, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.368, | |
| "grad_norm": 24.752384185791016, | |
| "learning_rate": 4.585858585858586e-05, | |
| "loss": 4.7445, | |
| "mean_token_accuracy": 0.7875040620565414, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.376, | |
| "grad_norm": 23.33396339416504, | |
| "learning_rate": 4.575757575757576e-05, | |
| "loss": 4.9179, | |
| "mean_token_accuracy": 0.7820334136486053, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.384, | |
| "grad_norm": 25.201007843017578, | |
| "learning_rate": 4.565656565656566e-05, | |
| "loss": 5.5614, | |
| "mean_token_accuracy": 0.7553493082523346, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.392, | |
| "grad_norm": 26.965972900390625, | |
| "learning_rate": 4.555555555555556e-05, | |
| "loss": 5.2295, | |
| "mean_token_accuracy": 0.7683407664299011, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 27.544897079467773, | |
| "learning_rate": 4.545454545454546e-05, | |
| "loss": 5.6261, | |
| "mean_token_accuracy": 0.7545335441827774, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.408, | |
| "grad_norm": 26.644731521606445, | |
| "learning_rate": 4.535353535353535e-05, | |
| "loss": 4.6757, | |
| "mean_token_accuracy": 0.7908278107643127, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.416, | |
| "grad_norm": 21.93665885925293, | |
| "learning_rate": 4.525252525252526e-05, | |
| "loss": 4.631, | |
| "mean_token_accuracy": 0.7804701924324036, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.424, | |
| "grad_norm": 22.430749893188477, | |
| "learning_rate": 4.515151515151516e-05, | |
| "loss": 5.1858, | |
| "mean_token_accuracy": 0.7598922550678253, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.432, | |
| "grad_norm": 22.974781036376953, | |
| "learning_rate": 4.5050505050505056e-05, | |
| "loss": 4.8737, | |
| "mean_token_accuracy": 0.7714487016201019, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 25.078826904296875, | |
| "learning_rate": 4.494949494949495e-05, | |
| "loss": 5.1416, | |
| "mean_token_accuracy": 0.7775991857051849, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.448, | |
| "grad_norm": 26.377742767333984, | |
| "learning_rate": 4.484848484848485e-05, | |
| "loss": 4.3445, | |
| "mean_token_accuracy": 0.8000840097665787, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.456, | |
| "grad_norm": 23.724504470825195, | |
| "learning_rate": 4.474747474747475e-05, | |
| "loss": 5.0277, | |
| "mean_token_accuracy": 0.7753082364797592, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.464, | |
| "grad_norm": 24.300643920898438, | |
| "learning_rate": 4.464646464646465e-05, | |
| "loss": 4.0889, | |
| "mean_token_accuracy": 0.805213525891304, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.472, | |
| "grad_norm": 24.581886291503906, | |
| "learning_rate": 4.454545454545455e-05, | |
| "loss": 5.0518, | |
| "mean_token_accuracy": 0.7843358516693115, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 26.722314834594727, | |
| "learning_rate": 4.4444444444444447e-05, | |
| "loss": 5.087, | |
| "mean_token_accuracy": 0.7961481958627701, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.488, | |
| "grad_norm": 25.016273498535156, | |
| "learning_rate": 4.4343434343434346e-05, | |
| "loss": 5.5151, | |
| "mean_token_accuracy": 0.7682371735572815, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.496, | |
| "grad_norm": 24.71924591064453, | |
| "learning_rate": 4.4242424242424246e-05, | |
| "loss": 4.6307, | |
| "mean_token_accuracy": 0.7860427796840668, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.504, | |
| "grad_norm": 21.29850959777832, | |
| "learning_rate": 4.4141414141414145e-05, | |
| "loss": 4.7079, | |
| "mean_token_accuracy": 0.7890657931566238, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.512, | |
| "grad_norm": 23.304716110229492, | |
| "learning_rate": 4.4040404040404044e-05, | |
| "loss": 5.0797, | |
| "mean_token_accuracy": 0.7593720108270645, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 21.623689651489258, | |
| "learning_rate": 4.3939393939393944e-05, | |
| "loss": 4.749, | |
| "mean_token_accuracy": 0.778199702501297, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.528, | |
| "grad_norm": 19.37482452392578, | |
| "learning_rate": 4.383838383838384e-05, | |
| "loss": 4.6279, | |
| "mean_token_accuracy": 0.7791639715433121, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.536, | |
| "grad_norm": 22.020885467529297, | |
| "learning_rate": 4.3737373737373736e-05, | |
| "loss": 4.9881, | |
| "mean_token_accuracy": 0.7844051718711853, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.544, | |
| "grad_norm": 23.672704696655273, | |
| "learning_rate": 4.3636363636363636e-05, | |
| "loss": 5.7155, | |
| "mean_token_accuracy": 0.7384811043739319, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.552, | |
| "grad_norm": 22.16874122619629, | |
| "learning_rate": 4.3535353535353535e-05, | |
| "loss": 5.272, | |
| "mean_token_accuracy": 0.7523021399974823, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 21.230941772460938, | |
| "learning_rate": 4.343434343434344e-05, | |
| "loss": 4.7733, | |
| "mean_token_accuracy": 0.7712208926677704, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.568, | |
| "grad_norm": 23.96208381652832, | |
| "learning_rate": 4.3333333333333334e-05, | |
| "loss": 5.2342, | |
| "mean_token_accuracy": 0.7734561115503311, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.576, | |
| "grad_norm": 19.650798797607422, | |
| "learning_rate": 4.3232323232323234e-05, | |
| "loss": 4.0045, | |
| "mean_token_accuracy": 0.8018752634525299, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.584, | |
| "grad_norm": 21.702190399169922, | |
| "learning_rate": 4.313131313131313e-05, | |
| "loss": 5.4078, | |
| "mean_token_accuracy": 0.7569945156574249, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.592, | |
| "grad_norm": 20.384090423583984, | |
| "learning_rate": 4.303030303030303e-05, | |
| "loss": 4.2726, | |
| "mean_token_accuracy": 0.8059134036302567, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 20.51775360107422, | |
| "learning_rate": 4.292929292929293e-05, | |
| "loss": 4.3555, | |
| "mean_token_accuracy": 0.7980977892875671, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.608, | |
| "grad_norm": 19.564922332763672, | |
| "learning_rate": 4.282828282828283e-05, | |
| "loss": 4.5646, | |
| "mean_token_accuracy": 0.7788865268230438, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.616, | |
| "grad_norm": 24.02853012084961, | |
| "learning_rate": 4.2727272727272724e-05, | |
| "loss": 4.8484, | |
| "mean_token_accuracy": 0.794918566942215, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.624, | |
| "grad_norm": 19.45448875427246, | |
| "learning_rate": 4.262626262626263e-05, | |
| "loss": 4.201, | |
| "mean_token_accuracy": 0.79364313185215, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.632, | |
| "grad_norm": 20.79984474182129, | |
| "learning_rate": 4.252525252525253e-05, | |
| "loss": 4.7681, | |
| "mean_token_accuracy": 0.7742249518632889, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 19.014602661132812, | |
| "learning_rate": 4.242424242424243e-05, | |
| "loss": 4.3379, | |
| "mean_token_accuracy": 0.7996639758348465, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.648, | |
| "grad_norm": 20.98752212524414, | |
| "learning_rate": 4.232323232323233e-05, | |
| "loss": 4.4264, | |
| "mean_token_accuracy": 0.7978147119283676, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.656, | |
| "grad_norm": 19.011503219604492, | |
| "learning_rate": 4.222222222222222e-05, | |
| "loss": 4.4577, | |
| "mean_token_accuracy": 0.7895375341176987, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.664, | |
| "grad_norm": 18.420442581176758, | |
| "learning_rate": 4.212121212121212e-05, | |
| "loss": 4.2517, | |
| "mean_token_accuracy": 0.8092215359210968, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.672, | |
| "grad_norm": 20.489845275878906, | |
| "learning_rate": 4.202020202020202e-05, | |
| "loss": 4.8926, | |
| "mean_token_accuracy": 0.7734092026948929, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 19.172943115234375, | |
| "learning_rate": 4.191919191919192e-05, | |
| "loss": 4.0021, | |
| "mean_token_accuracy": 0.8282384127378464, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.688, | |
| "grad_norm": 18.97472381591797, | |
| "learning_rate": 4.181818181818182e-05, | |
| "loss": 4.2862, | |
| "mean_token_accuracy": 0.8147316575050354, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.696, | |
| "grad_norm": 22.11132049560547, | |
| "learning_rate": 4.171717171717172e-05, | |
| "loss": 4.7801, | |
| "mean_token_accuracy": 0.7815755158662796, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.704, | |
| "grad_norm": 24.296100616455078, | |
| "learning_rate": 4.161616161616162e-05, | |
| "loss": 4.7305, | |
| "mean_token_accuracy": 0.8020643591880798, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.712, | |
| "grad_norm": 20.803821563720703, | |
| "learning_rate": 4.151515151515152e-05, | |
| "loss": 5.2232, | |
| "mean_token_accuracy": 0.7705590724945068, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 18.738903045654297, | |
| "learning_rate": 4.141414141414142e-05, | |
| "loss": 4.2926, | |
| "mean_token_accuracy": 0.8065244108438492, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.728, | |
| "grad_norm": 21.6732120513916, | |
| "learning_rate": 4.131313131313132e-05, | |
| "loss": 4.3887, | |
| "mean_token_accuracy": 0.7921336442232132, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.736, | |
| "grad_norm": 21.643827438354492, | |
| "learning_rate": 4.1212121212121216e-05, | |
| "loss": 4.5504, | |
| "mean_token_accuracy": 0.7796259820461273, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.744, | |
| "grad_norm": 19.445819854736328, | |
| "learning_rate": 4.111111111111111e-05, | |
| "loss": 4.2139, | |
| "mean_token_accuracy": 0.7942475080490112, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.752, | |
| "grad_norm": 20.41779899597168, | |
| "learning_rate": 4.101010101010101e-05, | |
| "loss": 4.4009, | |
| "mean_token_accuracy": 0.7928617000579834, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 20.34018325805664, | |
| "learning_rate": 4.0909090909090915e-05, | |
| "loss": 4.2759, | |
| "mean_token_accuracy": 0.800753116607666, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.768, | |
| "grad_norm": 19.045612335205078, | |
| "learning_rate": 4.0808080808080814e-05, | |
| "loss": 4.0605, | |
| "mean_token_accuracy": 0.8086353838443756, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.776, | |
| "grad_norm": 20.315345764160156, | |
| "learning_rate": 4.070707070707071e-05, | |
| "loss": 4.3944, | |
| "mean_token_accuracy": 0.7914804071187973, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.784, | |
| "grad_norm": 19.407426834106445, | |
| "learning_rate": 4.0606060606060606e-05, | |
| "loss": 3.9872, | |
| "mean_token_accuracy": 0.8120452463626862, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.792, | |
| "grad_norm": 22.875099182128906, | |
| "learning_rate": 4.0505050505050506e-05, | |
| "loss": 4.3298, | |
| "mean_token_accuracy": 0.7921182066202164, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 23.487300872802734, | |
| "learning_rate": 4.0404040404040405e-05, | |
| "loss": 4.966, | |
| "mean_token_accuracy": 0.7685753107070923, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.808, | |
| "grad_norm": 19.982942581176758, | |
| "learning_rate": 4.0303030303030305e-05, | |
| "loss": 4.0714, | |
| "mean_token_accuracy": 0.8111653327941895, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.816, | |
| "grad_norm": 19.33985137939453, | |
| "learning_rate": 4.0202020202020204e-05, | |
| "loss": 3.9349, | |
| "mean_token_accuracy": 0.8289202004671097, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.824, | |
| "grad_norm": 19.506378173828125, | |
| "learning_rate": 4.01010101010101e-05, | |
| "loss": 4.1454, | |
| "mean_token_accuracy": 0.8035620599985123, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.832, | |
| "grad_norm": 20.6740665435791, | |
| "learning_rate": 4e-05, | |
| "loss": 5.1995, | |
| "mean_token_accuracy": 0.7696829289197922, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 19.577415466308594, | |
| "learning_rate": 3.98989898989899e-05, | |
| "loss": 4.1304, | |
| "mean_token_accuracy": 0.7980705797672272, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.848, | |
| "grad_norm": 21.14516258239746, | |
| "learning_rate": 3.97979797979798e-05, | |
| "loss": 4.1035, | |
| "mean_token_accuracy": 0.8171603679656982, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.856, | |
| "grad_norm": 19.88467788696289, | |
| "learning_rate": 3.96969696969697e-05, | |
| "loss": 4.7231, | |
| "mean_token_accuracy": 0.7842641174793243, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.864, | |
| "grad_norm": 19.563447952270508, | |
| "learning_rate": 3.9595959595959594e-05, | |
| "loss": 4.9512, | |
| "mean_token_accuracy": 0.7920339405536652, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.872, | |
| "grad_norm": 19.343595504760742, | |
| "learning_rate": 3.9494949494949494e-05, | |
| "loss": 4.4894, | |
| "mean_token_accuracy": 0.7923117727041245, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 19.845014572143555, | |
| "learning_rate": 3.939393939393939e-05, | |
| "loss": 3.9204, | |
| "mean_token_accuracy": 0.8179315477609634, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.888, | |
| "grad_norm": 20.17218780517578, | |
| "learning_rate": 3.929292929292929e-05, | |
| "loss": 4.6602, | |
| "mean_token_accuracy": 0.7860896587371826, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.896, | |
| "grad_norm": 17.958740234375, | |
| "learning_rate": 3.91919191919192e-05, | |
| "loss": 3.9505, | |
| "mean_token_accuracy": 0.802581325173378, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.904, | |
| "grad_norm": 17.446216583251953, | |
| "learning_rate": 3.909090909090909e-05, | |
| "loss": 3.6399, | |
| "mean_token_accuracy": 0.8280811607837677, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.912, | |
| "grad_norm": 19.220077514648438, | |
| "learning_rate": 3.898989898989899e-05, | |
| "loss": 3.9278, | |
| "mean_token_accuracy": 0.8266900181770325, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 21.730310440063477, | |
| "learning_rate": 3.888888888888889e-05, | |
| "loss": 4.5073, | |
| "mean_token_accuracy": 0.7951602935791016, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.928, | |
| "grad_norm": 21.945823669433594, | |
| "learning_rate": 3.878787878787879e-05, | |
| "loss": 4.2317, | |
| "mean_token_accuracy": 0.8075039386749268, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.936, | |
| "grad_norm": 21.828777313232422, | |
| "learning_rate": 3.868686868686869e-05, | |
| "loss": 4.9559, | |
| "mean_token_accuracy": 0.7718204110860825, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.944, | |
| "grad_norm": 18.954065322875977, | |
| "learning_rate": 3.858585858585859e-05, | |
| "loss": 4.1636, | |
| "mean_token_accuracy": 0.8161665499210358, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.952, | |
| "grad_norm": 18.235891342163086, | |
| "learning_rate": 3.848484848484848e-05, | |
| "loss": 3.5472, | |
| "mean_token_accuracy": 0.8299020826816559, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 19.251985549926758, | |
| "learning_rate": 3.838383838383838e-05, | |
| "loss": 4.362, | |
| "mean_token_accuracy": 0.7960271835327148, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.968, | |
| "grad_norm": 19.986879348754883, | |
| "learning_rate": 3.828282828282829e-05, | |
| "loss": 4.3958, | |
| "mean_token_accuracy": 0.7863996028900146, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.976, | |
| "grad_norm": 19.505414962768555, | |
| "learning_rate": 3.818181818181819e-05, | |
| "loss": 3.7777, | |
| "mean_token_accuracy": 0.8199738562107086, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.984, | |
| "grad_norm": 20.4809627532959, | |
| "learning_rate": 3.8080808080808087e-05, | |
| "loss": 4.0683, | |
| "mean_token_accuracy": 0.7989889979362488, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.992, | |
| "grad_norm": 17.48236083984375, | |
| "learning_rate": 3.797979797979798e-05, | |
| "loss": 3.8542, | |
| "mean_token_accuracy": 0.8198198229074478, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 20.318952560424805, | |
| "learning_rate": 3.787878787878788e-05, | |
| "loss": 4.1094, | |
| "mean_token_accuracy": 0.7986647337675095, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 1.008, | |
| "grad_norm": 17.65445327758789, | |
| "learning_rate": 3.777777777777778e-05, | |
| "loss": 3.1601, | |
| "mean_token_accuracy": 0.8280518800020218, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 1.016, | |
| "grad_norm": 17.846628189086914, | |
| "learning_rate": 3.767676767676768e-05, | |
| "loss": 3.3474, | |
| "mean_token_accuracy": 0.8235991448163986, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 1.024, | |
| "grad_norm": 16.856306076049805, | |
| "learning_rate": 3.757575757575758e-05, | |
| "loss": 3.3474, | |
| "mean_token_accuracy": 0.8168051540851593, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 1.032, | |
| "grad_norm": 17.891374588012695, | |
| "learning_rate": 3.747474747474748e-05, | |
| "loss": 3.3601, | |
| "mean_token_accuracy": 0.8238555788993835, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "grad_norm": 16.272624969482422, | |
| "learning_rate": 3.7373737373737376e-05, | |
| "loss": 2.6233, | |
| "mean_token_accuracy": 0.8651646226644516, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 1.048, | |
| "grad_norm": 18.739025115966797, | |
| "learning_rate": 3.7272727272727276e-05, | |
| "loss": 3.2236, | |
| "mean_token_accuracy": 0.8394419103860855, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 1.056, | |
| "grad_norm": 19.770313262939453, | |
| "learning_rate": 3.7171717171717175e-05, | |
| "loss": 3.114, | |
| "mean_token_accuracy": 0.8411016166210175, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 1.064, | |
| "grad_norm": 17.718353271484375, | |
| "learning_rate": 3.7070707070707075e-05, | |
| "loss": 2.1838, | |
| "mean_token_accuracy": 0.8674204647541046, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 1.072, | |
| "grad_norm": 19.21759605407715, | |
| "learning_rate": 3.6969696969696974e-05, | |
| "loss": 3.137, | |
| "mean_token_accuracy": 0.8263226449489594, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "grad_norm": 21.520931243896484, | |
| "learning_rate": 3.686868686868687e-05, | |
| "loss": 3.1435, | |
| "mean_token_accuracy": 0.8464161157608032, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 1.088, | |
| "grad_norm": 22.635766983032227, | |
| "learning_rate": 3.6767676767676766e-05, | |
| "loss": 3.1886, | |
| "mean_token_accuracy": 0.8231067955493927, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 1.096, | |
| "grad_norm": 25.0559024810791, | |
| "learning_rate": 3.6666666666666666e-05, | |
| "loss": 3.6031, | |
| "mean_token_accuracy": 0.8235105574131012, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 1.104, | |
| "grad_norm": 21.050643920898438, | |
| "learning_rate": 3.656565656565657e-05, | |
| "loss": 3.1022, | |
| "mean_token_accuracy": 0.8318233489990234, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 1.112, | |
| "grad_norm": 21.77235221862793, | |
| "learning_rate": 3.6464646464646465e-05, | |
| "loss": 3.4294, | |
| "mean_token_accuracy": 0.8234484493732452, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "grad_norm": 19.92261505126953, | |
| "learning_rate": 3.6363636363636364e-05, | |
| "loss": 2.895, | |
| "mean_token_accuracy": 0.8333221822977066, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 1.1280000000000001, | |
| "grad_norm": 19.0081844329834, | |
| "learning_rate": 3.6262626262626264e-05, | |
| "loss": 3.1505, | |
| "mean_token_accuracy": 0.8352071642875671, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 1.1360000000000001, | |
| "grad_norm": 17.831018447875977, | |
| "learning_rate": 3.616161616161616e-05, | |
| "loss": 3.1531, | |
| "mean_token_accuracy": 0.8377863019704819, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 1.144, | |
| "grad_norm": 18.68701171875, | |
| "learning_rate": 3.606060606060606e-05, | |
| "loss": 3.1063, | |
| "mean_token_accuracy": 0.8171389847993851, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 1.152, | |
| "grad_norm": 16.475784301757812, | |
| "learning_rate": 3.595959595959596e-05, | |
| "loss": 2.9425, | |
| "mean_token_accuracy": 0.841956228017807, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "grad_norm": 18.2690372467041, | |
| "learning_rate": 3.5858585858585855e-05, | |
| "loss": 2.7428, | |
| "mean_token_accuracy": 0.832164078950882, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 1.168, | |
| "grad_norm": 17.770587921142578, | |
| "learning_rate": 3.575757575757576e-05, | |
| "loss": 2.9521, | |
| "mean_token_accuracy": 0.8395673930644989, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 1.176, | |
| "grad_norm": 17.223388671875, | |
| "learning_rate": 3.565656565656566e-05, | |
| "loss": 2.77, | |
| "mean_token_accuracy": 0.8558414876461029, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 1.184, | |
| "grad_norm": 18.294189453125, | |
| "learning_rate": 3.555555555555556e-05, | |
| "loss": 2.9188, | |
| "mean_token_accuracy": 0.8403861075639725, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 1.192, | |
| "grad_norm": 16.28548812866211, | |
| "learning_rate": 3.545454545454546e-05, | |
| "loss": 2.6033, | |
| "mean_token_accuracy": 0.856376975774765, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "grad_norm": 17.804237365722656, | |
| "learning_rate": 3.535353535353535e-05, | |
| "loss": 2.699, | |
| "mean_token_accuracy": 0.849004715681076, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 1.208, | |
| "grad_norm": 22.294708251953125, | |
| "learning_rate": 3.525252525252525e-05, | |
| "loss": 3.3516, | |
| "mean_token_accuracy": 0.8136928379535675, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 1.216, | |
| "grad_norm": 19.004802703857422, | |
| "learning_rate": 3.515151515151515e-05, | |
| "loss": 3.014, | |
| "mean_token_accuracy": 0.8440276980400085, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 1.224, | |
| "grad_norm": 22.38014030456543, | |
| "learning_rate": 3.505050505050505e-05, | |
| "loss": 3.2117, | |
| "mean_token_accuracy": 0.8412062674760818, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 1.232, | |
| "grad_norm": 20.0627384185791, | |
| "learning_rate": 3.494949494949495e-05, | |
| "loss": 2.9902, | |
| "mean_token_accuracy": 0.829803541302681, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "grad_norm": 19.082454681396484, | |
| "learning_rate": 3.484848484848485e-05, | |
| "loss": 3.0242, | |
| "mean_token_accuracy": 0.8390958160161972, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 1.248, | |
| "grad_norm": 18.173250198364258, | |
| "learning_rate": 3.474747474747475e-05, | |
| "loss": 2.7003, | |
| "mean_token_accuracy": 0.8489221632480621, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 1.256, | |
| "grad_norm": 20.8232479095459, | |
| "learning_rate": 3.464646464646465e-05, | |
| "loss": 2.9356, | |
| "mean_token_accuracy": 0.8414415568113327, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 1.264, | |
| "grad_norm": 22.360126495361328, | |
| "learning_rate": 3.454545454545455e-05, | |
| "loss": 3.4103, | |
| "mean_token_accuracy": 0.8287290334701538, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 1.272, | |
| "grad_norm": 17.922536849975586, | |
| "learning_rate": 3.444444444444445e-05, | |
| "loss": 2.8419, | |
| "mean_token_accuracy": 0.8393957614898682, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "grad_norm": 19.15587043762207, | |
| "learning_rate": 3.434343434343435e-05, | |
| "loss": 3.2652, | |
| "mean_token_accuracy": 0.8440860509872437, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 1.288, | |
| "grad_norm": 17.887313842773438, | |
| "learning_rate": 3.424242424242424e-05, | |
| "loss": 3.2203, | |
| "mean_token_accuracy": 0.8331134468317032, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 1.296, | |
| "grad_norm": 17.222763061523438, | |
| "learning_rate": 3.414141414141414e-05, | |
| "loss": 2.6044, | |
| "mean_token_accuracy": 0.8538916707038879, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 1.304, | |
| "grad_norm": 17.67905044555664, | |
| "learning_rate": 3.4040404040404045e-05, | |
| "loss": 3.0019, | |
| "mean_token_accuracy": 0.8431571871042252, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 1.312, | |
| "grad_norm": 16.879220962524414, | |
| "learning_rate": 3.3939393939393945e-05, | |
| "loss": 2.6565, | |
| "mean_token_accuracy": 0.8555881530046463, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "grad_norm": 19.58002281188965, | |
| "learning_rate": 3.3838383838383844e-05, | |
| "loss": 3.1684, | |
| "mean_token_accuracy": 0.8247578740119934, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 1.328, | |
| "grad_norm": 18.66584587097168, | |
| "learning_rate": 3.373737373737374e-05, | |
| "loss": 3.0177, | |
| "mean_token_accuracy": 0.8371108621358871, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 1.336, | |
| "grad_norm": 18.553882598876953, | |
| "learning_rate": 3.3636363636363636e-05, | |
| "loss": 2.9127, | |
| "mean_token_accuracy": 0.8480137139558792, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 1.3439999999999999, | |
| "grad_norm": 16.836511611938477, | |
| "learning_rate": 3.3535353535353536e-05, | |
| "loss": 2.8304, | |
| "mean_token_accuracy": 0.8449047356843948, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 1.3519999999999999, | |
| "grad_norm": 18.28154754638672, | |
| "learning_rate": 3.3434343434343435e-05, | |
| "loss": 2.8143, | |
| "mean_token_accuracy": 0.8557403534650803, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 1.3599999999999999, | |
| "grad_norm": 18.682939529418945, | |
| "learning_rate": 3.3333333333333335e-05, | |
| "loss": 2.9991, | |
| "mean_token_accuracy": 0.8342336267232895, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 1.3679999999999999, | |
| "grad_norm": 19.349193572998047, | |
| "learning_rate": 3.3232323232323234e-05, | |
| "loss": 2.6821, | |
| "mean_token_accuracy": 0.853237047791481, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 1.376, | |
| "grad_norm": 18.103790283203125, | |
| "learning_rate": 3.3131313131313134e-05, | |
| "loss": 2.8468, | |
| "mean_token_accuracy": 0.8595820367336273, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 1.384, | |
| "grad_norm": 19.089265823364258, | |
| "learning_rate": 3.303030303030303e-05, | |
| "loss": 2.9513, | |
| "mean_token_accuracy": 0.8398459106683731, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 1.392, | |
| "grad_norm": 18.12497901916504, | |
| "learning_rate": 3.292929292929293e-05, | |
| "loss": 2.8923, | |
| "mean_token_accuracy": 0.8398792594671249, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "grad_norm": 20.43326187133789, | |
| "learning_rate": 3.282828282828283e-05, | |
| "loss": 3.34, | |
| "mean_token_accuracy": 0.8242058306932449, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 1.408, | |
| "grad_norm": 22.64555549621582, | |
| "learning_rate": 3.272727272727273e-05, | |
| "loss": 3.4045, | |
| "mean_token_accuracy": 0.8219707757234573, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 1.416, | |
| "grad_norm": 19.150569915771484, | |
| "learning_rate": 3.2626262626262624e-05, | |
| "loss": 3.0239, | |
| "mean_token_accuracy": 0.8366655707359314, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 1.424, | |
| "grad_norm": 19.570180892944336, | |
| "learning_rate": 3.2525252525252524e-05, | |
| "loss": 3.1822, | |
| "mean_token_accuracy": 0.8368578106164932, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 1.432, | |
| "grad_norm": 19.738082885742188, | |
| "learning_rate": 3.2424242424242423e-05, | |
| "loss": 3.1123, | |
| "mean_token_accuracy": 0.83487968146801, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "grad_norm": 19.594087600708008, | |
| "learning_rate": 3.232323232323233e-05, | |
| "loss": 2.9401, | |
| "mean_token_accuracy": 0.8384669721126556, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 1.448, | |
| "grad_norm": 21.615190505981445, | |
| "learning_rate": 3.222222222222223e-05, | |
| "loss": 3.1546, | |
| "mean_token_accuracy": 0.8384737074375153, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 1.456, | |
| "grad_norm": 20.175748825073242, | |
| "learning_rate": 3.212121212121212e-05, | |
| "loss": 3.2557, | |
| "mean_token_accuracy": 0.8329617381095886, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 1.464, | |
| "grad_norm": 20.50943374633789, | |
| "learning_rate": 3.202020202020202e-05, | |
| "loss": 3.1785, | |
| "mean_token_accuracy": 0.8321622610092163, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 1.472, | |
| "grad_norm": 20.06108856201172, | |
| "learning_rate": 3.191919191919192e-05, | |
| "loss": 3.5482, | |
| "mean_token_accuracy": 0.8127909004688263, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "grad_norm": 18.441608428955078, | |
| "learning_rate": 3.181818181818182e-05, | |
| "loss": 3.1898, | |
| "mean_token_accuracy": 0.8352639973163605, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 1.488, | |
| "grad_norm": 18.315155029296875, | |
| "learning_rate": 3.171717171717172e-05, | |
| "loss": 3.0578, | |
| "mean_token_accuracy": 0.8414150923490524, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 1.496, | |
| "grad_norm": 19.802762985229492, | |
| "learning_rate": 3.161616161616161e-05, | |
| "loss": 3.4275, | |
| "mean_token_accuracy": 0.8302872627973557, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 1.504, | |
| "grad_norm": 17.772506713867188, | |
| "learning_rate": 3.151515151515151e-05, | |
| "loss": 2.5637, | |
| "mean_token_accuracy": 0.8702727556228638, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 1.512, | |
| "grad_norm": 16.841842651367188, | |
| "learning_rate": 3.141414141414142e-05, | |
| "loss": 2.7877, | |
| "mean_token_accuracy": 0.8445184230804443, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "grad_norm": 18.028287887573242, | |
| "learning_rate": 3.131313131313132e-05, | |
| "loss": 2.8235, | |
| "mean_token_accuracy": 0.8519050180912018, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 1.528, | |
| "grad_norm": 17.80512809753418, | |
| "learning_rate": 3.121212121212122e-05, | |
| "loss": 2.8722, | |
| "mean_token_accuracy": 0.8488593399524689, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 1.536, | |
| "grad_norm": 18.017911911010742, | |
| "learning_rate": 3.111111111111111e-05, | |
| "loss": 2.815, | |
| "mean_token_accuracy": 0.8511413782835007, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 1.544, | |
| "grad_norm": 16.424795150756836, | |
| "learning_rate": 3.101010101010101e-05, | |
| "loss": 2.6713, | |
| "mean_token_accuracy": 0.8467608839273453, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 1.552, | |
| "grad_norm": 18.472646713256836, | |
| "learning_rate": 3.090909090909091e-05, | |
| "loss": 3.0529, | |
| "mean_token_accuracy": 0.848779022693634, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "grad_norm": 19.97772216796875, | |
| "learning_rate": 3.080808080808081e-05, | |
| "loss": 3.2967, | |
| "mean_token_accuracy": 0.8236434757709503, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 1.568, | |
| "grad_norm": 18.231199264526367, | |
| "learning_rate": 3.070707070707071e-05, | |
| "loss": 2.7142, | |
| "mean_token_accuracy": 0.8442593514919281, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 1.576, | |
| "grad_norm": 18.801584243774414, | |
| "learning_rate": 3.060606060606061e-05, | |
| "loss": 2.8128, | |
| "mean_token_accuracy": 0.853375568985939, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 1.584, | |
| "grad_norm": 17.9156551361084, | |
| "learning_rate": 3.050505050505051e-05, | |
| "loss": 2.76, | |
| "mean_token_accuracy": 0.8531969636678696, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 1.592, | |
| "grad_norm": 20.197153091430664, | |
| "learning_rate": 3.0404040404040406e-05, | |
| "loss": 3.3442, | |
| "mean_token_accuracy": 0.820599153637886, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "grad_norm": 19.50455665588379, | |
| "learning_rate": 3.0303030303030306e-05, | |
| "loss": 3.1541, | |
| "mean_token_accuracy": 0.8246084749698639, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 1.608, | |
| "grad_norm": 18.644977569580078, | |
| "learning_rate": 3.0202020202020205e-05, | |
| "loss": 2.9864, | |
| "mean_token_accuracy": 0.8428880572319031, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 1.616, | |
| "grad_norm": 18.98227882385254, | |
| "learning_rate": 3.01010101010101e-05, | |
| "loss": 2.8429, | |
| "mean_token_accuracy": 0.8443266749382019, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 1.624, | |
| "grad_norm": 21.41233253479004, | |
| "learning_rate": 3e-05, | |
| "loss": 3.2002, | |
| "mean_token_accuracy": 0.8272002190351486, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 1.6320000000000001, | |
| "grad_norm": 18.17258071899414, | |
| "learning_rate": 2.98989898989899e-05, | |
| "loss": 2.9346, | |
| "mean_token_accuracy": 0.836423709988594, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 1.6400000000000001, | |
| "grad_norm": 18.524250030517578, | |
| "learning_rate": 2.9797979797979796e-05, | |
| "loss": 2.9417, | |
| "mean_token_accuracy": 0.8409391045570374, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 1.6480000000000001, | |
| "grad_norm": 19.283735275268555, | |
| "learning_rate": 2.96969696969697e-05, | |
| "loss": 2.6658, | |
| "mean_token_accuracy": 0.8474785685539246, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 1.6560000000000001, | |
| "grad_norm": 17.894819259643555, | |
| "learning_rate": 2.95959595959596e-05, | |
| "loss": 2.7392, | |
| "mean_token_accuracy": 0.845216765999794, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 1.6640000000000001, | |
| "grad_norm": 19.564115524291992, | |
| "learning_rate": 2.9494949494949498e-05, | |
| "loss": 3.4316, | |
| "mean_token_accuracy": 0.8185593038797379, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 1.6720000000000002, | |
| "grad_norm": 17.54917335510254, | |
| "learning_rate": 2.9393939393939394e-05, | |
| "loss": 3.0063, | |
| "mean_token_accuracy": 0.8485386520624161, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 1.6800000000000002, | |
| "grad_norm": 18.805709838867188, | |
| "learning_rate": 2.9292929292929294e-05, | |
| "loss": 3.2245, | |
| "mean_token_accuracy": 0.8334101438522339, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 1.688, | |
| "grad_norm": 16.054325103759766, | |
| "learning_rate": 2.9191919191919193e-05, | |
| "loss": 2.4667, | |
| "mean_token_accuracy": 0.8625558912754059, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 1.696, | |
| "grad_norm": 17.8482723236084, | |
| "learning_rate": 2.909090909090909e-05, | |
| "loss": 2.9457, | |
| "mean_token_accuracy": 0.8457487970590591, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 1.704, | |
| "grad_norm": 18.403047561645508, | |
| "learning_rate": 2.898989898989899e-05, | |
| "loss": 2.7452, | |
| "mean_token_accuracy": 0.8584460020065308, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 1.712, | |
| "grad_norm": 17.21164321899414, | |
| "learning_rate": 2.8888888888888888e-05, | |
| "loss": 2.4914, | |
| "mean_token_accuracy": 0.8537319600582123, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "grad_norm": 18.591094970703125, | |
| "learning_rate": 2.878787878787879e-05, | |
| "loss": 2.9312, | |
| "mean_token_accuracy": 0.8463590443134308, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 1.728, | |
| "grad_norm": 17.793514251708984, | |
| "learning_rate": 2.868686868686869e-05, | |
| "loss": 2.653, | |
| "mean_token_accuracy": 0.8544672429561615, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 1.736, | |
| "grad_norm": 16.645397186279297, | |
| "learning_rate": 2.8585858585858587e-05, | |
| "loss": 2.4487, | |
| "mean_token_accuracy": 0.8595747947692871, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 1.744, | |
| "grad_norm": 18.782949447631836, | |
| "learning_rate": 2.8484848484848486e-05, | |
| "loss": 3.1283, | |
| "mean_token_accuracy": 0.8335093259811401, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 1.752, | |
| "grad_norm": 19.578367233276367, | |
| "learning_rate": 2.8383838383838386e-05, | |
| "loss": 2.8753, | |
| "mean_token_accuracy": 0.8500068634748459, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "grad_norm": 20.041643142700195, | |
| "learning_rate": 2.8282828282828282e-05, | |
| "loss": 3.409, | |
| "mean_token_accuracy": 0.8245423436164856, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 1.768, | |
| "grad_norm": 19.69051170349121, | |
| "learning_rate": 2.818181818181818e-05, | |
| "loss": 3.0326, | |
| "mean_token_accuracy": 0.8391107320785522, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 1.776, | |
| "grad_norm": 18.727209091186523, | |
| "learning_rate": 2.808080808080808e-05, | |
| "loss": 2.8157, | |
| "mean_token_accuracy": 0.8591902405023575, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 1.784, | |
| "grad_norm": 16.66358184814453, | |
| "learning_rate": 2.7979797979797984e-05, | |
| "loss": 2.7767, | |
| "mean_token_accuracy": 0.8462460786104202, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 1.792, | |
| "grad_norm": 15.349605560302734, | |
| "learning_rate": 2.7878787878787883e-05, | |
| "loss": 2.3544, | |
| "mean_token_accuracy": 0.8616818338632584, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "grad_norm": 19.82066535949707, | |
| "learning_rate": 2.777777777777778e-05, | |
| "loss": 3.1691, | |
| "mean_token_accuracy": 0.8447847366333008, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 1.808, | |
| "grad_norm": 20.39141845703125, | |
| "learning_rate": 2.767676767676768e-05, | |
| "loss": 2.9149, | |
| "mean_token_accuracy": 0.8456564694643021, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 1.8159999999999998, | |
| "grad_norm": 18.638967514038086, | |
| "learning_rate": 2.7575757575757578e-05, | |
| "loss": 2.889, | |
| "mean_token_accuracy": 0.8354881256818771, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 1.8239999999999998, | |
| "grad_norm": 17.088668823242188, | |
| "learning_rate": 2.7474747474747474e-05, | |
| "loss": 2.6078, | |
| "mean_token_accuracy": 0.8517204225063324, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 1.8319999999999999, | |
| "grad_norm": 18.29672622680664, | |
| "learning_rate": 2.7373737373737374e-05, | |
| "loss": 3.0306, | |
| "mean_token_accuracy": 0.8434745818376541, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 1.8399999999999999, | |
| "grad_norm": 19.094703674316406, | |
| "learning_rate": 2.7272727272727273e-05, | |
| "loss": 2.8052, | |
| "mean_token_accuracy": 0.8540522456169128, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 1.8479999999999999, | |
| "grad_norm": 18.637916564941406, | |
| "learning_rate": 2.717171717171717e-05, | |
| "loss": 2.7584, | |
| "mean_token_accuracy": 0.8538561910390854, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 1.8559999999999999, | |
| "grad_norm": 18.355613708496094, | |
| "learning_rate": 2.7070707070707075e-05, | |
| "loss": 3.1927, | |
| "mean_token_accuracy": 0.8381341099739075, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 1.8639999999999999, | |
| "grad_norm": 24.347230911254883, | |
| "learning_rate": 2.696969696969697e-05, | |
| "loss": 3.353, | |
| "mean_token_accuracy": 0.8150720745325089, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 1.8719999999999999, | |
| "grad_norm": 19.14170265197754, | |
| "learning_rate": 2.686868686868687e-05, | |
| "loss": 3.4434, | |
| "mean_token_accuracy": 0.8221272975206375, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "grad_norm": 16.074077606201172, | |
| "learning_rate": 2.676767676767677e-05, | |
| "loss": 2.6817, | |
| "mean_token_accuracy": 0.856314018368721, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 1.888, | |
| "grad_norm": 20.647747039794922, | |
| "learning_rate": 2.6666666666666667e-05, | |
| "loss": 3.2522, | |
| "mean_token_accuracy": 0.822910264134407, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 1.896, | |
| "grad_norm": 17.33372688293457, | |
| "learning_rate": 2.6565656565656566e-05, | |
| "loss": 2.878, | |
| "mean_token_accuracy": 0.8544587790966034, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 1.904, | |
| "grad_norm": 21.383373260498047, | |
| "learning_rate": 2.6464646464646466e-05, | |
| "loss": 3.0065, | |
| "mean_token_accuracy": 0.8334675282239914, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 1.912, | |
| "grad_norm": 17.983051300048828, | |
| "learning_rate": 2.636363636363636e-05, | |
| "loss": 2.6445, | |
| "mean_token_accuracy": 0.8565590083599091, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "grad_norm": 19.193273544311523, | |
| "learning_rate": 2.6262626262626268e-05, | |
| "loss": 2.8959, | |
| "mean_token_accuracy": 0.8444470465183258, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 1.928, | |
| "grad_norm": 18.387603759765625, | |
| "learning_rate": 2.6161616161616164e-05, | |
| "loss": 3.031, | |
| "mean_token_accuracy": 0.8487697094678879, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 1.936, | |
| "grad_norm": 17.07048988342285, | |
| "learning_rate": 2.6060606060606063e-05, | |
| "loss": 2.3949, | |
| "mean_token_accuracy": 0.8728950917720795, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 1.944, | |
| "grad_norm": 16.93642234802246, | |
| "learning_rate": 2.5959595959595963e-05, | |
| "loss": 2.8926, | |
| "mean_token_accuracy": 0.8463722467422485, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 1.952, | |
| "grad_norm": 18.29349136352539, | |
| "learning_rate": 2.585858585858586e-05, | |
| "loss": 3.0535, | |
| "mean_token_accuracy": 0.8376583307981491, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "grad_norm": 18.56781768798828, | |
| "learning_rate": 2.575757575757576e-05, | |
| "loss": 2.8888, | |
| "mean_token_accuracy": 0.8465193212032318, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 1.968, | |
| "grad_norm": 18.541624069213867, | |
| "learning_rate": 2.5656565656565658e-05, | |
| "loss": 2.3475, | |
| "mean_token_accuracy": 0.8656208217144012, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 1.976, | |
| "grad_norm": 17.897377014160156, | |
| "learning_rate": 2.5555555555555554e-05, | |
| "loss": 2.837, | |
| "mean_token_accuracy": 0.8604338765144348, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 1.984, | |
| "grad_norm": 18.500314712524414, | |
| "learning_rate": 2.5454545454545454e-05, | |
| "loss": 2.6232, | |
| "mean_token_accuracy": 0.859655350446701, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 1.992, | |
| "grad_norm": 18.447683334350586, | |
| "learning_rate": 2.5353535353535356e-05, | |
| "loss": 3.01, | |
| "mean_token_accuracy": 0.8308267742395401, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "grad_norm": 16.8708438873291, | |
| "learning_rate": 2.5252525252525256e-05, | |
| "loss": 2.81, | |
| "mean_token_accuracy": 0.8507586419582367, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 2.008, | |
| "grad_norm": 14.083173751831055, | |
| "learning_rate": 2.5151515151515155e-05, | |
| "loss": 1.5615, | |
| "mean_token_accuracy": 0.9069856852293015, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 2.016, | |
| "grad_norm": 15.285985946655273, | |
| "learning_rate": 2.505050505050505e-05, | |
| "loss": 1.7983, | |
| "mean_token_accuracy": 0.8917003720998764, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 2.024, | |
| "grad_norm": 14.96445083618164, | |
| "learning_rate": 2.494949494949495e-05, | |
| "loss": 1.6866, | |
| "mean_token_accuracy": 0.8935637921094894, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 2.032, | |
| "grad_norm": 15.746031761169434, | |
| "learning_rate": 2.4848484848484847e-05, | |
| "loss": 1.7632, | |
| "mean_token_accuracy": 0.8932289183139801, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "grad_norm": 16.720212936401367, | |
| "learning_rate": 2.474747474747475e-05, | |
| "loss": 1.7424, | |
| "mean_token_accuracy": 0.8972940593957901, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 2.048, | |
| "grad_norm": 15.393232345581055, | |
| "learning_rate": 2.464646464646465e-05, | |
| "loss": 1.5677, | |
| "mean_token_accuracy": 0.9046053141355515, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 2.056, | |
| "grad_norm": 17.94659996032715, | |
| "learning_rate": 2.4545454545454545e-05, | |
| "loss": 1.9221, | |
| "mean_token_accuracy": 0.8920014947652817, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 2.064, | |
| "grad_norm": 18.840778350830078, | |
| "learning_rate": 2.4444444444444445e-05, | |
| "loss": 1.7952, | |
| "mean_token_accuracy": 0.8805793821811676, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 2.072, | |
| "grad_norm": 17.553537368774414, | |
| "learning_rate": 2.4343434343434344e-05, | |
| "loss": 1.6415, | |
| "mean_token_accuracy": 0.9035749733448029, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "grad_norm": 16.594493865966797, | |
| "learning_rate": 2.4242424242424244e-05, | |
| "loss": 1.4085, | |
| "mean_token_accuracy": 0.9104350358247757, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 2.088, | |
| "grad_norm": 18.47384262084961, | |
| "learning_rate": 2.4141414141414143e-05, | |
| "loss": 1.7969, | |
| "mean_token_accuracy": 0.893615260720253, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 2.096, | |
| "grad_norm": 20.94403839111328, | |
| "learning_rate": 2.404040404040404e-05, | |
| "loss": 2.0246, | |
| "mean_token_accuracy": 0.8807271867990494, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 2.104, | |
| "grad_norm": 18.5487117767334, | |
| "learning_rate": 2.393939393939394e-05, | |
| "loss": 1.5092, | |
| "mean_token_accuracy": 0.9083298295736313, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 2.112, | |
| "grad_norm": 20.118314743041992, | |
| "learning_rate": 2.3838383838383842e-05, | |
| "loss": 1.7474, | |
| "mean_token_accuracy": 0.8932338207960129, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "grad_norm": 19.099689483642578, | |
| "learning_rate": 2.3737373737373738e-05, | |
| "loss": 1.8237, | |
| "mean_token_accuracy": 0.8876322358846664, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 2.128, | |
| "grad_norm": 20.295059204101562, | |
| "learning_rate": 2.3636363636363637e-05, | |
| "loss": 1.9511, | |
| "mean_token_accuracy": 0.8856483995914459, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 2.136, | |
| "grad_norm": 21.094022750854492, | |
| "learning_rate": 2.3535353535353537e-05, | |
| "loss": 1.7983, | |
| "mean_token_accuracy": 0.8887585699558258, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 2.144, | |
| "grad_norm": 19.475961685180664, | |
| "learning_rate": 2.3434343434343436e-05, | |
| "loss": 1.8121, | |
| "mean_token_accuracy": 0.8935129791498184, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 2.152, | |
| "grad_norm": 17.48745346069336, | |
| "learning_rate": 2.3333333333333336e-05, | |
| "loss": 1.6347, | |
| "mean_token_accuracy": 0.901911199092865, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "grad_norm": 19.021291732788086, | |
| "learning_rate": 2.3232323232323232e-05, | |
| "loss": 1.6945, | |
| "mean_token_accuracy": 0.8956849575042725, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 2.168, | |
| "grad_norm": 18.498403549194336, | |
| "learning_rate": 2.313131313131313e-05, | |
| "loss": 1.8713, | |
| "mean_token_accuracy": 0.8862475454807281, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 2.176, | |
| "grad_norm": 20.616243362426758, | |
| "learning_rate": 2.3030303030303034e-05, | |
| "loss": 1.9077, | |
| "mean_token_accuracy": 0.8816216289997101, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 2.184, | |
| "grad_norm": 18.55655288696289, | |
| "learning_rate": 2.292929292929293e-05, | |
| "loss": 1.893, | |
| "mean_token_accuracy": 0.8819779455661774, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 2.192, | |
| "grad_norm": 19.719331741333008, | |
| "learning_rate": 2.282828282828283e-05, | |
| "loss": 1.7443, | |
| "mean_token_accuracy": 0.8999358415603638, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "grad_norm": 15.951658248901367, | |
| "learning_rate": 2.272727272727273e-05, | |
| "loss": 1.5942, | |
| "mean_token_accuracy": 0.9009282290935516, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 2.208, | |
| "grad_norm": 17.43150520324707, | |
| "learning_rate": 2.262626262626263e-05, | |
| "loss": 1.597, | |
| "mean_token_accuracy": 0.8990740329027176, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 2.216, | |
| "grad_norm": 18.600112915039062, | |
| "learning_rate": 2.2525252525252528e-05, | |
| "loss": 1.8142, | |
| "mean_token_accuracy": 0.891479030251503, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 2.224, | |
| "grad_norm": 18.331689834594727, | |
| "learning_rate": 2.2424242424242424e-05, | |
| "loss": 1.6484, | |
| "mean_token_accuracy": 0.8858047872781754, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 2.232, | |
| "grad_norm": 17.937702178955078, | |
| "learning_rate": 2.2323232323232324e-05, | |
| "loss": 1.6012, | |
| "mean_token_accuracy": 0.8977932035923004, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "grad_norm": 18.553834915161133, | |
| "learning_rate": 2.2222222222222223e-05, | |
| "loss": 1.507, | |
| "mean_token_accuracy": 0.9132246375083923, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 2.248, | |
| "grad_norm": 17.619672775268555, | |
| "learning_rate": 2.2121212121212123e-05, | |
| "loss": 1.5574, | |
| "mean_token_accuracy": 0.903651237487793, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 2.2560000000000002, | |
| "grad_norm": 18.53679084777832, | |
| "learning_rate": 2.2020202020202022e-05, | |
| "loss": 1.8894, | |
| "mean_token_accuracy": 0.8831851631402969, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 2.2640000000000002, | |
| "grad_norm": 18.827486038208008, | |
| "learning_rate": 2.191919191919192e-05, | |
| "loss": 1.8957, | |
| "mean_token_accuracy": 0.8890838772058487, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 2.2720000000000002, | |
| "grad_norm": 18.129066467285156, | |
| "learning_rate": 2.1818181818181818e-05, | |
| "loss": 1.7498, | |
| "mean_token_accuracy": 0.8904573619365692, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 2.2800000000000002, | |
| "grad_norm": 20.051956176757812, | |
| "learning_rate": 2.171717171717172e-05, | |
| "loss": 1.8336, | |
| "mean_token_accuracy": 0.8877034038305283, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 2.288, | |
| "grad_norm": 18.455982208251953, | |
| "learning_rate": 2.1616161616161617e-05, | |
| "loss": 1.7406, | |
| "mean_token_accuracy": 0.8882922381162643, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 2.296, | |
| "grad_norm": 16.913650512695312, | |
| "learning_rate": 2.1515151515151516e-05, | |
| "loss": 1.5838, | |
| "mean_token_accuracy": 0.8983030468225479, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 2.304, | |
| "grad_norm": 16.81460952758789, | |
| "learning_rate": 2.1414141414141416e-05, | |
| "loss": 1.5973, | |
| "mean_token_accuracy": 0.9012346267700195, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 2.312, | |
| "grad_norm": 18.44367790222168, | |
| "learning_rate": 2.1313131313131315e-05, | |
| "loss": 1.8431, | |
| "mean_token_accuracy": 0.8918169885873795, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "grad_norm": 17.587459564208984, | |
| "learning_rate": 2.1212121212121215e-05, | |
| "loss": 1.7482, | |
| "mean_token_accuracy": 0.883071169257164, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 2.328, | |
| "grad_norm": 20.508304595947266, | |
| "learning_rate": 2.111111111111111e-05, | |
| "loss": 2.1301, | |
| "mean_token_accuracy": 0.8722483068704605, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 2.336, | |
| "grad_norm": 21.067611694335938, | |
| "learning_rate": 2.101010101010101e-05, | |
| "loss": 1.8906, | |
| "mean_token_accuracy": 0.8874034583568573, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 2.344, | |
| "grad_norm": 17.23782730102539, | |
| "learning_rate": 2.090909090909091e-05, | |
| "loss": 1.5809, | |
| "mean_token_accuracy": 0.9070325046777725, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 2.352, | |
| "grad_norm": 17.424673080444336, | |
| "learning_rate": 2.080808080808081e-05, | |
| "loss": 1.7886, | |
| "mean_token_accuracy": 0.8876490592956543, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "grad_norm": 18.26118278503418, | |
| "learning_rate": 2.070707070707071e-05, | |
| "loss": 1.6298, | |
| "mean_token_accuracy": 0.8885916471481323, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 2.368, | |
| "grad_norm": 19.218050003051758, | |
| "learning_rate": 2.0606060606060608e-05, | |
| "loss": 1.9432, | |
| "mean_token_accuracy": 0.8848688006401062, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 2.376, | |
| "grad_norm": 18.870819091796875, | |
| "learning_rate": 2.0505050505050504e-05, | |
| "loss": 1.7683, | |
| "mean_token_accuracy": 0.889656737446785, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 2.384, | |
| "grad_norm": 17.928733825683594, | |
| "learning_rate": 2.0404040404040407e-05, | |
| "loss": 1.8347, | |
| "mean_token_accuracy": 0.8826514780521393, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 2.392, | |
| "grad_norm": 17.18740463256836, | |
| "learning_rate": 2.0303030303030303e-05, | |
| "loss": 1.7417, | |
| "mean_token_accuracy": 0.8983327001333237, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "grad_norm": 17.51845932006836, | |
| "learning_rate": 2.0202020202020203e-05, | |
| "loss": 1.5803, | |
| "mean_token_accuracy": 0.9036453068256378, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 2.408, | |
| "grad_norm": 18.641651153564453, | |
| "learning_rate": 2.0101010101010102e-05, | |
| "loss": 1.9701, | |
| "mean_token_accuracy": 0.8859322518110275, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 2.416, | |
| "grad_norm": 19.06338119506836, | |
| "learning_rate": 2e-05, | |
| "loss": 1.593, | |
| "mean_token_accuracy": 0.8942733258008957, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 2.424, | |
| "grad_norm": 17.082141876220703, | |
| "learning_rate": 1.98989898989899e-05, | |
| "loss": 1.5682, | |
| "mean_token_accuracy": 0.8980588763952255, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 2.432, | |
| "grad_norm": 18.56012725830078, | |
| "learning_rate": 1.9797979797979797e-05, | |
| "loss": 1.9395, | |
| "mean_token_accuracy": 0.8804908245801926, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "grad_norm": 20.00580406188965, | |
| "learning_rate": 1.9696969696969697e-05, | |
| "loss": 1.8976, | |
| "mean_token_accuracy": 0.8813226372003555, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 2.448, | |
| "grad_norm": 17.66527557373047, | |
| "learning_rate": 1.95959595959596e-05, | |
| "loss": 1.7429, | |
| "mean_token_accuracy": 0.8991499990224838, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 2.456, | |
| "grad_norm": 22.222915649414062, | |
| "learning_rate": 1.9494949494949496e-05, | |
| "loss": 2.0225, | |
| "mean_token_accuracy": 0.8786364048719406, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 2.464, | |
| "grad_norm": 15.038418769836426, | |
| "learning_rate": 1.9393939393939395e-05, | |
| "loss": 1.5893, | |
| "mean_token_accuracy": 0.9014905393123627, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 2.472, | |
| "grad_norm": 16.89600944519043, | |
| "learning_rate": 1.9292929292929295e-05, | |
| "loss": 1.7931, | |
| "mean_token_accuracy": 0.8878951072692871, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "grad_norm": 18.997501373291016, | |
| "learning_rate": 1.919191919191919e-05, | |
| "loss": 1.7395, | |
| "mean_token_accuracy": 0.8750288486480713, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 2.488, | |
| "grad_norm": 17.06654930114746, | |
| "learning_rate": 1.9090909090909094e-05, | |
| "loss": 1.7309, | |
| "mean_token_accuracy": 0.8866889774799347, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 2.496, | |
| "grad_norm": 16.174907684326172, | |
| "learning_rate": 1.898989898989899e-05, | |
| "loss": 1.6361, | |
| "mean_token_accuracy": 0.8998618721961975, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 2.504, | |
| "grad_norm": 20.149904251098633, | |
| "learning_rate": 1.888888888888889e-05, | |
| "loss": 1.651, | |
| "mean_token_accuracy": 0.894085094332695, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 2.512, | |
| "grad_norm": 17.874101638793945, | |
| "learning_rate": 1.878787878787879e-05, | |
| "loss": 1.651, | |
| "mean_token_accuracy": 0.8896775543689728, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "grad_norm": 18.547687530517578, | |
| "learning_rate": 1.8686868686868688e-05, | |
| "loss": 1.9905, | |
| "mean_token_accuracy": 0.8854596465826035, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 2.528, | |
| "grad_norm": 19.14983367919922, | |
| "learning_rate": 1.8585858585858588e-05, | |
| "loss": 1.8654, | |
| "mean_token_accuracy": 0.884912833571434, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 2.536, | |
| "grad_norm": 18.558279037475586, | |
| "learning_rate": 1.8484848484848487e-05, | |
| "loss": 1.7852, | |
| "mean_token_accuracy": 0.8943625837564468, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 2.544, | |
| "grad_norm": 17.757083892822266, | |
| "learning_rate": 1.8383838383838383e-05, | |
| "loss": 1.5901, | |
| "mean_token_accuracy": 0.9012947678565979, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 2.552, | |
| "grad_norm": 17.890913009643555, | |
| "learning_rate": 1.8282828282828286e-05, | |
| "loss": 1.8079, | |
| "mean_token_accuracy": 0.8911891281604767, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "grad_norm": 16.844127655029297, | |
| "learning_rate": 1.8181818181818182e-05, | |
| "loss": 1.6101, | |
| "mean_token_accuracy": 0.9100003689527512, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 2.568, | |
| "grad_norm": 20.43346405029297, | |
| "learning_rate": 1.808080808080808e-05, | |
| "loss": 1.9681, | |
| "mean_token_accuracy": 0.8885725736618042, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 2.576, | |
| "grad_norm": 19.34762191772461, | |
| "learning_rate": 1.797979797979798e-05, | |
| "loss": 1.9601, | |
| "mean_token_accuracy": 0.8918892741203308, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 2.584, | |
| "grad_norm": 16.914316177368164, | |
| "learning_rate": 1.787878787878788e-05, | |
| "loss": 1.587, | |
| "mean_token_accuracy": 0.8948457092046738, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 2.592, | |
| "grad_norm": 18.791858673095703, | |
| "learning_rate": 1.777777777777778e-05, | |
| "loss": 1.8718, | |
| "mean_token_accuracy": 0.8815629184246063, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "grad_norm": 17.461881637573242, | |
| "learning_rate": 1.7676767676767676e-05, | |
| "loss": 1.6116, | |
| "mean_token_accuracy": 0.8986152410507202, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 2.608, | |
| "grad_norm": 19.637073516845703, | |
| "learning_rate": 1.7575757575757576e-05, | |
| "loss": 1.8468, | |
| "mean_token_accuracy": 0.8987504243850708, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 2.616, | |
| "grad_norm": 19.572315216064453, | |
| "learning_rate": 1.7474747474747475e-05, | |
| "loss": 1.8924, | |
| "mean_token_accuracy": 0.8777202218770981, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 2.624, | |
| "grad_norm": 19.34830665588379, | |
| "learning_rate": 1.7373737373737375e-05, | |
| "loss": 1.8247, | |
| "mean_token_accuracy": 0.8945279717445374, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 2.632, | |
| "grad_norm": 16.744380950927734, | |
| "learning_rate": 1.7272727272727274e-05, | |
| "loss": 1.615, | |
| "mean_token_accuracy": 0.8965769708156586, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "grad_norm": 16.65348243713379, | |
| "learning_rate": 1.7171717171717173e-05, | |
| "loss": 1.7394, | |
| "mean_token_accuracy": 0.8977630883455276, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 2.648, | |
| "grad_norm": 19.861833572387695, | |
| "learning_rate": 1.707070707070707e-05, | |
| "loss": 1.9065, | |
| "mean_token_accuracy": 0.8806384950876236, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 2.656, | |
| "grad_norm": 18.7454891204834, | |
| "learning_rate": 1.6969696969696972e-05, | |
| "loss": 1.7408, | |
| "mean_token_accuracy": 0.8817498087882996, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 2.664, | |
| "grad_norm": 18.72283935546875, | |
| "learning_rate": 1.686868686868687e-05, | |
| "loss": 1.6119, | |
| "mean_token_accuracy": 0.8949918895959854, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 2.672, | |
| "grad_norm": 17.392744064331055, | |
| "learning_rate": 1.6767676767676768e-05, | |
| "loss": 1.7595, | |
| "mean_token_accuracy": 0.8967541307210922, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "grad_norm": 17.753887176513672, | |
| "learning_rate": 1.6666666666666667e-05, | |
| "loss": 1.867, | |
| "mean_token_accuracy": 0.8946125656366348, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 2.6879999999999997, | |
| "grad_norm": 15.396509170532227, | |
| "learning_rate": 1.6565656565656567e-05, | |
| "loss": 1.4533, | |
| "mean_token_accuracy": 0.902470737695694, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 2.6959999999999997, | |
| "grad_norm": 17.420621871948242, | |
| "learning_rate": 1.6464646464646466e-05, | |
| "loss": 1.8612, | |
| "mean_token_accuracy": 0.8907473981380463, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 2.7039999999999997, | |
| "grad_norm": 16.46539306640625, | |
| "learning_rate": 1.6363636363636366e-05, | |
| "loss": 1.8107, | |
| "mean_token_accuracy": 0.8965497016906738, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 2.7119999999999997, | |
| "grad_norm": 18.55117416381836, | |
| "learning_rate": 1.6262626262626262e-05, | |
| "loss": 1.6489, | |
| "mean_token_accuracy": 0.8972436189651489, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 2.7199999999999998, | |
| "grad_norm": 20.577390670776367, | |
| "learning_rate": 1.6161616161616165e-05, | |
| "loss": 1.7792, | |
| "mean_token_accuracy": 0.8870959579944611, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 2.7279999999999998, | |
| "grad_norm": 16.724843978881836, | |
| "learning_rate": 1.606060606060606e-05, | |
| "loss": 1.7987, | |
| "mean_token_accuracy": 0.896402895450592, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 2.7359999999999998, | |
| "grad_norm": 19.218793869018555, | |
| "learning_rate": 1.595959595959596e-05, | |
| "loss": 1.9317, | |
| "mean_token_accuracy": 0.8903376162052155, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 2.7439999999999998, | |
| "grad_norm": 17.59774398803711, | |
| "learning_rate": 1.585858585858586e-05, | |
| "loss": 1.7176, | |
| "mean_token_accuracy": 0.9004463106393814, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 2.752, | |
| "grad_norm": 18.768749237060547, | |
| "learning_rate": 1.5757575757575756e-05, | |
| "loss": 1.772, | |
| "mean_token_accuracy": 0.8775418102741241, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "grad_norm": 19.772340774536133, | |
| "learning_rate": 1.565656565656566e-05, | |
| "loss": 1.8746, | |
| "mean_token_accuracy": 0.8900909423828125, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 2.768, | |
| "grad_norm": 17.75616455078125, | |
| "learning_rate": 1.5555555555555555e-05, | |
| "loss": 1.6445, | |
| "mean_token_accuracy": 0.8957866132259369, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 2.776, | |
| "grad_norm": 17.134241104125977, | |
| "learning_rate": 1.5454545454545454e-05, | |
| "loss": 1.6659, | |
| "mean_token_accuracy": 0.9030885994434357, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 2.784, | |
| "grad_norm": 16.95988655090332, | |
| "learning_rate": 1.5353535353535354e-05, | |
| "loss": 1.5641, | |
| "mean_token_accuracy": 0.8961215913295746, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 2.792, | |
| "grad_norm": 18.894676208496094, | |
| "learning_rate": 1.5252525252525255e-05, | |
| "loss": 1.6115, | |
| "mean_token_accuracy": 0.9038358479738235, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "grad_norm": 18.057437896728516, | |
| "learning_rate": 1.5151515151515153e-05, | |
| "loss": 1.8048, | |
| "mean_token_accuracy": 0.8895912170410156, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 2.808, | |
| "grad_norm": 16.97344398498535, | |
| "learning_rate": 1.505050505050505e-05, | |
| "loss": 1.7005, | |
| "mean_token_accuracy": 0.9056167453527451, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 2.816, | |
| "grad_norm": 18.987558364868164, | |
| "learning_rate": 1.494949494949495e-05, | |
| "loss": 1.9392, | |
| "mean_token_accuracy": 0.8795923292636871, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 2.824, | |
| "grad_norm": 19.48893165588379, | |
| "learning_rate": 1.484848484848485e-05, | |
| "loss": 1.884, | |
| "mean_token_accuracy": 0.8816027939319611, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 2.832, | |
| "grad_norm": 16.16484260559082, | |
| "learning_rate": 1.4747474747474749e-05, | |
| "loss": 1.3967, | |
| "mean_token_accuracy": 0.9068291187286377, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "grad_norm": 16.221166610717773, | |
| "learning_rate": 1.4646464646464647e-05, | |
| "loss": 1.6643, | |
| "mean_token_accuracy": 0.8908949345350266, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 2.848, | |
| "grad_norm": 20.419599533081055, | |
| "learning_rate": 1.4545454545454545e-05, | |
| "loss": 1.9597, | |
| "mean_token_accuracy": 0.8768892884254456, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 2.856, | |
| "grad_norm": 21.398174285888672, | |
| "learning_rate": 1.4444444444444444e-05, | |
| "loss": 1.8647, | |
| "mean_token_accuracy": 0.882171094417572, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 2.864, | |
| "grad_norm": 20.366880416870117, | |
| "learning_rate": 1.4343434343434345e-05, | |
| "loss": 1.7834, | |
| "mean_token_accuracy": 0.8946562111377716, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 2.872, | |
| "grad_norm": 18.22726821899414, | |
| "learning_rate": 1.4242424242424243e-05, | |
| "loss": 1.7605, | |
| "mean_token_accuracy": 0.901305228471756, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "grad_norm": 19.751203536987305, | |
| "learning_rate": 1.4141414141414141e-05, | |
| "loss": 1.6872, | |
| "mean_token_accuracy": 0.8890763968229294, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 2.888, | |
| "grad_norm": 15.955531120300293, | |
| "learning_rate": 1.404040404040404e-05, | |
| "loss": 1.6292, | |
| "mean_token_accuracy": 0.9044130593538284, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 2.896, | |
| "grad_norm": 17.15973472595215, | |
| "learning_rate": 1.3939393939393942e-05, | |
| "loss": 1.5112, | |
| "mean_token_accuracy": 0.9043047428131104, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 2.904, | |
| "grad_norm": 20.14824676513672, | |
| "learning_rate": 1.383838383838384e-05, | |
| "loss": 2.1377, | |
| "mean_token_accuracy": 0.874171257019043, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 2.912, | |
| "grad_norm": 17.76668930053711, | |
| "learning_rate": 1.3737373737373737e-05, | |
| "loss": 1.815, | |
| "mean_token_accuracy": 0.8899873048067093, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "grad_norm": 17.47455596923828, | |
| "learning_rate": 1.3636363636363637e-05, | |
| "loss": 1.8602, | |
| "mean_token_accuracy": 0.8958509713411331, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 2.928, | |
| "grad_norm": 18.125408172607422, | |
| "learning_rate": 1.3535353535353538e-05, | |
| "loss": 1.7344, | |
| "mean_token_accuracy": 0.8926298767328262, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 2.936, | |
| "grad_norm": 18.787460327148438, | |
| "learning_rate": 1.3434343434343436e-05, | |
| "loss": 1.8846, | |
| "mean_token_accuracy": 0.8847126960754395, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 2.944, | |
| "grad_norm": 17.53483009338379, | |
| "learning_rate": 1.3333333333333333e-05, | |
| "loss": 1.7037, | |
| "mean_token_accuracy": 0.8964135348796844, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 2.952, | |
| "grad_norm": 17.428064346313477, | |
| "learning_rate": 1.3232323232323233e-05, | |
| "loss": 1.7683, | |
| "mean_token_accuracy": 0.8908781260251999, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "grad_norm": 17.09117889404297, | |
| "learning_rate": 1.3131313131313134e-05, | |
| "loss": 1.6467, | |
| "mean_token_accuracy": 0.8997662663459778, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 2.968, | |
| "grad_norm": 18.157711029052734, | |
| "learning_rate": 1.3030303030303032e-05, | |
| "loss": 2.0013, | |
| "mean_token_accuracy": 0.878508061170578, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 2.976, | |
| "grad_norm": 17.718358993530273, | |
| "learning_rate": 1.292929292929293e-05, | |
| "loss": 1.7484, | |
| "mean_token_accuracy": 0.8951921463012695, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 2.984, | |
| "grad_norm": 19.314804077148438, | |
| "learning_rate": 1.2828282828282829e-05, | |
| "loss": 1.8593, | |
| "mean_token_accuracy": 0.8909815549850464, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 2.992, | |
| "grad_norm": 15.990935325622559, | |
| "learning_rate": 1.2727272727272727e-05, | |
| "loss": 1.7345, | |
| "mean_token_accuracy": 0.8998604416847229, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "grad_norm": 17.464195251464844, | |
| "learning_rate": 1.2626262626262628e-05, | |
| "loss": 1.9088, | |
| "mean_token_accuracy": 0.8819395452737808, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 3.008, | |
| "grad_norm": 11.721142768859863, | |
| "learning_rate": 1.2525252525252526e-05, | |
| "loss": 1.0405, | |
| "mean_token_accuracy": 0.9373282045125961, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 3.016, | |
| "grad_norm": 14.365158081054688, | |
| "learning_rate": 1.2424242424242424e-05, | |
| "loss": 1.3254, | |
| "mean_token_accuracy": 0.9172510653734207, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 3.024, | |
| "grad_norm": 12.863395690917969, | |
| "learning_rate": 1.2323232323232325e-05, | |
| "loss": 1.0233, | |
| "mean_token_accuracy": 0.9359502047300339, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 3.032, | |
| "grad_norm": 13.299514770507812, | |
| "learning_rate": 1.2222222222222222e-05, | |
| "loss": 1.0952, | |
| "mean_token_accuracy": 0.9341908991336823, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 3.04, | |
| "grad_norm": 15.089731216430664, | |
| "learning_rate": 1.2121212121212122e-05, | |
| "loss": 1.1399, | |
| "mean_token_accuracy": 0.9282894879579544, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 3.048, | |
| "grad_norm": 13.675347328186035, | |
| "learning_rate": 1.202020202020202e-05, | |
| "loss": 1.1441, | |
| "mean_token_accuracy": 0.9303115904331207, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 3.056, | |
| "grad_norm": 14.102622985839844, | |
| "learning_rate": 1.1919191919191921e-05, | |
| "loss": 1.0362, | |
| "mean_token_accuracy": 0.9242476671934128, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 3.064, | |
| "grad_norm": 13.803597450256348, | |
| "learning_rate": 1.1818181818181819e-05, | |
| "loss": 1.0437, | |
| "mean_token_accuracy": 0.9370378255844116, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 3.072, | |
| "grad_norm": 13.217544555664062, | |
| "learning_rate": 1.1717171717171718e-05, | |
| "loss": 0.9713, | |
| "mean_token_accuracy": 0.9372242540121078, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 3.08, | |
| "grad_norm": 14.189465522766113, | |
| "learning_rate": 1.1616161616161616e-05, | |
| "loss": 0.947, | |
| "mean_token_accuracy": 0.9372307360172272, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 3.088, | |
| "grad_norm": 18.668476104736328, | |
| "learning_rate": 1.1515151515151517e-05, | |
| "loss": 1.0271, | |
| "mean_token_accuracy": 0.9349544644355774, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 3.096, | |
| "grad_norm": 15.80893325805664, | |
| "learning_rate": 1.1414141414141415e-05, | |
| "loss": 0.9894, | |
| "mean_token_accuracy": 0.9350813329219818, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 3.104, | |
| "grad_norm": 13.34670639038086, | |
| "learning_rate": 1.1313131313131314e-05, | |
| "loss": 0.8648, | |
| "mean_token_accuracy": 0.9458875060081482, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 3.112, | |
| "grad_norm": 13.572864532470703, | |
| "learning_rate": 1.1212121212121212e-05, | |
| "loss": 0.8647, | |
| "mean_token_accuracy": 0.9380109906196594, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 3.12, | |
| "grad_norm": 14.114604949951172, | |
| "learning_rate": 1.1111111111111112e-05, | |
| "loss": 0.9754, | |
| "mean_token_accuracy": 0.9376205205917358, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 3.128, | |
| "grad_norm": 17.51618003845215, | |
| "learning_rate": 1.1010101010101011e-05, | |
| "loss": 1.1896, | |
| "mean_token_accuracy": 0.9313846081495285, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 3.136, | |
| "grad_norm": 16.113466262817383, | |
| "learning_rate": 1.0909090909090909e-05, | |
| "loss": 0.9537, | |
| "mean_token_accuracy": 0.9409548491239548, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 3.144, | |
| "grad_norm": 16.324718475341797, | |
| "learning_rate": 1.0808080808080808e-05, | |
| "loss": 1.0058, | |
| "mean_token_accuracy": 0.9435614198446274, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 3.152, | |
| "grad_norm": 14.306757926940918, | |
| "learning_rate": 1.0707070707070708e-05, | |
| "loss": 0.8795, | |
| "mean_token_accuracy": 0.9390368461608887, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 3.16, | |
| "grad_norm": 17.119672775268555, | |
| "learning_rate": 1.0606060606060607e-05, | |
| "loss": 1.0024, | |
| "mean_token_accuracy": 0.9306483864784241, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 3.168, | |
| "grad_norm": 16.017135620117188, | |
| "learning_rate": 1.0505050505050505e-05, | |
| "loss": 0.9437, | |
| "mean_token_accuracy": 0.9379599690437317, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 3.176, | |
| "grad_norm": 14.418791770935059, | |
| "learning_rate": 1.0404040404040405e-05, | |
| "loss": 0.8662, | |
| "mean_token_accuracy": 0.9417054504156113, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 3.184, | |
| "grad_norm": 18.20338249206543, | |
| "learning_rate": 1.0303030303030304e-05, | |
| "loss": 1.048, | |
| "mean_token_accuracy": 0.9236321002244949, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 3.192, | |
| "grad_norm": 16.065502166748047, | |
| "learning_rate": 1.0202020202020204e-05, | |
| "loss": 0.9848, | |
| "mean_token_accuracy": 0.9392602741718292, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 3.2, | |
| "grad_norm": 15.719825744628906, | |
| "learning_rate": 1.0101010101010101e-05, | |
| "loss": 0.9635, | |
| "mean_token_accuracy": 0.9424310624599457, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 3.208, | |
| "grad_norm": 17.026273727416992, | |
| "learning_rate": 1e-05, | |
| "loss": 1.0384, | |
| "mean_token_accuracy": 0.9312147498130798, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 3.216, | |
| "grad_norm": 15.657279014587402, | |
| "learning_rate": 9.898989898989899e-06, | |
| "loss": 0.9847, | |
| "mean_token_accuracy": 0.9377696067094803, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 3.224, | |
| "grad_norm": 15.536697387695312, | |
| "learning_rate": 9.7979797979798e-06, | |
| "loss": 0.8828, | |
| "mean_token_accuracy": 0.9396137297153473, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 3.232, | |
| "grad_norm": 15.973769187927246, | |
| "learning_rate": 9.696969696969698e-06, | |
| "loss": 1.0313, | |
| "mean_token_accuracy": 0.9414243400096893, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 3.24, | |
| "grad_norm": 17.689178466796875, | |
| "learning_rate": 9.595959595959595e-06, | |
| "loss": 1.1151, | |
| "mean_token_accuracy": 0.9186365902423859, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 3.248, | |
| "grad_norm": 16.18196678161621, | |
| "learning_rate": 9.494949494949495e-06, | |
| "loss": 1.0122, | |
| "mean_token_accuracy": 0.9342964887619019, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 3.2560000000000002, | |
| "grad_norm": 14.099705696105957, | |
| "learning_rate": 9.393939393939394e-06, | |
| "loss": 0.9217, | |
| "mean_token_accuracy": 0.9330043792724609, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 3.2640000000000002, | |
| "grad_norm": 16.102678298950195, | |
| "learning_rate": 9.292929292929294e-06, | |
| "loss": 1.0442, | |
| "mean_token_accuracy": 0.9256764650344849, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 3.2720000000000002, | |
| "grad_norm": 16.18111228942871, | |
| "learning_rate": 9.191919191919192e-06, | |
| "loss": 0.9579, | |
| "mean_token_accuracy": 0.92878857254982, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 3.2800000000000002, | |
| "grad_norm": 16.256258010864258, | |
| "learning_rate": 9.090909090909091e-06, | |
| "loss": 1.0232, | |
| "mean_token_accuracy": 0.9233647882938385, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 3.288, | |
| "grad_norm": 15.954553604125977, | |
| "learning_rate": 8.98989898989899e-06, | |
| "loss": 0.9598, | |
| "mean_token_accuracy": 0.9369918555021286, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 3.296, | |
| "grad_norm": 16.494140625, | |
| "learning_rate": 8.88888888888889e-06, | |
| "loss": 1.0289, | |
| "mean_token_accuracy": 0.9376811236143112, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 3.304, | |
| "grad_norm": 14.779414176940918, | |
| "learning_rate": 8.787878787878788e-06, | |
| "loss": 0.9465, | |
| "mean_token_accuracy": 0.9370988309383392, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 3.312, | |
| "grad_norm": 15.431049346923828, | |
| "learning_rate": 8.686868686868687e-06, | |
| "loss": 0.9653, | |
| "mean_token_accuracy": 0.9339174628257751, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 3.32, | |
| "grad_norm": 15.794412612915039, | |
| "learning_rate": 8.585858585858587e-06, | |
| "loss": 1.0466, | |
| "mean_token_accuracy": 0.9339481443166733, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 3.328, | |
| "grad_norm": 17.15564727783203, | |
| "learning_rate": 8.484848484848486e-06, | |
| "loss": 1.1203, | |
| "mean_token_accuracy": 0.9287491142749786, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 3.336, | |
| "grad_norm": 16.112878799438477, | |
| "learning_rate": 8.383838383838384e-06, | |
| "loss": 1.0063, | |
| "mean_token_accuracy": 0.9275001287460327, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 3.344, | |
| "grad_norm": 15.973852157592773, | |
| "learning_rate": 8.282828282828283e-06, | |
| "loss": 1.0703, | |
| "mean_token_accuracy": 0.92690809071064, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 3.352, | |
| "grad_norm": 14.684673309326172, | |
| "learning_rate": 8.181818181818183e-06, | |
| "loss": 0.9457, | |
| "mean_token_accuracy": 0.9447111636400223, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 3.36, | |
| "grad_norm": 15.560981750488281, | |
| "learning_rate": 8.080808080808082e-06, | |
| "loss": 1.0273, | |
| "mean_token_accuracy": 0.934164434671402, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 3.368, | |
| "grad_norm": 15.820408821105957, | |
| "learning_rate": 7.97979797979798e-06, | |
| "loss": 0.976, | |
| "mean_token_accuracy": 0.9367633759975433, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 3.376, | |
| "grad_norm": 14.710906982421875, | |
| "learning_rate": 7.878787878787878e-06, | |
| "loss": 0.9833, | |
| "mean_token_accuracy": 0.9340205192565918, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 3.384, | |
| "grad_norm": 14.58287239074707, | |
| "learning_rate": 7.777777777777777e-06, | |
| "loss": 0.9453, | |
| "mean_token_accuracy": 0.9295031130313873, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 3.392, | |
| "grad_norm": 15.10856819152832, | |
| "learning_rate": 7.676767676767677e-06, | |
| "loss": 0.9536, | |
| "mean_token_accuracy": 0.9329349398612976, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 3.4, | |
| "grad_norm": 12.695680618286133, | |
| "learning_rate": 7.5757575757575764e-06, | |
| "loss": 0.8262, | |
| "mean_token_accuracy": 0.9397930353879929, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 3.408, | |
| "grad_norm": 16.12411880493164, | |
| "learning_rate": 7.474747474747475e-06, | |
| "loss": 1.1105, | |
| "mean_token_accuracy": 0.9315387904644012, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 3.416, | |
| "grad_norm": 15.864951133728027, | |
| "learning_rate": 7.3737373737373745e-06, | |
| "loss": 0.9715, | |
| "mean_token_accuracy": 0.9408050626516342, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 3.424, | |
| "grad_norm": 14.210537910461426, | |
| "learning_rate": 7.272727272727272e-06, | |
| "loss": 1.0179, | |
| "mean_token_accuracy": 0.933457687497139, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 3.432, | |
| "grad_norm": 13.187994003295898, | |
| "learning_rate": 7.171717171717173e-06, | |
| "loss": 0.912, | |
| "mean_token_accuracy": 0.9418987780809402, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 3.44, | |
| "grad_norm": 17.883209228515625, | |
| "learning_rate": 7.0707070707070704e-06, | |
| "loss": 1.1905, | |
| "mean_token_accuracy": 0.9247805923223495, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 3.448, | |
| "grad_norm": 14.041234016418457, | |
| "learning_rate": 6.969696969696971e-06, | |
| "loss": 0.9638, | |
| "mean_token_accuracy": 0.9382515847682953, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 3.456, | |
| "grad_norm": 14.920849800109863, | |
| "learning_rate": 6.8686868686868685e-06, | |
| "loss": 1.0225, | |
| "mean_token_accuracy": 0.9311228841543198, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 3.464, | |
| "grad_norm": 14.72715950012207, | |
| "learning_rate": 6.767676767676769e-06, | |
| "loss": 1.0436, | |
| "mean_token_accuracy": 0.9386313855648041, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 3.472, | |
| "grad_norm": 14.456624984741211, | |
| "learning_rate": 6.666666666666667e-06, | |
| "loss": 0.976, | |
| "mean_token_accuracy": 0.9332603514194489, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 3.48, | |
| "grad_norm": 14.296424865722656, | |
| "learning_rate": 6.565656565656567e-06, | |
| "loss": 0.9394, | |
| "mean_token_accuracy": 0.9352370202541351, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 3.488, | |
| "grad_norm": 14.372357368469238, | |
| "learning_rate": 6.464646464646465e-06, | |
| "loss": 1.0074, | |
| "mean_token_accuracy": 0.9338279366493225, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 3.496, | |
| "grad_norm": 14.993144989013672, | |
| "learning_rate": 6.363636363636363e-06, | |
| "loss": 1.028, | |
| "mean_token_accuracy": 0.932237908244133, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 3.504, | |
| "grad_norm": 15.180071830749512, | |
| "learning_rate": 6.262626262626263e-06, | |
| "loss": 0.949, | |
| "mean_token_accuracy": 0.937222346663475, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 3.512, | |
| "grad_norm": 16.305255889892578, | |
| "learning_rate": 6.161616161616162e-06, | |
| "loss": 1.0678, | |
| "mean_token_accuracy": 0.9322217702865601, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 3.52, | |
| "grad_norm": 14.723209381103516, | |
| "learning_rate": 6.060606060606061e-06, | |
| "loss": 0.9263, | |
| "mean_token_accuracy": 0.943725198507309, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 3.528, | |
| "grad_norm": 13.980120658874512, | |
| "learning_rate": 5.9595959595959605e-06, | |
| "loss": 0.8992, | |
| "mean_token_accuracy": 0.93913933634758, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 3.536, | |
| "grad_norm": 15.935212135314941, | |
| "learning_rate": 5.858585858585859e-06, | |
| "loss": 1.0253, | |
| "mean_token_accuracy": 0.9386613219976425, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 3.544, | |
| "grad_norm": 13.340989112854004, | |
| "learning_rate": 5.7575757575757586e-06, | |
| "loss": 0.903, | |
| "mean_token_accuracy": 0.9501822739839554, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 3.552, | |
| "grad_norm": 15.33117389678955, | |
| "learning_rate": 5.656565656565657e-06, | |
| "loss": 0.9156, | |
| "mean_token_accuracy": 0.9392779171466827, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 3.56, | |
| "grad_norm": 15.94556999206543, | |
| "learning_rate": 5.555555555555556e-06, | |
| "loss": 1.0368, | |
| "mean_token_accuracy": 0.9274450391530991, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 3.568, | |
| "grad_norm": 16.81086540222168, | |
| "learning_rate": 5.4545454545454545e-06, | |
| "loss": 0.9722, | |
| "mean_token_accuracy": 0.929038405418396, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 3.576, | |
| "grad_norm": 14.2332124710083, | |
| "learning_rate": 5.353535353535354e-06, | |
| "loss": 0.8955, | |
| "mean_token_accuracy": 0.9410098195075989, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 3.584, | |
| "grad_norm": 16.48716926574707, | |
| "learning_rate": 5.2525252525252526e-06, | |
| "loss": 1.0025, | |
| "mean_token_accuracy": 0.9347028732299805, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 3.592, | |
| "grad_norm": 14.993062019348145, | |
| "learning_rate": 5.151515151515152e-06, | |
| "loss": 0.943, | |
| "mean_token_accuracy": 0.940876841545105, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 3.6, | |
| "grad_norm": 15.412555694580078, | |
| "learning_rate": 5.050505050505051e-06, | |
| "loss": 1.019, | |
| "mean_token_accuracy": 0.9378762990236282, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 3.608, | |
| "grad_norm": 14.794282913208008, | |
| "learning_rate": 4.949494949494949e-06, | |
| "loss": 0.9683, | |
| "mean_token_accuracy": 0.9323435574769974, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 3.616, | |
| "grad_norm": 14.959280967712402, | |
| "learning_rate": 4.848484848484849e-06, | |
| "loss": 1.0428, | |
| "mean_token_accuracy": 0.9346833378076553, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 3.624, | |
| "grad_norm": 17.50050926208496, | |
| "learning_rate": 4.747474747474747e-06, | |
| "loss": 1.0809, | |
| "mean_token_accuracy": 0.9283154457807541, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 3.632, | |
| "grad_norm": 14.101909637451172, | |
| "learning_rate": 4.646464646464647e-06, | |
| "loss": 0.9455, | |
| "mean_token_accuracy": 0.9361487329006195, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 3.64, | |
| "grad_norm": 14.593957901000977, | |
| "learning_rate": 4.5454545454545455e-06, | |
| "loss": 0.9624, | |
| "mean_token_accuracy": 0.9361502379179001, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 3.648, | |
| "grad_norm": 15.813642501831055, | |
| "learning_rate": 4.444444444444445e-06, | |
| "loss": 0.9207, | |
| "mean_token_accuracy": 0.939594179391861, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 3.656, | |
| "grad_norm": 13.442476272583008, | |
| "learning_rate": 4.343434343434344e-06, | |
| "loss": 0.8886, | |
| "mean_token_accuracy": 0.938508003950119, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 3.664, | |
| "grad_norm": 18.873268127441406, | |
| "learning_rate": 4.242424242424243e-06, | |
| "loss": 1.1159, | |
| "mean_token_accuracy": 0.9257403612136841, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 3.672, | |
| "grad_norm": 14.351343154907227, | |
| "learning_rate": 4.141414141414142e-06, | |
| "loss": 0.9803, | |
| "mean_token_accuracy": 0.9281805902719498, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 3.68, | |
| "grad_norm": 17.297321319580078, | |
| "learning_rate": 4.040404040404041e-06, | |
| "loss": 1.0052, | |
| "mean_token_accuracy": 0.9295466840267181, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 3.6879999999999997, | |
| "grad_norm": 21.608966827392578, | |
| "learning_rate": 3.939393939393939e-06, | |
| "loss": 1.013, | |
| "mean_token_accuracy": 0.9357683062553406, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 3.6959999999999997, | |
| "grad_norm": 16.225805282592773, | |
| "learning_rate": 3.8383838383838385e-06, | |
| "loss": 1.0466, | |
| "mean_token_accuracy": 0.9302550107240677, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 3.7039999999999997, | |
| "grad_norm": 14.666239738464355, | |
| "learning_rate": 3.7373737373737375e-06, | |
| "loss": 0.9911, | |
| "mean_token_accuracy": 0.9366718828678131, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 3.7119999999999997, | |
| "grad_norm": 15.249946594238281, | |
| "learning_rate": 3.636363636363636e-06, | |
| "loss": 1.0193, | |
| "mean_token_accuracy": 0.9248189479112625, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 3.7199999999999998, | |
| "grad_norm": 15.235973358154297, | |
| "learning_rate": 3.5353535353535352e-06, | |
| "loss": 1.0371, | |
| "mean_token_accuracy": 0.9370195269584656, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 3.7279999999999998, | |
| "grad_norm": 15.745083808898926, | |
| "learning_rate": 3.4343434343434343e-06, | |
| "loss": 0.951, | |
| "mean_token_accuracy": 0.9338164031505585, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 3.7359999999999998, | |
| "grad_norm": 14.804433822631836, | |
| "learning_rate": 3.3333333333333333e-06, | |
| "loss": 1.0366, | |
| "mean_token_accuracy": 0.9359022080898285, | |
| "step": 467 | |
| }, | |
| { | |
| "epoch": 3.7439999999999998, | |
| "grad_norm": 13.7555570602417, | |
| "learning_rate": 3.2323232323232324e-06, | |
| "loss": 0.977, | |
| "mean_token_accuracy": 0.9416959583759308, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 3.752, | |
| "grad_norm": 14.858553886413574, | |
| "learning_rate": 3.1313131313131314e-06, | |
| "loss": 0.9739, | |
| "mean_token_accuracy": 0.9415158182382584, | |
| "step": 469 | |
| }, | |
| { | |
| "epoch": 3.76, | |
| "grad_norm": 16.706321716308594, | |
| "learning_rate": 3.0303030303030305e-06, | |
| "loss": 1.0242, | |
| "mean_token_accuracy": 0.9358144998550415, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 3.768, | |
| "grad_norm": 15.634292602539062, | |
| "learning_rate": 2.9292929292929295e-06, | |
| "loss": 1.0251, | |
| "mean_token_accuracy": 0.9278706014156342, | |
| "step": 471 | |
| }, | |
| { | |
| "epoch": 3.776, | |
| "grad_norm": 13.777891159057617, | |
| "learning_rate": 2.8282828282828286e-06, | |
| "loss": 0.8874, | |
| "mean_token_accuracy": 0.9410149455070496, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 3.784, | |
| "grad_norm": 16.45195198059082, | |
| "learning_rate": 2.7272727272727272e-06, | |
| "loss": 0.9863, | |
| "mean_token_accuracy": 0.9311119616031647, | |
| "step": 473 | |
| }, | |
| { | |
| "epoch": 3.792, | |
| "grad_norm": 16.67084312438965, | |
| "learning_rate": 2.6262626262626263e-06, | |
| "loss": 0.9869, | |
| "mean_token_accuracy": 0.9290085285902023, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 3.8, | |
| "grad_norm": 16.300039291381836, | |
| "learning_rate": 2.5252525252525253e-06, | |
| "loss": 1.0872, | |
| "mean_token_accuracy": 0.9302412569522858, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 3.808, | |
| "grad_norm": 13.501500129699707, | |
| "learning_rate": 2.4242424242424244e-06, | |
| "loss": 0.917, | |
| "mean_token_accuracy": 0.9408380538225174, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 3.816, | |
| "grad_norm": 16.53856658935547, | |
| "learning_rate": 2.3232323232323234e-06, | |
| "loss": 1.1446, | |
| "mean_token_accuracy": 0.9272817373275757, | |
| "step": 477 | |
| }, | |
| { | |
| "epoch": 3.824, | |
| "grad_norm": 14.656960487365723, | |
| "learning_rate": 2.2222222222222225e-06, | |
| "loss": 0.8935, | |
| "mean_token_accuracy": 0.9378921240568161, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 3.832, | |
| "grad_norm": 14.75877857208252, | |
| "learning_rate": 2.1212121212121216e-06, | |
| "loss": 0.9782, | |
| "mean_token_accuracy": 0.9403275102376938, | |
| "step": 479 | |
| }, | |
| { | |
| "epoch": 3.84, | |
| "grad_norm": 15.582100868225098, | |
| "learning_rate": 2.0202020202020206e-06, | |
| "loss": 1.0506, | |
| "mean_token_accuracy": 0.9281135201454163, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 3.848, | |
| "grad_norm": 16.705493927001953, | |
| "learning_rate": 1.9191919191919192e-06, | |
| "loss": 1.0155, | |
| "mean_token_accuracy": 0.9309724718332291, | |
| "step": 481 | |
| }, | |
| { | |
| "epoch": 3.856, | |
| "grad_norm": 14.00627326965332, | |
| "learning_rate": 1.818181818181818e-06, | |
| "loss": 0.9561, | |
| "mean_token_accuracy": 0.9410349428653717, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 3.864, | |
| "grad_norm": 16.71605682373047, | |
| "learning_rate": 1.7171717171717171e-06, | |
| "loss": 1.0531, | |
| "mean_token_accuracy": 0.9354703277349472, | |
| "step": 483 | |
| }, | |
| { | |
| "epoch": 3.872, | |
| "grad_norm": 12.789509773254395, | |
| "learning_rate": 1.6161616161616162e-06, | |
| "loss": 0.8623, | |
| "mean_token_accuracy": 0.9429314136505127, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 3.88, | |
| "grad_norm": 13.605375289916992, | |
| "learning_rate": 1.5151515151515152e-06, | |
| "loss": 1.017, | |
| "mean_token_accuracy": 0.9328679293394089, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 3.888, | |
| "grad_norm": 15.087848663330078, | |
| "learning_rate": 1.4141414141414143e-06, | |
| "loss": 1.0554, | |
| "mean_token_accuracy": 0.9394855350255966, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 3.896, | |
| "grad_norm": 15.166902542114258, | |
| "learning_rate": 1.3131313131313131e-06, | |
| "loss": 0.9055, | |
| "mean_token_accuracy": 0.9326141625642776, | |
| "step": 487 | |
| }, | |
| { | |
| "epoch": 3.904, | |
| "grad_norm": 15.783126831054688, | |
| "learning_rate": 1.2121212121212122e-06, | |
| "loss": 0.9401, | |
| "mean_token_accuracy": 0.9369719177484512, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 3.912, | |
| "grad_norm": 15.481629371643066, | |
| "learning_rate": 1.1111111111111112e-06, | |
| "loss": 0.9853, | |
| "mean_token_accuracy": 0.9318475723266602, | |
| "step": 489 | |
| }, | |
| { | |
| "epoch": 3.92, | |
| "grad_norm": 16.257150650024414, | |
| "learning_rate": 1.0101010101010103e-06, | |
| "loss": 1.0291, | |
| "mean_token_accuracy": 0.9357573091983795, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 3.928, | |
| "grad_norm": 14.175272941589355, | |
| "learning_rate": 9.09090909090909e-07, | |
| "loss": 0.8596, | |
| "mean_token_accuracy": 0.9388793110847473, | |
| "step": 491 | |
| }, | |
| { | |
| "epoch": 3.936, | |
| "grad_norm": 14.242705345153809, | |
| "learning_rate": 8.080808080808081e-07, | |
| "loss": 0.9267, | |
| "mean_token_accuracy": 0.9343722760677338, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 3.944, | |
| "grad_norm": 17.908700942993164, | |
| "learning_rate": 7.070707070707071e-07, | |
| "loss": 1.0493, | |
| "mean_token_accuracy": 0.923067107796669, | |
| "step": 493 | |
| }, | |
| { | |
| "epoch": 3.952, | |
| "grad_norm": 14.200380325317383, | |
| "learning_rate": 6.060606060606061e-07, | |
| "loss": 0.9237, | |
| "mean_token_accuracy": 0.9427360892295837, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 3.96, | |
| "grad_norm": 14.010753631591797, | |
| "learning_rate": 5.050505050505052e-07, | |
| "loss": 0.8624, | |
| "mean_token_accuracy": 0.9410801976919174, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 3.968, | |
| "grad_norm": 17.56831169128418, | |
| "learning_rate": 4.0404040404040405e-07, | |
| "loss": 1.002, | |
| "mean_token_accuracy": 0.9286356568336487, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 3.976, | |
| "grad_norm": 12.637655258178711, | |
| "learning_rate": 3.0303030303030305e-07, | |
| "loss": 0.8658, | |
| "mean_token_accuracy": 0.9444387257099152, | |
| "step": 497 | |
| }, | |
| { | |
| "epoch": 3.984, | |
| "grad_norm": 17.883581161499023, | |
| "learning_rate": 2.0202020202020202e-07, | |
| "loss": 1.068, | |
| "mean_token_accuracy": 0.9314180314540863, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 3.992, | |
| "grad_norm": 15.118261337280273, | |
| "learning_rate": 1.0101010101010101e-07, | |
| "loss": 1.0254, | |
| "mean_token_accuracy": 0.9345874488353729, | |
| "step": 499 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "grad_norm": 14.052021026611328, | |
| "learning_rate": 0.0, | |
| "loss": 1.0111, | |
| "mean_token_accuracy": 0.9361301213502884, | |
| "step": 500 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 500, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 4, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 8795704000512000.0, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |