fixed-spin-reg / trainer_state.json
semran1's picture
Upload folder using huggingface_hub
944fd70 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.0125,
"eval_steps": 500,
"global_step": 500,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 2.5e-05,
"grad_norm": 8.947145462036133,
"learning_rate": 2.0000000000000002e-07,
"loss": 127.4759,
"loss/crossentropy": 2.8921360969543457,
"loss/hidden": 0.373046875,
"loss/logits": 0.1864406317472458,
"loss/reg": 126.91645812988281,
"step": 1
},
{
"epoch": 5e-05,
"grad_norm": 13.883983612060547,
"learning_rate": 4.0000000000000003e-07,
"loss": 115.8264,
"loss/crossentropy": 2.8499865531921387,
"loss/hidden": 0.6484375,
"loss/logits": 0.27192822098731995,
"loss/reg": 114.90603637695312,
"step": 2
},
{
"epoch": 7.5e-05,
"grad_norm": 19.773080825805664,
"learning_rate": 6.000000000000001e-07,
"loss": 116.012,
"loss/crossentropy": 3.1656363010406494,
"loss/hidden": 0.72265625,
"loss/logits": 0.3859825134277344,
"loss/reg": 114.9033432006836,
"step": 3
},
{
"epoch": 0.0001,
"grad_norm": 18.47258758544922,
"learning_rate": 8.000000000000001e-07,
"loss": 115.903,
"loss/crossentropy": 2.8063976764678955,
"loss/hidden": 0.6875,
"loss/logits": 0.31685781478881836,
"loss/reg": 114.89862823486328,
"step": 4
},
{
"epoch": 0.000125,
"grad_norm": 18.90447998046875,
"learning_rate": 1.0000000000000002e-06,
"loss": 115.9715,
"loss/crossentropy": 3.020209312438965,
"loss/hidden": 0.7265625,
"loss/logits": 0.3528100252151489,
"loss/reg": 114.89217376708984,
"step": 5
},
{
"epoch": 0.00015,
"grad_norm": 15.197858810424805,
"learning_rate": 1.2000000000000002e-06,
"loss": 115.9385,
"loss/crossentropy": 2.891916036605835,
"loss/hidden": 0.6953125,
"loss/logits": 0.35902339220046997,
"loss/reg": 114.8841781616211,
"step": 6
},
{
"epoch": 0.000175,
"grad_norm": 15.15339469909668,
"learning_rate": 1.4000000000000001e-06,
"loss": 115.8638,
"loss/crossentropy": 2.888406753540039,
"loss/hidden": 0.6796875,
"loss/logits": 0.3095901906490326,
"loss/reg": 114.8744888305664,
"step": 7
},
{
"epoch": 0.0002,
"grad_norm": 13.61527156829834,
"learning_rate": 1.6000000000000001e-06,
"loss": 115.85,
"loss/crossentropy": 2.9885716438293457,
"loss/hidden": 0.671875,
"loss/logits": 0.3148782253265381,
"loss/reg": 114.86323547363281,
"step": 8
},
{
"epoch": 0.000225,
"grad_norm": 13.632570266723633,
"learning_rate": 1.8000000000000001e-06,
"loss": 115.7999,
"loss/crossentropy": 2.6390163898468018,
"loss/hidden": 0.6640625,
"loss/logits": 0.2855065166950226,
"loss/reg": 114.85034942626953,
"step": 9
},
{
"epoch": 0.00025,
"grad_norm": 12.151652336120605,
"learning_rate": 2.0000000000000003e-06,
"loss": 115.7602,
"loss/crossentropy": 2.9643092155456543,
"loss/hidden": 0.6328125,
"loss/logits": 0.2915680408477783,
"loss/reg": 114.83580780029297,
"step": 10
},
{
"epoch": 0.000275,
"grad_norm": 10.124210357666016,
"learning_rate": 2.2e-06,
"loss": 115.6818,
"loss/crossentropy": 2.7221367359161377,
"loss/hidden": 0.609375,
"loss/logits": 0.2528327405452728,
"loss/reg": 114.81963348388672,
"step": 11
},
{
"epoch": 0.0003,
"grad_norm": 12.582488059997559,
"learning_rate": 2.4000000000000003e-06,
"loss": 112.7148,
"loss/crossentropy": 3.236349582672119,
"loss/hidden": 0.7890625,
"loss/logits": 0.3880160450935364,
"loss/reg": 111.53773498535156,
"step": 12
},
{
"epoch": 0.000325,
"grad_norm": 9.126688957214355,
"learning_rate": 2.6e-06,
"loss": 112.5513,
"loss/crossentropy": 2.9401042461395264,
"loss/hidden": 0.7109375,
"loss/logits": 0.3218693733215332,
"loss/reg": 111.51847839355469,
"step": 13
},
{
"epoch": 0.00035,
"grad_norm": 5.668403625488281,
"learning_rate": 2.8000000000000003e-06,
"loss": 112.3898,
"loss/crossentropy": 2.5529444217681885,
"loss/hidden": 0.66796875,
"loss/logits": 0.22446033358573914,
"loss/reg": 111.49739837646484,
"step": 14
},
{
"epoch": 0.000375,
"grad_norm": 5.202491283416748,
"learning_rate": 3e-06,
"loss": 112.4113,
"loss/crossentropy": 3.1544463634490967,
"loss/hidden": 0.65625,
"loss/logits": 0.28114181756973267,
"loss/reg": 111.47393035888672,
"step": 15
},
{
"epoch": 0.0004,
"grad_norm": 4.127603054046631,
"grad_norm_var": 23.16038376789459,
"learning_rate": 3.2000000000000003e-06,
"loss": 112.257,
"loss/crossentropy": 3.014888286590576,
"loss/hidden": 0.6015625,
"loss/logits": 0.20710057020187378,
"loss/reg": 111.44828796386719,
"step": 16
},
{
"epoch": 0.000425,
"grad_norm": 3.9600610733032227,
"grad_norm_var": 26.934472753712136,
"learning_rate": 3.4000000000000005e-06,
"loss": 112.1822,
"loss/crossentropy": 2.7658724784851074,
"loss/hidden": 0.57421875,
"loss/logits": 0.18775172531604767,
"loss/reg": 111.42024230957031,
"step": 17
},
{
"epoch": 0.00045,
"grad_norm": 4.745978355407715,
"grad_norm_var": 29.825747343372463,
"learning_rate": 3.6000000000000003e-06,
"loss": 112.1734,
"loss/crossentropy": 3.176030158996582,
"loss/hidden": 0.58203125,
"loss/logits": 0.2011646032333374,
"loss/reg": 111.39022064208984,
"step": 18
},
{
"epoch": 0.000475,
"grad_norm": 3.7388970851898193,
"grad_norm_var": 27.998639503348578,
"learning_rate": 3.8000000000000005e-06,
"loss": 112.0506,
"loss/crossentropy": 3.0487682819366455,
"loss/hidden": 0.53515625,
"loss/logits": 0.15711598098278046,
"loss/reg": 111.3582763671875,
"step": 19
},
{
"epoch": 0.0005,
"grad_norm": 3.213416576385498,
"grad_norm_var": 26.127740304788446,
"learning_rate": 4.000000000000001e-06,
"loss": 111.9841,
"loss/crossentropy": 2.9822959899902344,
"loss/hidden": 0.51171875,
"loss/logits": 0.14767958223819733,
"loss/reg": 111.32469177246094,
"step": 20
},
{
"epoch": 0.000525,
"grad_norm": 5.186110019683838,
"grad_norm_var": 20.59027009605237,
"learning_rate": 4.2000000000000004e-06,
"loss": 111.9338,
"loss/crossentropy": 2.6938610076904297,
"loss/hidden": 0.50390625,
"loss/logits": 0.14082029461860657,
"loss/reg": 111.2890625,
"step": 21
},
{
"epoch": 0.00055,
"grad_norm": 3.830970525741577,
"grad_norm_var": 18.64965973207226,
"learning_rate": 4.4e-06,
"loss": 111.3757,
"loss/crossentropy": 2.60425066947937,
"loss/hidden": 0.5234375,
"loss/logits": 0.11805371195077896,
"loss/reg": 110.73417663574219,
"step": 22
},
{
"epoch": 0.000575,
"grad_norm": 4.170963764190674,
"grad_norm_var": 15.535602850312825,
"learning_rate": 4.600000000000001e-06,
"loss": 111.3937,
"loss/crossentropy": 3.180182695388794,
"loss/hidden": 0.55859375,
"loss/logits": 0.14081788063049316,
"loss/reg": 110.69429779052734,
"step": 23
},
{
"epoch": 0.0006,
"grad_norm": 2.6631879806518555,
"grad_norm_var": 13.65312757585727,
"learning_rate": 4.800000000000001e-06,
"loss": 111.2621,
"loss/crossentropy": 2.8149499893188477,
"loss/hidden": 0.48828125,
"loss/logits": 0.12221543490886688,
"loss/reg": 110.65160369873047,
"step": 24
},
{
"epoch": 0.000625,
"grad_norm": 2.618382453918457,
"grad_norm_var": 10.772089347301707,
"learning_rate": 5e-06,
"loss": 111.1819,
"loss/crossentropy": 2.8573601245880127,
"loss/hidden": 0.458984375,
"loss/logits": 0.11608506739139557,
"loss/reg": 110.60684204101562,
"step": 25
},
{
"epoch": 0.00065,
"grad_norm": 2.31030011177063,
"grad_norm_var": 8.516386613252116,
"learning_rate": 5.2e-06,
"loss": 111.117,
"loss/crossentropy": 3.055554151535034,
"loss/hidden": 0.439453125,
"loss/logits": 0.11802805215120316,
"loss/reg": 110.5595474243164,
"step": 26
},
{
"epoch": 0.000675,
"grad_norm": 2.017902374267578,
"grad_norm_var": 7.3058512031776175,
"learning_rate": 5.400000000000001e-06,
"loss": 110.9925,
"loss/crossentropy": 3.038731813430786,
"loss/hidden": 0.388671875,
"loss/logits": 0.09406228363513947,
"loss/reg": 110.50971984863281,
"step": 27
},
{
"epoch": 0.0007,
"grad_norm": 2.0806379318237305,
"grad_norm_var": 3.1583128509652356,
"learning_rate": 5.600000000000001e-06,
"loss": 110.9178,
"loss/crossentropy": 2.8045244216918945,
"loss/hidden": 0.37109375,
"loss/logits": 0.0893578976392746,
"loss/reg": 110.45733642578125,
"step": 28
},
{
"epoch": 0.000725,
"grad_norm": 1.8444489240646362,
"grad_norm_var": 1.5350877177341649,
"learning_rate": 5.8e-06,
"loss": 110.8323,
"loss/crossentropy": 2.9446072578430176,
"loss/hidden": 0.3515625,
"loss/logits": 0.07847409695386887,
"loss/reg": 110.40225219726562,
"step": 29
},
{
"epoch": 0.00075,
"grad_norm": 1.900687575340271,
"grad_norm_var": 1.3763151431979745,
"learning_rate": 6e-06,
"loss": 110.766,
"loss/crossentropy": 2.991264581680298,
"loss/hidden": 0.345703125,
"loss/logits": 0.07573088258504868,
"loss/reg": 110.34453582763672,
"step": 30
},
{
"epoch": 0.000775,
"grad_norm": 1.8288308382034302,
"grad_norm_var": 1.254712449102712,
"learning_rate": 6.200000000000001e-06,
"loss": 110.6691,
"loss/crossentropy": 2.695741653442383,
"loss/hidden": 0.31640625,
"loss/logits": 0.06882831454277039,
"loss/reg": 110.28388214111328,
"step": 31
},
{
"epoch": 0.0008,
"grad_norm": 7.6339874267578125,
"grad_norm_var": 2.4849026575088917,
"learning_rate": 6.4000000000000006e-06,
"loss": 110.524,
"loss/crossentropy": 2.8983194828033447,
"loss/hidden": 0.30078125,
"loss/logits": 0.0650840774178505,
"loss/reg": 110.15809631347656,
"step": 32
},
{
"epoch": 0.000825,
"grad_norm": 1.7091763019561768,
"grad_norm_var": 2.6211828816924214,
"learning_rate": 6.600000000000001e-06,
"loss": 110.421,
"loss/crossentropy": 3.1094229221343994,
"loss/hidden": 0.2734375,
"loss/logits": 0.055526312440633774,
"loss/reg": 110.09200286865234,
"step": 33
},
{
"epoch": 0.00085,
"grad_norm": 2.408731698989868,
"grad_norm_var": 2.486549186444365,
"learning_rate": 6.800000000000001e-06,
"loss": 110.3723,
"loss/crossentropy": 3.0703020095825195,
"loss/hidden": 0.28515625,
"loss/logits": 0.06386865675449371,
"loss/reg": 110.02332305908203,
"step": 34
},
{
"epoch": 0.000875,
"grad_norm": 5.639738082885742,
"grad_norm_var": 2.8813226444205484,
"learning_rate": 7e-06,
"loss": 110.2663,
"loss/crossentropy": 2.6366381645202637,
"loss/hidden": 0.259765625,
"loss/logits": 0.05466887354850769,
"loss/reg": 109.95182800292969,
"step": 35
},
{
"epoch": 0.0009,
"grad_norm": 1.5474259853363037,
"grad_norm_var": 3.049833938595448,
"learning_rate": 7.2000000000000005e-06,
"loss": 110.163,
"loss/crossentropy": 2.9982497692108154,
"loss/hidden": 0.2373046875,
"loss/logits": 0.0486503466963768,
"loss/reg": 109.8770751953125,
"step": 36
},
{
"epoch": 0.000925,
"grad_norm": 1.822043776512146,
"grad_norm_var": 2.8155889180573683,
"learning_rate": 7.4e-06,
"loss": 110.0863,
"loss/crossentropy": 2.5750184059143066,
"loss/hidden": 0.244140625,
"loss/logits": 0.04214131087064743,
"loss/reg": 109.80003356933594,
"step": 37
},
{
"epoch": 0.00095,
"grad_norm": 1.646996259689331,
"grad_norm_var": 2.8358215165690646,
"learning_rate": 7.600000000000001e-06,
"loss": 109.9793,
"loss/crossentropy": 2.6531567573547363,
"loss/hidden": 0.216796875,
"loss/logits": 0.04228401929140091,
"loss/reg": 109.72017669677734,
"step": 38
},
{
"epoch": 0.000975,
"grad_norm": 1.6782788038253784,
"grad_norm_var": 2.7486431517133116,
"learning_rate": 7.800000000000002e-06,
"loss": 109.9137,
"loss/crossentropy": 2.8859128952026367,
"loss/hidden": 0.2333984375,
"loss/logits": 0.04267453774809837,
"loss/reg": 109.63758087158203,
"step": 39
},
{
"epoch": 0.001,
"grad_norm": 1.1451095342636108,
"grad_norm_var": 2.8767352862187083,
"learning_rate": 8.000000000000001e-06,
"loss": 109.7821,
"loss/crossentropy": 2.717745304107666,
"loss/hidden": 0.1982421875,
"loss/logits": 0.03221501410007477,
"loss/reg": 109.55165100097656,
"step": 40
},
{
"epoch": 0.001025,
"grad_norm": 1.5123186111450195,
"grad_norm_var": 2.934195649765045,
"learning_rate": 8.2e-06,
"loss": 109.6931,
"loss/crossentropy": 3.1204404830932617,
"loss/hidden": 0.193359375,
"loss/logits": 0.03670802339911461,
"loss/reg": 109.46299743652344,
"step": 41
},
{
"epoch": 0.00105,
"grad_norm": 1.5102463960647583,
"grad_norm_var": 2.985947226480973,
"learning_rate": 8.400000000000001e-06,
"loss": 109.5902,
"loss/crossentropy": 2.754708766937256,
"loss/hidden": 0.193359375,
"loss/logits": 0.0338013619184494,
"loss/reg": 109.36308288574219,
"step": 42
},
{
"epoch": 0.001075,
"grad_norm": 1.7040822505950928,
"grad_norm_var": 3.00685228461388,
"learning_rate": 8.6e-06,
"loss": 109.4694,
"loss/crossentropy": 2.99281907081604,
"loss/hidden": 0.169921875,
"loss/logits": 0.030927911400794983,
"loss/reg": 109.26850891113281,
"step": 43
},
{
"epoch": 0.0011,
"grad_norm": 2.202953338623047,
"grad_norm_var": 3.0033814137683112,
"learning_rate": 8.8e-06,
"loss": 109.3558,
"loss/crossentropy": 2.8213930130004883,
"loss/hidden": 0.1572265625,
"loss/logits": 0.027427691966295242,
"loss/reg": 109.17113494873047,
"step": 44
},
{
"epoch": 0.001125,
"grad_norm": 2.543388843536377,
"grad_norm_var": 2.986013784076106,
"learning_rate": 9e-06,
"loss": 109.2579,
"loss/crossentropy": 2.9007112979888916,
"loss/hidden": 0.1572265625,
"loss/logits": 0.02932737208902836,
"loss/reg": 109.07130432128906,
"step": 45
},
{
"epoch": 0.00115,
"grad_norm": 2.870173692703247,
"grad_norm_var": 2.9799395606014794,
"learning_rate": 9.200000000000002e-06,
"loss": 109.153,
"loss/crossentropy": 2.8239502906799316,
"loss/hidden": 0.1572265625,
"loss/logits": 0.028089674189686775,
"loss/reg": 108.96764373779297,
"step": 46
},
{
"epoch": 0.001175,
"grad_norm": 1.966118574142456,
"grad_norm_var": 2.9695142383622777,
"learning_rate": 9.4e-06,
"loss": 109.0394,
"loss/crossentropy": 2.8659369945526123,
"loss/hidden": 0.1513671875,
"loss/logits": 0.025958221405744553,
"loss/reg": 108.862060546875,
"step": 47
},
{
"epoch": 0.0012,
"grad_norm": 1.9000473022460938,
"grad_norm_var": 1.0773867075877055,
"learning_rate": 9.600000000000001e-06,
"loss": 108.9211,
"loss/crossentropy": 2.765244722366333,
"loss/hidden": 0.1455078125,
"loss/logits": 0.02295522391796112,
"loss/reg": 108.75267028808594,
"step": 48
},
{
"epoch": 0.001225,
"grad_norm": 2.161703109741211,
"grad_norm_var": 1.0658244305761075,
"learning_rate": 9.800000000000001e-06,
"loss": 108.8061,
"loss/crossentropy": 2.999718427658081,
"loss/hidden": 0.140625,
"loss/logits": 0.025498513132333755,
"loss/reg": 108.64002227783203,
"step": 49
},
{
"epoch": 0.00125,
"grad_norm": 2.4916954040527344,
"grad_norm_var": 1.0692138980283508,
"learning_rate": 1e-05,
"loss": 108.6839,
"loss/crossentropy": 2.717797040939331,
"loss/hidden": 0.138671875,
"loss/logits": 0.021402183920145035,
"loss/reg": 108.52385711669922,
"step": 50
},
{
"epoch": 0.001275,
"grad_norm": 1.824416995048523,
"grad_norm_var": 0.20190934714180309,
"learning_rate": 1.02e-05,
"loss": 108.5772,
"loss/crossentropy": 3.0746214389801025,
"loss/hidden": 0.1484375,
"loss/logits": 0.023359911516308784,
"loss/reg": 108.4053726196289,
"step": 51
},
{
"epoch": 0.0013,
"grad_norm": 2.363981008529663,
"grad_norm_var": 0.20433165463635838,
"learning_rate": 1.04e-05,
"loss": 108.4089,
"loss/crossentropy": 2.7881290912628174,
"loss/hidden": 0.115234375,
"loss/logits": 0.0219597015529871,
"loss/reg": 108.27169036865234,
"step": 52
},
{
"epoch": 0.001325,
"grad_norm": 2.2850818634033203,
"grad_norm_var": 0.2092781831522757,
"learning_rate": 1.0600000000000002e-05,
"loss": 108.2794,
"loss/crossentropy": 2.877086639404297,
"loss/hidden": 0.11328125,
"loss/logits": 0.019944053143262863,
"loss/reg": 108.1462173461914,
"step": 53
},
{
"epoch": 0.00135,
"grad_norm": 2.2029547691345215,
"grad_norm_var": 0.203324965479775,
"learning_rate": 1.0800000000000002e-05,
"loss": 108.1448,
"loss/crossentropy": 2.5684406757354736,
"loss/hidden": 0.10888671875,
"loss/logits": 0.018222380429506302,
"loss/reg": 108.01766204833984,
"step": 54
},
{
"epoch": 0.001375,
"grad_norm": 4.355873107910156,
"grad_norm_var": 0.5284712138594291,
"learning_rate": 1.1000000000000001e-05,
"loss": 108.0238,
"loss/crossentropy": 3.1050121784210205,
"loss/hidden": 0.11474609375,
"loss/logits": 0.023576244711875916,
"loss/reg": 107.8854751586914,
"step": 55
},
{
"epoch": 0.0014,
"grad_norm": 8.262518882751465,
"grad_norm_var": 2.702968942542203,
"learning_rate": 1.1200000000000001e-05,
"loss": 107.8767,
"loss/crossentropy": 2.6639468669891357,
"loss/hidden": 0.109375,
"loss/logits": 0.017037922516465187,
"loss/reg": 107.75032806396484,
"step": 56
},
{
"epoch": 0.001425,
"grad_norm": 1.5480672121047974,
"grad_norm_var": 2.6976983052976697,
"learning_rate": 1.14e-05,
"loss": 107.734,
"loss/crossentropy": 2.7018520832061768,
"loss/hidden": 0.10693359375,
"loss/logits": 0.016180139034986496,
"loss/reg": 107.61087036132812,
"step": 57
},
{
"epoch": 0.00145,
"grad_norm": 1.2781832218170166,
"grad_norm_var": 2.7359303926605016,
"learning_rate": 1.16e-05,
"loss": 107.5901,
"loss/crossentropy": 2.8058838844299316,
"loss/hidden": 0.1044921875,
"loss/logits": 0.016446152701973915,
"loss/reg": 107.46919250488281,
"step": 58
},
{
"epoch": 0.001475,
"grad_norm": 1.5537561178207397,
"grad_norm_var": 2.755752608942243,
"learning_rate": 1.18e-05,
"loss": 107.4512,
"loss/crossentropy": 2.6371188163757324,
"loss/hidden": 0.109375,
"loss/logits": 0.018213655799627304,
"loss/reg": 107.32366180419922,
"step": 59
},
{
"epoch": 0.0015,
"grad_norm": 3.3459622859954834,
"grad_norm_var": 2.774887598663352,
"learning_rate": 1.2e-05,
"loss": 107.3005,
"loss/crossentropy": 3.190295457839966,
"loss/hidden": 0.109375,
"loss/logits": 0.016267115250229836,
"loss/reg": 107.1749038696289,
"step": 60
},
{
"epoch": 0.001525,
"grad_norm": 1.5608716011047363,
"grad_norm_var": 2.8537229826550923,
"learning_rate": 1.22e-05,
"loss": 107.1405,
"loss/crossentropy": 2.689225673675537,
"loss/hidden": 0.10205078125,
"loss/logits": 0.015642033889889717,
"loss/reg": 107.0228271484375,
"step": 61
},
{
"epoch": 0.00155,
"grad_norm": 2.9783740043640137,
"grad_norm_var": 2.858017521797208,
"learning_rate": 1.2400000000000002e-05,
"loss": 106.9817,
"loss/crossentropy": 3.0421159267425537,
"loss/hidden": 0.10205078125,
"loss/logits": 0.016691621392965317,
"loss/reg": 106.86290740966797,
"step": 62
},
{
"epoch": 0.001575,
"grad_norm": 1.751490831375122,
"grad_norm_var": 2.8798941991380254,
"learning_rate": 1.2600000000000001e-05,
"loss": 106.8293,
"loss/crossentropy": 2.954704523086548,
"loss/hidden": 0.10693359375,
"loss/logits": 0.018012236803770065,
"loss/reg": 106.7043228149414,
"step": 63
},
{
"epoch": 0.0016,
"grad_norm": 2.2218408584594727,
"grad_norm_var": 2.855623539050459,
"learning_rate": 1.2800000000000001e-05,
"loss": 106.6657,
"loss/crossentropy": 2.735278844833374,
"loss/hidden": 0.1044921875,
"loss/logits": 0.019818950444459915,
"loss/reg": 106.54141998291016,
"step": 64
},
{
"epoch": 0.001625,
"grad_norm": 1.6660186052322388,
"grad_norm_var": 2.9023713600347114,
"learning_rate": 1.3000000000000001e-05,
"loss": 106.4947,
"loss/crossentropy": 3.00019907951355,
"loss/hidden": 0.1044921875,
"loss/logits": 0.015060758218169212,
"loss/reg": 106.3750991821289,
"step": 65
},
{
"epoch": 0.00165,
"grad_norm": 1.7433619499206543,
"grad_norm_var": 2.948745965901263,
"learning_rate": 1.3200000000000002e-05,
"loss": 106.3047,
"loss/crossentropy": 2.4090218544006348,
"loss/hidden": 0.0859375,
"loss/logits": 0.013767421245574951,
"loss/reg": 106.20496368408203,
"step": 66
},
{
"epoch": 0.001675,
"grad_norm": 1.8968372344970703,
"grad_norm_var": 2.941981354522321,
"learning_rate": 1.3400000000000002e-05,
"loss": 106.158,
"loss/crossentropy": 2.8800578117370605,
"loss/hidden": 0.10693359375,
"loss/logits": 0.019842734560370445,
"loss/reg": 106.03123474121094,
"step": 67
},
{
"epoch": 0.0017,
"grad_norm": 2.075490951538086,
"grad_norm_var": 2.95485559076868,
"learning_rate": 1.3600000000000002e-05,
"loss": 105.9487,
"loss/crossentropy": 3.121166944503784,
"loss/hidden": 0.07861328125,
"loss/logits": 0.015778005123138428,
"loss/reg": 105.85435485839844,
"step": 68
},
{
"epoch": 0.001725,
"grad_norm": 2.8943285942077637,
"grad_norm_var": 2.9569066036314915,
"learning_rate": 1.38e-05,
"loss": 105.765,
"loss/crossentropy": 2.750092029571533,
"loss/hidden": 0.0732421875,
"loss/logits": 0.01827952079474926,
"loss/reg": 105.67351531982422,
"step": 69
},
{
"epoch": 0.00175,
"grad_norm": 1.740116000175476,
"grad_norm_var": 2.993779212225659,
"learning_rate": 1.4e-05,
"loss": 105.5679,
"loss/crossentropy": 2.808587074279785,
"loss/hidden": 0.06396484375,
"loss/logits": 0.014089204370975494,
"loss/reg": 105.48979949951172,
"step": 70
},
{
"epoch": 0.001775,
"grad_norm": 3.365384101867676,
"grad_norm_var": 2.817206299433534,
"learning_rate": 1.4200000000000001e-05,
"loss": 105.4282,
"loss/crossentropy": 3.0790698528289795,
"loss/hidden": 0.109375,
"loss/logits": 0.01681649312376976,
"loss/reg": 105.30203247070312,
"step": 71
},
{
"epoch": 0.0018,
"grad_norm": 1.2450242042541504,
"grad_norm_var": 0.49637522068055445,
"learning_rate": 1.4400000000000001e-05,
"loss": 105.1742,
"loss/crossentropy": 2.7480971813201904,
"loss/hidden": 0.05712890625,
"loss/logits": 0.011092130094766617,
"loss/reg": 105.10601806640625,
"step": 72
},
{
"epoch": 0.001825,
"grad_norm": 1.7799638509750366,
"grad_norm_var": 0.48409086806424567,
"learning_rate": 1.46e-05,
"loss": 104.9855,
"loss/crossentropy": 2.8875551223754883,
"loss/hidden": 0.0615234375,
"loss/logits": 0.013171407394111156,
"loss/reg": 104.91083526611328,
"step": 73
},
{
"epoch": 0.00185,
"grad_norm": 3.0080645084381104,
"grad_norm_var": 0.4888197405301819,
"learning_rate": 1.48e-05,
"loss": 104.7925,
"loss/crossentropy": 2.947901725769043,
"loss/hidden": 0.06396484375,
"loss/logits": 0.015885423868894577,
"loss/reg": 104.71261596679688,
"step": 74
},
{
"epoch": 0.001875,
"grad_norm": 2.132399320602417,
"grad_norm_var": 0.46168637093159354,
"learning_rate": 1.5000000000000002e-05,
"loss": 104.5821,
"loss/crossentropy": 2.8153300285339355,
"loss/hidden": 0.059326171875,
"loss/logits": 0.012619425542652607,
"loss/reg": 104.51014709472656,
"step": 75
},
{
"epoch": 0.0019,
"grad_norm": 4.225804328918457,
"grad_norm_var": 0.642997495639734,
"learning_rate": 1.5200000000000002e-05,
"loss": 104.3881,
"loss/crossentropy": 3.0654807090759277,
"loss/hidden": 0.06591796875,
"loss/logits": 0.017460748553276062,
"loss/reg": 104.30471801757812,
"step": 76
},
{
"epoch": 0.001925,
"grad_norm": 1.3250384330749512,
"grad_norm_var": 0.6687036530869581,
"learning_rate": 1.54e-05,
"loss": 104.1632,
"loss/crossentropy": 2.7858409881591797,
"loss/hidden": 0.05712890625,
"loss/logits": 0.010722491890192032,
"loss/reg": 104.09539031982422,
"step": 77
},
{
"epoch": 0.00195,
"grad_norm": 2.416010618209839,
"grad_norm_var": 0.6340868082909626,
"learning_rate": 1.5600000000000003e-05,
"loss": 103.955,
"loss/crossentropy": 3.242889642715454,
"loss/hidden": 0.05908203125,
"loss/logits": 0.013778477907180786,
"loss/reg": 103.8821792602539,
"step": 78
},
{
"epoch": 0.001975,
"grad_norm": 2.7921998500823975,
"grad_norm_var": 0.6370527990787508,
"learning_rate": 1.58e-05,
"loss": 103.7406,
"loss/crossentropy": 2.815262794494629,
"loss/hidden": 0.0615234375,
"loss/logits": 0.012466374784708023,
"loss/reg": 103.66657257080078,
"step": 79
},
{
"epoch": 0.002,
"grad_norm": 2.4101223945617676,
"grad_norm_var": 0.6377332514460964,
"learning_rate": 1.6000000000000003e-05,
"loss": 103.5152,
"loss/crossentropy": 2.6628899574279785,
"loss/hidden": 0.05712890625,
"loss/logits": 0.011213818565011024,
"loss/reg": 103.44680786132812,
"step": 80
},
{
"epoch": 0.002025,
"grad_norm": 4.508922576904297,
"grad_norm_var": 0.9045377168218747,
"learning_rate": 1.62e-05,
"loss": 103.3373,
"loss/crossentropy": 2.809832811355591,
"loss/hidden": 0.099609375,
"loss/logits": 0.014138867147266865,
"loss/reg": 103.2235107421875,
"step": 81
},
{
"epoch": 0.00205,
"grad_norm": 2.5658440589904785,
"grad_norm_var": 0.8668634995421407,
"learning_rate": 1.64e-05,
"loss": 103.0584,
"loss/crossentropy": 2.9585628509521484,
"loss/hidden": 0.05712890625,
"loss/logits": 0.011128315702080727,
"loss/reg": 102.99018859863281,
"step": 82
},
{
"epoch": 0.002075,
"grad_norm": 1.03963041305542,
"grad_norm_var": 0.9844523199274171,
"learning_rate": 1.66e-05,
"loss": 102.8268,
"loss/crossentropy": 2.9794199466705322,
"loss/hidden": 0.05712890625,
"loss/logits": 0.010468900203704834,
"loss/reg": 102.75920104980469,
"step": 83
},
{
"epoch": 0.0021,
"grad_norm": 2.0196003913879395,
"grad_norm_var": 0.9875894888249034,
"learning_rate": 1.6800000000000002e-05,
"loss": 102.5951,
"loss/crossentropy": 3.0950756072998047,
"loss/hidden": 0.05712890625,
"loss/logits": 0.012872470542788506,
"loss/reg": 102.52505493164062,
"step": 84
},
{
"epoch": 0.002125,
"grad_norm": 2.780224323272705,
"grad_norm_var": 0.9818985175549242,
"learning_rate": 1.7e-05,
"loss": 102.3623,
"loss/crossentropy": 2.8365509510040283,
"loss/hidden": 0.0615234375,
"loss/logits": 0.013720070943236351,
"loss/reg": 102.28706359863281,
"step": 85
},
{
"epoch": 0.00215,
"grad_norm": 4.0338544845581055,
"grad_norm_var": 1.0906703730721057,
"learning_rate": 1.72e-05,
"loss": 102.1238,
"loss/crossentropy": 2.8407351970672607,
"loss/hidden": 0.06396484375,
"loss/logits": 0.013968261890113354,
"loss/reg": 102.04585266113281,
"step": 86
},
{
"epoch": 0.002175,
"grad_norm": 1.787563681602478,
"grad_norm_var": 1.0858789976663803,
"learning_rate": 1.7400000000000003e-05,
"loss": 101.8693,
"loss/crossentropy": 2.7350587844848633,
"loss/hidden": 0.05712890625,
"loss/logits": 0.010529162362217903,
"loss/reg": 101.80165100097656,
"step": 87
},
{
"epoch": 0.0022,
"grad_norm": 2.832834005355835,
"grad_norm_var": 0.9768321018939086,
"learning_rate": 1.76e-05,
"loss": 101.6227,
"loss/crossentropy": 2.9185311794281006,
"loss/hidden": 0.05712890625,
"loss/logits": 0.01262863352894783,
"loss/reg": 101.55294036865234,
"step": 88
},
{
"epoch": 0.002225,
"grad_norm": 3.1689658164978027,
"grad_norm_var": 0.944871850947833,
"learning_rate": 1.7800000000000002e-05,
"loss": 101.3782,
"loss/crossentropy": 3.041252374649048,
"loss/hidden": 0.0615234375,
"loss/logits": 0.01572200283408165,
"loss/reg": 101.30099487304688,
"step": 89
},
{
"epoch": 0.00225,
"grad_norm": 1.7799317836761475,
"grad_norm_var": 0.9871302861463281,
"learning_rate": 1.8e-05,
"loss": 101.1145,
"loss/crossentropy": 2.7331812381744385,
"loss/hidden": 0.05712890625,
"loss/logits": 0.011585134081542492,
"loss/reg": 101.04578399658203,
"step": 90
},
{
"epoch": 0.002275,
"grad_norm": 2.1665501594543457,
"grad_norm_var": 0.9850116745237391,
"learning_rate": 1.8200000000000002e-05,
"loss": 100.8592,
"loss/crossentropy": 2.4707322120666504,
"loss/hidden": 0.059326171875,
"loss/logits": 0.013475198298692703,
"loss/reg": 100.7863998413086,
"step": 91
},
{
"epoch": 0.0023,
"grad_norm": 1.3672914505004883,
"grad_norm_var": 0.8820831680751643,
"learning_rate": 1.8400000000000003e-05,
"loss": 100.5874,
"loss/crossentropy": 2.577828884124756,
"loss/hidden": 0.05712890625,
"loss/logits": 0.01066804863512516,
"loss/reg": 100.51959228515625,
"step": 92
},
{
"epoch": 0.002325,
"grad_norm": 1.8506819009780884,
"grad_norm_var": 0.8214079520174674,
"learning_rate": 1.86e-05,
"loss": 100.323,
"loss/crossentropy": 2.847485065460205,
"loss/hidden": 0.05712890625,
"loss/logits": 0.011566294357180595,
"loss/reg": 100.25430297851562,
"step": 93
},
{
"epoch": 0.00235,
"grad_norm": 3.947273015975952,
"grad_norm_var": 0.956929905592916,
"learning_rate": 1.88e-05,
"loss": 100.0834,
"loss/crossentropy": 2.833409070968628,
"loss/hidden": 0.0849609375,
"loss/logits": 0.013771452009677887,
"loss/reg": 99.98463439941406,
"step": 94
},
{
"epoch": 0.002375,
"grad_norm": 1.522775650024414,
"grad_norm_var": 1.0193112536277242,
"learning_rate": 1.9e-05,
"loss": 99.7793,
"loss/crossentropy": 2.616933822631836,
"loss/hidden": 0.05712890625,
"loss/logits": 0.01004397589713335,
"loss/reg": 99.71215057373047,
"step": 95
},
{
"epoch": 0.0024,
"grad_norm": 2.6764161586761475,
"grad_norm_var": 1.0210357175361253,
"learning_rate": 1.9200000000000003e-05,
"loss": 99.5082,
"loss/crossentropy": 3.357677936553955,
"loss/hidden": 0.059326171875,
"loss/logits": 0.01289614848792553,
"loss/reg": 99.43595886230469,
"step": 96
},
{
"epoch": 0.002425,
"grad_norm": 2.411071538925171,
"grad_norm_var": 0.7350196269207759,
"learning_rate": 1.94e-05,
"loss": 99.2327,
"loss/crossentropy": 2.9822490215301514,
"loss/hidden": 0.0615234375,
"loss/logits": 0.014493342489004135,
"loss/reg": 99.15665435791016,
"step": 97
},
{
"epoch": 0.00245,
"grad_norm": 5.154429912567139,
"grad_norm_var": 1.2207546039745256,
"learning_rate": 1.9600000000000002e-05,
"loss": 98.9606,
"loss/crossentropy": 2.878635883331299,
"loss/hidden": 0.068359375,
"loss/logits": 0.018285058438777924,
"loss/reg": 98.87391662597656,
"step": 98
},
{
"epoch": 0.002475,
"grad_norm": 5.486807823181152,
"grad_norm_var": 1.5709261870313864,
"learning_rate": 1.98e-05,
"loss": 98.6813,
"loss/crossentropy": 3.3747873306274414,
"loss/hidden": 0.07275390625,
"loss/logits": 0.02096904069185257,
"loss/reg": 98.58756256103516,
"step": 99
},
{
"epoch": 0.0025,
"grad_norm": 2.2651255130767822,
"grad_norm_var": 1.548765033441659,
"learning_rate": 2e-05,
"loss": 98.378,
"loss/crossentropy": 2.667487621307373,
"loss/hidden": 0.06396484375,
"loss/logits": 0.015711724758148193,
"loss/reg": 98.29834747314453,
"step": 100
},
{
"epoch": 0.002525,
"grad_norm": 2.8461413383483887,
"grad_norm_var": 1.548625602237729,
"learning_rate": 2e-05,
"loss": 98.0811,
"loss/crossentropy": 2.8070576190948486,
"loss/hidden": 0.05908203125,
"loss/logits": 0.01651117019355297,
"loss/reg": 98.00546264648438,
"step": 101
},
{
"epoch": 0.00255,
"grad_norm": 3.3563625812530518,
"grad_norm_var": 1.4686659287860662,
"learning_rate": 2e-05,
"loss": 97.79,
"loss/crossentropy": 3.0256845951080322,
"loss/hidden": 0.06591796875,
"loss/logits": 0.01792600378394127,
"loss/reg": 97.70619201660156,
"step": 102
},
{
"epoch": 0.002575,
"grad_norm": 3.0281784534454346,
"grad_norm_var": 1.3992474053637958,
"learning_rate": 2e-05,
"loss": 97.485,
"loss/crossentropy": 2.7928712368011475,
"loss/hidden": 0.05908203125,
"loss/logits": 0.016789790242910385,
"loss/reg": 97.40911102294922,
"step": 103
},
{
"epoch": 0.0026,
"grad_norm": 2.899623394012451,
"grad_norm_var": 1.399228163006243,
"learning_rate": 2e-05,
"loss": 97.1889,
"loss/crossentropy": 2.6017889976501465,
"loss/hidden": 0.0615234375,
"loss/logits": 0.014727754518389702,
"loss/reg": 97.11270141601562,
"step": 104
},
{
"epoch": 0.002625,
"grad_norm": 4.760987758636475,
"grad_norm_var": 1.6209967019940714,
"learning_rate": 2e-05,
"loss": 96.9561,
"loss/crossentropy": 2.79957914352417,
"loss/hidden": 0.1044921875,
"loss/logits": 0.036032598465681076,
"loss/reg": 96.81553649902344,
"step": 105
},
{
"epoch": 0.00265,
"grad_norm": 1.9602891206741333,
"grad_norm_var": 1.5944119405426551,
"learning_rate": 2e-05,
"loss": 96.5917,
"loss/crossentropy": 3.0477099418640137,
"loss/hidden": 0.059326171875,
"loss/logits": 0.014111923985183239,
"loss/reg": 96.51824951171875,
"step": 106
},
{
"epoch": 0.002675,
"grad_norm": 3.8924121856689453,
"grad_norm_var": 1.593099738565498,
"learning_rate": 2e-05,
"loss": 96.306,
"loss/crossentropy": 2.725721597671509,
"loss/hidden": 0.06591796875,
"loss/logits": 0.01935645565390587,
"loss/reg": 96.22071838378906,
"step": 107
},
{
"epoch": 0.0027,
"grad_norm": 2.1384472846984863,
"grad_norm_var": 1.45322790572492,
"learning_rate": 2e-05,
"loss": 95.9971,
"loss/crossentropy": 2.9953653812408447,
"loss/hidden": 0.05908203125,
"loss/logits": 0.015377204865217209,
"loss/reg": 95.92265319824219,
"step": 108
},
{
"epoch": 0.002725,
"grad_norm": 2.844248056411743,
"grad_norm_var": 1.3444790509526579,
"learning_rate": 2e-05,
"loss": 95.7164,
"loss/crossentropy": 3.295865297317505,
"loss/hidden": 0.06591796875,
"loss/logits": 0.02641688659787178,
"loss/reg": 95.62409973144531,
"step": 109
},
{
"epoch": 0.00275,
"grad_norm": 1.7702182531356812,
"grad_norm_var": 1.423617390900579,
"learning_rate": 2e-05,
"loss": 95.4,
"loss/crossentropy": 2.6738994121551514,
"loss/hidden": 0.059326171875,
"loss/logits": 0.014722645282745361,
"loss/reg": 95.32594299316406,
"step": 110
},
{
"epoch": 0.002775,
"grad_norm": 1.3760358095169067,
"grad_norm_var": 1.4551049156645277,
"learning_rate": 2e-05,
"loss": 95.0977,
"loss/crossentropy": 2.6674644947052,
"loss/hidden": 0.05712890625,
"loss/logits": 0.013482127338647842,
"loss/reg": 95.02709197998047,
"step": 111
},
{
"epoch": 0.0028,
"grad_norm": 1.5607060194015503,
"grad_norm_var": 1.589101356479648,
"learning_rate": 2e-05,
"loss": 94.8153,
"loss/crossentropy": 2.852834939956665,
"loss/hidden": 0.08251953125,
"loss/logits": 0.014033878222107887,
"loss/reg": 94.71875,
"step": 112
},
{
"epoch": 0.002825,
"grad_norm": 3.75045108795166,
"grad_norm_var": 1.5988275534787362,
"learning_rate": 2e-05,
"loss": 94.5009,
"loss/crossentropy": 2.9057774543762207,
"loss/hidden": 0.061279296875,
"loss/logits": 0.020055318251252174,
"loss/reg": 94.4195556640625,
"step": 113
},
{
"epoch": 0.00285,
"grad_norm": 2.267350912094116,
"grad_norm_var": 1.3166800230582432,
"learning_rate": 2e-05,
"loss": 94.1944,
"loss/crossentropy": 2.7482869625091553,
"loss/hidden": 0.056884765625,
"loss/logits": 0.017577335238456726,
"loss/reg": 94.1199722290039,
"step": 114
},
{
"epoch": 0.002875,
"grad_norm": 2.194246768951416,
"grad_norm_var": 0.8532155162290316,
"learning_rate": 2e-05,
"loss": 93.8912,
"loss/crossentropy": 2.743758201599121,
"loss/hidden": 0.05712890625,
"loss/logits": 0.013515879400074482,
"loss/reg": 93.82050323486328,
"step": 115
},
{
"epoch": 0.0029,
"grad_norm": 1.7801834344863892,
"grad_norm_var": 0.8948634812187672,
"learning_rate": 2e-05,
"loss": 93.5896,
"loss/crossentropy": 2.744076728820801,
"loss/hidden": 0.0546875,
"loss/logits": 0.014069687575101852,
"loss/reg": 93.5208740234375,
"step": 116
},
{
"epoch": 0.002925,
"grad_norm": 1.7759678363800049,
"grad_norm_var": 0.9386863932751007,
"learning_rate": 2e-05,
"loss": 93.2962,
"loss/crossentropy": 2.806662082672119,
"loss/hidden": 0.059326171875,
"loss/logits": 0.015871506184339523,
"loss/reg": 93.22098541259766,
"step": 117
},
{
"epoch": 0.00295,
"grad_norm": 2.6695425510406494,
"grad_norm_var": 0.8975061402584836,
"learning_rate": 2e-05,
"loss": 92.9929,
"loss/crossentropy": 2.797207832336426,
"loss/hidden": 0.056884765625,
"loss/logits": 0.01545205432921648,
"loss/reg": 92.92053985595703,
"step": 118
},
{
"epoch": 0.002975,
"grad_norm": 2.0127737522125244,
"grad_norm_var": 0.8960978495062031,
"learning_rate": 2e-05,
"loss": 92.6882,
"loss/crossentropy": 3.020660638809204,
"loss/hidden": 0.0546875,
"loss/logits": 0.013032155111432076,
"loss/reg": 92.62051391601562,
"step": 119
},
{
"epoch": 0.003,
"grad_norm": 2.1293652057647705,
"grad_norm_var": 0.88991297365724,
"learning_rate": 2e-05,
"loss": 92.3946,
"loss/crossentropy": 2.962179660797119,
"loss/hidden": 0.059326171875,
"loss/logits": 0.014865359291434288,
"loss/reg": 92.32038116455078,
"step": 120
},
{
"epoch": 0.003025,
"grad_norm": 2.659437894821167,
"grad_norm_var": 0.5128431927902019,
"learning_rate": 2e-05,
"loss": 92.0991,
"loss/crossentropy": 2.7553904056549072,
"loss/hidden": 0.0615234375,
"loss/logits": 0.017567995935678482,
"loss/reg": 92.01996612548828,
"step": 121
},
{
"epoch": 0.00305,
"grad_norm": 3.1532716751098633,
"grad_norm_var": 0.5479399334418377,
"learning_rate": 2e-05,
"loss": 91.7898,
"loss/crossentropy": 3.20906138420105,
"loss/hidden": 0.05908203125,
"loss/logits": 0.01771578937768936,
"loss/reg": 91.71299743652344,
"step": 122
},
{
"epoch": 0.003075,
"grad_norm": 2.2405543327331543,
"grad_norm_var": 0.3839241818330661,
"learning_rate": 2e-05,
"loss": 91.4828,
"loss/crossentropy": 2.6381375789642334,
"loss/hidden": 0.056884765625,
"loss/logits": 0.013064563274383545,
"loss/reg": 91.41288757324219,
"step": 123
},
{
"epoch": 0.0031,
"grad_norm": 3.476731300354004,
"grad_norm_var": 0.4723567159636746,
"learning_rate": 2e-05,
"loss": 91.1941,
"loss/crossentropy": 2.8266453742980957,
"loss/hidden": 0.06396484375,
"loss/logits": 0.01818031072616577,
"loss/reg": 91.11192321777344,
"step": 124
},
{
"epoch": 0.003125,
"grad_norm": 1.7917191982269287,
"grad_norm_var": 0.4727696793815556,
"learning_rate": 2e-05,
"loss": 90.8828,
"loss/crossentropy": 3.0270473957061768,
"loss/hidden": 0.05712890625,
"loss/logits": 0.013925662264227867,
"loss/reg": 90.81177520751953,
"step": 125
},
{
"epoch": 0.00315,
"grad_norm": 3.2372536659240723,
"grad_norm_var": 0.5059943836231217,
"learning_rate": 2e-05,
"loss": 90.5864,
"loss/crossentropy": 2.838686227798462,
"loss/hidden": 0.05908203125,
"loss/logits": 0.016410067677497864,
"loss/reg": 90.51092529296875,
"step": 126
},
{
"epoch": 0.003175,
"grad_norm": 1.6228362321853638,
"grad_norm_var": 0.4767731820654664,
"learning_rate": 2e-05,
"loss": 90.2798,
"loss/crossentropy": 2.6141273975372314,
"loss/hidden": 0.05712890625,
"loss/logits": 0.011823797598481178,
"loss/reg": 90.21080780029297,
"step": 127
},
{
"epoch": 0.0032,
"grad_norm": 2.675074577331543,
"grad_norm_var": 0.4304030863895955,
"learning_rate": 2e-05,
"loss": 89.9815,
"loss/crossentropy": 2.7882752418518066,
"loss/hidden": 0.05712890625,
"loss/logits": 0.014153627678751945,
"loss/reg": 89.91018676757812,
"step": 128
},
{
"epoch": 0.003225,
"grad_norm": 2.7944464683532715,
"grad_norm_var": 0.32364587258530453,
"learning_rate": 2e-05,
"loss": 89.685,
"loss/crossentropy": 2.5254697799682617,
"loss/hidden": 0.05908203125,
"loss/logits": 0.016538312658667564,
"loss/reg": 89.6093978881836,
"step": 129
},
{
"epoch": 0.00325,
"grad_norm": 1.8436682224273682,
"grad_norm_var": 0.34264366805227553,
"learning_rate": 2e-05,
"loss": 89.3803,
"loss/crossentropy": 2.928762197494507,
"loss/hidden": 0.056884765625,
"loss/logits": 0.014276012778282166,
"loss/reg": 89.30912017822266,
"step": 130
},
{
"epoch": 0.003275,
"grad_norm": 2.6479427814483643,
"grad_norm_var": 0.3443586211365847,
"learning_rate": 2e-05,
"loss": 89.084,
"loss/crossentropy": 2.7882349491119385,
"loss/hidden": 0.05908203125,
"loss/logits": 0.016245754435658455,
"loss/reg": 89.00865936279297,
"step": 131
},
{
"epoch": 0.0033,
"grad_norm": 4.339890956878662,
"grad_norm_var": 0.5399623455134483,
"learning_rate": 2e-05,
"loss": 88.7743,
"loss/crossentropy": 2.843156337738037,
"loss/hidden": 0.059326171875,
"loss/logits": 0.014006150886416435,
"loss/reg": 88.70095825195312,
"step": 132
},
{
"epoch": 0.003325,
"grad_norm": 1.3859492540359497,
"grad_norm_var": 0.5906001827105968,
"learning_rate": 2e-05,
"loss": 88.4686,
"loss/crossentropy": 2.56451153755188,
"loss/hidden": 0.05712890625,
"loss/logits": 0.011048642918467522,
"loss/reg": 88.40039825439453,
"step": 133
},
{
"epoch": 0.00335,
"grad_norm": 3.4068877696990967,
"grad_norm_var": 0.6370671369808453,
"learning_rate": 2e-05,
"loss": 88.1735,
"loss/crossentropy": 2.813845634460449,
"loss/hidden": 0.056884765625,
"loss/logits": 0.016296520829200745,
"loss/reg": 88.1003646850586,
"step": 134
},
{
"epoch": 0.003375,
"grad_norm": 2.2145471572875977,
"grad_norm_var": 0.6241198038182761,
"learning_rate": 2e-05,
"loss": 87.8709,
"loss/crossentropy": 2.9568309783935547,
"loss/hidden": 0.05712890625,
"loss/logits": 0.013690088875591755,
"loss/reg": 87.80007934570312,
"step": 135
},
{
"epoch": 0.0034,
"grad_norm": 2.6741232872009277,
"grad_norm_var": 0.6083942369195152,
"learning_rate": 2e-05,
"loss": 87.5719,
"loss/crossentropy": 2.7441060543060303,
"loss/hidden": 0.056884765625,
"loss/logits": 0.014466654509305954,
"loss/reg": 87.50055694580078,
"step": 136
},
{
"epoch": 0.003425,
"grad_norm": 2.338223934173584,
"grad_norm_var": 0.6138078516363993,
"learning_rate": 2e-05,
"loss": 87.2744,
"loss/crossentropy": 2.4464330673217773,
"loss/hidden": 0.059326171875,
"loss/logits": 0.014399213716387749,
"loss/reg": 87.20066833496094,
"step": 137
},
{
"epoch": 0.00345,
"grad_norm": 1.8222837448120117,
"grad_norm_var": 0.6290386070508833,
"learning_rate": 2e-05,
"loss": 86.9703,
"loss/crossentropy": 2.8742918968200684,
"loss/hidden": 0.05712890625,
"loss/logits": 0.011516067199409008,
"loss/reg": 86.90167999267578,
"step": 138
},
{
"epoch": 0.003475,
"grad_norm": 2.03706955909729,
"grad_norm_var": 0.6395340114512105,
"learning_rate": 2e-05,
"loss": 86.6734,
"loss/crossentropy": 2.651973009109497,
"loss/hidden": 0.056884765625,
"loss/logits": 0.01390259712934494,
"loss/reg": 86.60263061523438,
"step": 139
},
{
"epoch": 0.0035,
"grad_norm": 2.9460866451263428,
"grad_norm_var": 0.5893915505304466,
"learning_rate": 2e-05,
"loss": 86.4269,
"loss/crossentropy": 3.348736047744751,
"loss/hidden": 0.10693359375,
"loss/logits": 0.016606930643320084,
"loss/reg": 86.30339050292969,
"step": 140
},
{
"epoch": 0.003525,
"grad_norm": 2.397338390350342,
"grad_norm_var": 0.5562422152076701,
"learning_rate": 2e-05,
"loss": 86.0776,
"loss/crossentropy": 2.462433099746704,
"loss/hidden": 0.059326171875,
"loss/logits": 0.01363131683319807,
"loss/reg": 86.00459289550781,
"step": 141
},
{
"epoch": 0.00355,
"grad_norm": 2.2476322650909424,
"grad_norm_var": 0.5233350316059858,
"learning_rate": 2e-05,
"loss": 85.7713,
"loss/crossentropy": 2.6098899841308594,
"loss/hidden": 0.056884765625,
"loss/logits": 0.015527555719017982,
"loss/reg": 85.69889831542969,
"step": 142
},
{
"epoch": 0.003575,
"grad_norm": 1.603454828262329,
"grad_norm_var": 0.5255273885155527,
"learning_rate": 2e-05,
"loss": 85.4701,
"loss/crossentropy": 2.62600040435791,
"loss/hidden": 0.05712890625,
"loss/logits": 0.012892360799014568,
"loss/reg": 85.40011596679688,
"step": 143
},
{
"epoch": 0.0036,
"grad_norm": 1.814164400100708,
"grad_norm_var": 0.5472671850991829,
"learning_rate": 2e-05,
"loss": 85.1766,
"loss/crossentropy": 2.7545013427734375,
"loss/hidden": 0.059326171875,
"loss/logits": 0.015227731317281723,
"loss/reg": 85.10199737548828,
"step": 144
},
{
"epoch": 0.003625,
"grad_norm": 3.5215349197387695,
"grad_norm_var": 0.6178589741537791,
"learning_rate": 2e-05,
"loss": 84.8787,
"loss/crossentropy": 3.249708890914917,
"loss/hidden": 0.05908203125,
"loss/logits": 0.015877550467848778,
"loss/reg": 84.80371856689453,
"step": 145
},
{
"epoch": 0.00365,
"grad_norm": 1.9731359481811523,
"grad_norm_var": 0.6083958559047544,
"learning_rate": 2e-05,
"loss": 84.5763,
"loss/crossentropy": 3.2176551818847656,
"loss/hidden": 0.05712890625,
"loss/logits": 0.013086882419884205,
"loss/reg": 84.50607299804688,
"step": 146
},
{
"epoch": 0.003675,
"grad_norm": 4.117058753967285,
"grad_norm_var": 0.7799786660544029,
"learning_rate": 2e-05,
"loss": 84.2939,
"loss/crossentropy": 2.667962074279785,
"loss/hidden": 0.0634765625,
"loss/logits": 0.022213537245988846,
"loss/reg": 84.20823669433594,
"step": 147
},
{
"epoch": 0.0037,
"grad_norm": 3.0872607231140137,
"grad_norm_var": 0.5795145425234849,
"learning_rate": 2e-05,
"loss": 84.0006,
"loss/crossentropy": 2.904244899749756,
"loss/hidden": 0.07080078125,
"loss/logits": 0.01870722323656082,
"loss/reg": 83.91111755371094,
"step": 148
},
{
"epoch": 0.003725,
"grad_norm": 2.1129748821258545,
"grad_norm_var": 0.507061159842186,
"learning_rate": 2e-05,
"loss": 83.6913,
"loss/crossentropy": 2.4297919273376465,
"loss/hidden": 0.0615234375,
"loss/logits": 0.01596939191222191,
"loss/reg": 83.61376953125,
"step": 149
},
{
"epoch": 0.00375,
"grad_norm": 1.9368386268615723,
"grad_norm_var": 0.46821439415792837,
"learning_rate": 2e-05,
"loss": 83.3904,
"loss/crossentropy": 2.7391281127929688,
"loss/hidden": 0.059326171875,
"loss/logits": 0.014091565273702145,
"loss/reg": 83.31694030761719,
"step": 150
},
{
"epoch": 0.003775,
"grad_norm": 3.6663706302642822,
"grad_norm_var": 0.5586835942509762,
"learning_rate": 2e-05,
"loss": 83.1007,
"loss/crossentropy": 2.8713481426239014,
"loss/hidden": 0.0615234375,
"loss/logits": 0.01851273886859417,
"loss/reg": 83.02061462402344,
"step": 151
},
{
"epoch": 0.0038,
"grad_norm": 2.134021043777466,
"grad_norm_var": 0.5657064796702647,
"learning_rate": 2e-05,
"loss": 82.7923,
"loss/crossentropy": 3.2020580768585205,
"loss/hidden": 0.0615234375,
"loss/logits": 0.01641719415783882,
"loss/reg": 82.71440124511719,
"step": 152
},
{
"epoch": 0.003825,
"grad_norm": 2.897209405899048,
"grad_norm_var": 0.574317300050713,
"learning_rate": 2e-05,
"loss": 82.5031,
"loss/crossentropy": 2.975339412689209,
"loss/hidden": 0.06396484375,
"loss/logits": 0.021052952855825424,
"loss/reg": 82.41812133789062,
"step": 153
},
{
"epoch": 0.00385,
"grad_norm": 1.751135230064392,
"grad_norm_var": 0.5812492457165689,
"learning_rate": 2e-05,
"loss": 82.2003,
"loss/crossentropy": 2.38250994682312,
"loss/hidden": 0.0615234375,
"loss/logits": 0.016858752816915512,
"loss/reg": 82.12188720703125,
"step": 154
},
{
"epoch": 0.003875,
"grad_norm": 2.5146450996398926,
"grad_norm_var": 0.5650580150903565,
"learning_rate": 2e-05,
"loss": 81.9018,
"loss/crossentropy": 2.493891954421997,
"loss/hidden": 0.0615234375,
"loss/logits": 0.013709386810660362,
"loss/reg": 81.82661437988281,
"step": 155
},
{
"epoch": 0.0039,
"grad_norm": 2.2257137298583984,
"grad_norm_var": 0.5589724988129731,
"learning_rate": 2e-05,
"loss": 81.606,
"loss/crossentropy": 2.713038682937622,
"loss/hidden": 0.059326171875,
"loss/logits": 0.015239729546010494,
"loss/reg": 81.53138732910156,
"step": 156
},
{
"epoch": 0.003925,
"grad_norm": 5.219791412353516,
"grad_norm_var": 1.01821672382146,
"learning_rate": 2e-05,
"loss": 81.3182,
"loss/crossentropy": 2.7188966274261475,
"loss/hidden": 0.06396484375,
"loss/logits": 0.017340324819087982,
"loss/reg": 81.23693084716797,
"step": 157
},
{
"epoch": 0.00395,
"grad_norm": 1.601794958114624,
"grad_norm_var": 1.0812106477349142,
"learning_rate": 2e-05,
"loss": 81.0152,
"loss/crossentropy": 2.647388219833374,
"loss/hidden": 0.059326171875,
"loss/logits": 0.013692192733287811,
"loss/reg": 80.9421615600586,
"step": 158
},
{
"epoch": 0.003975,
"grad_norm": 2.963209390640259,
"grad_norm_var": 1.009555342899499,
"learning_rate": 2e-05,
"loss": 80.7271,
"loss/crossentropy": 2.849640369415283,
"loss/hidden": 0.05908203125,
"loss/logits": 0.019920967519283295,
"loss/reg": 80.64808654785156,
"step": 159
},
{
"epoch": 0.004,
"grad_norm": 2.3866851329803467,
"grad_norm_var": 0.9608132022483118,
"learning_rate": 2e-05,
"loss": 80.4313,
"loss/crossentropy": 2.6206085681915283,
"loss/hidden": 0.0615234375,
"loss/logits": 0.015263768844306469,
"loss/reg": 80.35447692871094,
"step": 160
},
{
"epoch": 0.004025,
"grad_norm": 2.6866133213043213,
"grad_norm_var": 0.9192531230990411,
"learning_rate": 2e-05,
"loss": 80.1358,
"loss/crossentropy": 2.6675243377685547,
"loss/hidden": 0.059326171875,
"loss/logits": 0.015180731192231178,
"loss/reg": 80.061279296875,
"step": 161
},
{
"epoch": 0.00405,
"grad_norm": 2.2380247116088867,
"grad_norm_var": 0.8978023926442598,
"learning_rate": 2e-05,
"loss": 79.8366,
"loss/crossentropy": 2.9008312225341797,
"loss/hidden": 0.059326171875,
"loss/logits": 0.014321748167276382,
"loss/reg": 79.76295471191406,
"step": 162
},
{
"epoch": 0.004075,
"grad_norm": 5.501147747039795,
"grad_norm_var": 1.2751311244587156,
"learning_rate": 2e-05,
"loss": 79.5506,
"loss/crossentropy": 2.774265766143799,
"loss/hidden": 0.06396484375,
"loss/logits": 0.016338951885700226,
"loss/reg": 79.47030639648438,
"step": 163
},
{
"epoch": 0.0041,
"grad_norm": 1.578844666481018,
"grad_norm_var": 1.3611156779384619,
"learning_rate": 2e-05,
"loss": 79.2495,
"loss/crossentropy": 2.8486151695251465,
"loss/hidden": 0.05712890625,
"loss/logits": 0.01422686967998743,
"loss/reg": 79.17816925048828,
"step": 164
},
{
"epoch": 0.004125,
"grad_norm": 1.8603459596633911,
"grad_norm_var": 1.3853304523501924,
"learning_rate": 2e-05,
"loss": 78.9609,
"loss/crossentropy": 2.9763364791870117,
"loss/hidden": 0.059326171875,
"loss/logits": 0.015356909483671188,
"loss/reg": 78.88626098632812,
"step": 165
},
{
"epoch": 0.00415,
"grad_norm": 2.4613988399505615,
"grad_norm_var": 1.3493160215050024,
"learning_rate": 2e-05,
"loss": 78.6768,
"loss/crossentropy": 2.6457760334014893,
"loss/hidden": 0.0615234375,
"loss/logits": 0.01994500868022442,
"loss/reg": 78.59528350830078,
"step": 166
},
{
"epoch": 0.004175,
"grad_norm": 2.5572216510772705,
"grad_norm_var": 1.2877918838448412,
"learning_rate": 2e-05,
"loss": 78.3792,
"loss/crossentropy": 2.726158380508423,
"loss/hidden": 0.059326171875,
"loss/logits": 0.015283918008208275,
"loss/reg": 78.30459594726562,
"step": 167
},
{
"epoch": 0.0042,
"grad_norm": 3.149622917175293,
"grad_norm_var": 1.2808819694844975,
"learning_rate": 2e-05,
"loss": 78.0974,
"loss/crossentropy": 2.464186906814575,
"loss/hidden": 0.068359375,
"loss/logits": 0.014741174876689911,
"loss/reg": 78.01426696777344,
"step": 168
},
{
"epoch": 0.004225,
"grad_norm": 1.4037340879440308,
"grad_norm_var": 1.3859120968246847,
"learning_rate": 2e-05,
"loss": 77.8036,
"loss/crossentropy": 2.8845841884613037,
"loss/hidden": 0.0615234375,
"loss/logits": 0.017704475671052933,
"loss/reg": 77.724365234375,
"step": 169
},
{
"epoch": 0.00425,
"grad_norm": 4.424335479736328,
"grad_norm_var": 1.518842252075294,
"learning_rate": 2e-05,
"loss": 77.5411,
"loss/crossentropy": 3.1228435039520264,
"loss/hidden": 0.0859375,
"loss/logits": 0.020280467346310616,
"loss/reg": 77.4349136352539,
"step": 170
},
{
"epoch": 0.004275,
"grad_norm": 1.2430707216262817,
"grad_norm_var": 1.6679938506527754,
"learning_rate": 2e-05,
"loss": 77.2194,
"loss/crossentropy": 2.7156295776367188,
"loss/hidden": 0.059326171875,
"loss/logits": 0.014160854741930962,
"loss/reg": 77.14595794677734,
"step": 171
},
{
"epoch": 0.0043,
"grad_norm": 2.0627477169036865,
"grad_norm_var": 1.6803689194951175,
"learning_rate": 2e-05,
"loss": 76.9251,
"loss/crossentropy": 2.817188262939453,
"loss/hidden": 0.059326171875,
"loss/logits": 0.01637103408575058,
"loss/reg": 76.84935760498047,
"step": 172
},
{
"epoch": 0.004325,
"grad_norm": 2.538249969482422,
"grad_norm_var": 1.2319590329252075,
"learning_rate": 2e-05,
"loss": 76.637,
"loss/crossentropy": 2.71427321434021,
"loss/hidden": 0.0615234375,
"loss/logits": 0.014527076855301857,
"loss/reg": 76.56098937988281,
"step": 173
},
{
"epoch": 0.00435,
"grad_norm": 1.848551869392395,
"grad_norm_var": 1.2048617288824057,
"learning_rate": 2e-05,
"loss": 76.3474,
"loss/crossentropy": 2.9161787033081055,
"loss/hidden": 0.059326171875,
"loss/logits": 0.015176617540419102,
"loss/reg": 76.27291870117188,
"step": 174
},
{
"epoch": 0.004375,
"grad_norm": 2.4840588569641113,
"grad_norm_var": 1.1932266879105464,
"learning_rate": 2e-05,
"loss": 76.0652,
"loss/crossentropy": 3.0494370460510254,
"loss/hidden": 0.0615234375,
"loss/logits": 0.018287423998117447,
"loss/reg": 75.98535919189453,
"step": 175
},
{
"epoch": 0.0044,
"grad_norm": 1.4861845970153809,
"grad_norm_var": 1.2607000339864975,
"learning_rate": 2e-05,
"loss": 75.7693,
"loss/crossentropy": 2.6618382930755615,
"loss/hidden": 0.05712890625,
"loss/logits": 0.013419630005955696,
"loss/reg": 75.69873046875,
"step": 176
},
{
"epoch": 0.004425,
"grad_norm": 2.652388334274292,
"grad_norm_var": 1.2597859494322088,
"learning_rate": 2e-05,
"loss": 75.4947,
"loss/crossentropy": 3.0685219764709473,
"loss/hidden": 0.0615234375,
"loss/logits": 0.020900297909975052,
"loss/reg": 75.41223907470703,
"step": 177
},
{
"epoch": 0.00445,
"grad_norm": 2.228562831878662,
"grad_norm_var": 1.2600818300469971,
"learning_rate": 2e-05,
"loss": 75.2042,
"loss/crossentropy": 2.6521048545837402,
"loss/hidden": 0.0615234375,
"loss/logits": 0.01636704057455063,
"loss/reg": 75.1263427734375,
"step": 178
},
{
"epoch": 0.004475,
"grad_norm": 2.145296096801758,
"grad_norm_var": 0.6065573977317332,
"learning_rate": 2e-05,
"loss": 74.9156,
"loss/crossentropy": 2.7815215587615967,
"loss/hidden": 0.059326171875,
"loss/logits": 0.01532922312617302,
"loss/reg": 74.84095764160156,
"step": 179
},
{
"epoch": 0.0045,
"grad_norm": 3.027330160140991,
"grad_norm_var": 0.6065639312435943,
"learning_rate": 2e-05,
"loss": 74.6397,
"loss/crossentropy": 2.828174591064453,
"loss/hidden": 0.06396484375,
"loss/logits": 0.020114243030548096,
"loss/reg": 74.55561828613281,
"step": 180
},
{
"epoch": 0.004525,
"grad_norm": 1.8776589632034302,
"grad_norm_var": 0.6054562283667906,
"learning_rate": 2e-05,
"loss": 74.3442,
"loss/crossentropy": 2.7055771350860596,
"loss/hidden": 0.059326171875,
"loss/logits": 0.01365465484559536,
"loss/reg": 74.27119445800781,
"step": 181
},
{
"epoch": 0.00455,
"grad_norm": 3.372591733932495,
"grad_norm_var": 0.6709551658510691,
"learning_rate": 2e-05,
"loss": 74.0724,
"loss/crossentropy": 3.0145347118377686,
"loss/hidden": 0.07275390625,
"loss/logits": 0.022765886038541794,
"loss/reg": 73.97685241699219,
"step": 182
},
{
"epoch": 0.004575,
"grad_norm": 1.4099830389022827,
"grad_norm_var": 0.7301368956932454,
"learning_rate": 2e-05,
"loss": 73.7664,
"loss/crossentropy": 2.795443058013916,
"loss/hidden": 0.059326171875,
"loss/logits": 0.013609878718852997,
"loss/reg": 73.69348907470703,
"step": 183
},
{
"epoch": 0.0046,
"grad_norm": 1.8791447877883911,
"grad_norm_var": 0.6929646819722675,
"learning_rate": 2e-05,
"loss": 73.4903,
"loss/crossentropy": 2.60862135887146,
"loss/hidden": 0.0615234375,
"loss/logits": 0.017894212156534195,
"loss/reg": 73.41085052490234,
"step": 184
},
{
"epoch": 0.004625,
"grad_norm": 4.824383735656738,
"grad_norm_var": 1.0359055758899136,
"learning_rate": 2e-05,
"loss": 73.2051,
"loss/crossentropy": 2.708460569381714,
"loss/hidden": 0.0615234375,
"loss/logits": 0.015174289233982563,
"loss/reg": 73.12835693359375,
"step": 185
},
{
"epoch": 0.00465,
"grad_norm": 1.6892882585525513,
"grad_norm_var": 0.7903900820577945,
"learning_rate": 2e-05,
"loss": 72.9203,
"loss/crossentropy": 2.5275192260742188,
"loss/hidden": 0.059326171875,
"loss/logits": 0.014362558722496033,
"loss/reg": 72.84661102294922,
"step": 186
},
{
"epoch": 0.004675,
"grad_norm": 2.1509838104248047,
"grad_norm_var": 0.7141933855878756,
"learning_rate": 2e-05,
"loss": 72.6431,
"loss/crossentropy": 2.828364849090576,
"loss/hidden": 0.0615234375,
"loss/logits": 0.016414958983659744,
"loss/reg": 72.56517028808594,
"step": 187
},
{
"epoch": 0.0047,
"grad_norm": 2.066053867340088,
"grad_norm_var": 0.7140653095859344,
"learning_rate": 2e-05,
"loss": 72.3592,
"loss/crossentropy": 2.918135166168213,
"loss/hidden": 0.059326171875,
"loss/logits": 0.015539775602519512,
"loss/reg": 72.2843246459961,
"step": 188
},
{
"epoch": 0.004725,
"grad_norm": 1.7003145217895508,
"grad_norm_var": 0.7374802094477738,
"learning_rate": 2e-05,
"loss": 72.0782,
"loss/crossentropy": 2.828505754470825,
"loss/hidden": 0.059326171875,
"loss/logits": 0.01498749665915966,
"loss/reg": 72.00386047363281,
"step": 189
},
{
"epoch": 0.00475,
"grad_norm": 2.66283917427063,
"grad_norm_var": 0.7296169710249801,
"learning_rate": 2e-05,
"loss": 71.8013,
"loss/crossentropy": 2.9118659496307373,
"loss/hidden": 0.0615234375,
"loss/logits": 0.015920046716928482,
"loss/reg": 71.723876953125,
"step": 190
},
{
"epoch": 0.004775,
"grad_norm": 2.2389256954193115,
"grad_norm_var": 0.7291075429880787,
"learning_rate": 2e-05,
"loss": 71.5216,
"loss/crossentropy": 2.6960675716400146,
"loss/hidden": 0.0615234375,
"loss/logits": 0.014915665611624718,
"loss/reg": 71.4451675415039,
"step": 191
},
{
"epoch": 0.0048,
"grad_norm": 2.652817487716675,
"grad_norm_var": 0.6816330763082435,
"learning_rate": 2e-05,
"loss": 71.2388,
"loss/crossentropy": 2.7465293407440186,
"loss/hidden": 0.0615234375,
"loss/logits": 0.01686110533773899,
"loss/reg": 71.160400390625,
"step": 192
},
{
"epoch": 0.004825,
"grad_norm": 2.2822165489196777,
"grad_norm_var": 0.6782911578411197,
"learning_rate": 2e-05,
"loss": 70.9582,
"loss/crossentropy": 2.776250123977661,
"loss/hidden": 0.059326171875,
"loss/logits": 0.016233820468187332,
"loss/reg": 70.88268280029297,
"step": 193
},
{
"epoch": 0.00485,
"grad_norm": 2.540499210357666,
"grad_norm_var": 0.6777404274996485,
"learning_rate": 2e-05,
"loss": 70.6821,
"loss/crossentropy": 2.798632860183716,
"loss/hidden": 0.0615234375,
"loss/logits": 0.015345659106969833,
"loss/reg": 70.60520935058594,
"step": 194
},
{
"epoch": 0.004875,
"grad_norm": 1.9856505393981934,
"grad_norm_var": 0.6849150734256606,
"learning_rate": 2e-05,
"loss": 70.405,
"loss/crossentropy": 2.6718432903289795,
"loss/hidden": 0.0615234375,
"loss/logits": 0.014745904132723808,
"loss/reg": 70.32874298095703,
"step": 195
},
{
"epoch": 0.0049,
"grad_norm": 2.3494439125061035,
"grad_norm_var": 0.6567124398874036,
"learning_rate": 2e-05,
"loss": 70.1346,
"loss/crossentropy": 2.9501113891601562,
"loss/hidden": 0.0615234375,
"loss/logits": 0.02032644674181938,
"loss/reg": 70.05279541015625,
"step": 196
},
{
"epoch": 0.004925,
"grad_norm": 3.5448572635650635,
"grad_norm_var": 0.7242858678478853,
"learning_rate": 2e-05,
"loss": 69.8519,
"loss/crossentropy": 2.6818275451660156,
"loss/hidden": 0.059326171875,
"loss/logits": 0.014945561066269875,
"loss/reg": 69.7776107788086,
"step": 197
},
{
"epoch": 0.00495,
"grad_norm": 1.8181650638580322,
"grad_norm_var": 0.6860304672396395,
"learning_rate": 2e-05,
"loss": 69.5793,
"loss/crossentropy": 2.56965708732605,
"loss/hidden": 0.0615234375,
"loss/logits": 0.014989674091339111,
"loss/reg": 69.50276947021484,
"step": 198
},
{
"epoch": 0.004975,
"grad_norm": 3.4242897033691406,
"grad_norm_var": 0.6838728374435107,
"learning_rate": 2e-05,
"loss": 69.3558,
"loss/crossentropy": 3.0659897327423096,
"loss/hidden": 0.109375,
"loss/logits": 0.017705179750919342,
"loss/reg": 69.22872161865234,
"step": 199
},
{
"epoch": 0.005,
"grad_norm": 1.973741888999939,
"grad_norm_var": 0.6767511902871185,
"learning_rate": 2e-05,
"loss": 69.0331,
"loss/crossentropy": 2.637833595275879,
"loss/hidden": 0.059326171875,
"loss/logits": 0.01859479956328869,
"loss/reg": 68.95521545410156,
"step": 200
},
{
"epoch": 0.005025,
"grad_norm": 2.6389999389648438,
"grad_norm_var": 0.2962159441771213,
"learning_rate": 2e-05,
"loss": 68.7623,
"loss/crossentropy": 3.0950043201446533,
"loss/hidden": 0.0615234375,
"loss/logits": 0.01863202080130577,
"loss/reg": 68.68214416503906,
"step": 201
},
{
"epoch": 0.00505,
"grad_norm": 2.583447217941284,
"grad_norm_var": 0.26652776132981815,
"learning_rate": 2e-05,
"loss": 68.4826,
"loss/crossentropy": 3.238170862197876,
"loss/hidden": 0.0615234375,
"loss/logits": 0.019120125100016594,
"loss/reg": 68.40199279785156,
"step": 202
},
{
"epoch": 0.005075,
"grad_norm": 1.9552291631698608,
"grad_norm_var": 0.275770096339541,
"learning_rate": 2e-05,
"loss": 68.2168,
"loss/crossentropy": 2.745107650756836,
"loss/hidden": 0.06591796875,
"loss/logits": 0.02080545760691166,
"loss/reg": 68.13004302978516,
"step": 203
},
{
"epoch": 0.0051,
"grad_norm": 2.0590291023254395,
"grad_norm_var": 0.2760869902247992,
"learning_rate": 2e-05,
"loss": 67.9385,
"loss/crossentropy": 2.9923768043518066,
"loss/hidden": 0.0615234375,
"loss/logits": 0.018183503299951553,
"loss/reg": 67.85881805419922,
"step": 204
},
{
"epoch": 0.005125,
"grad_norm": 2.838015079498291,
"grad_norm_var": 0.25074774510742726,
"learning_rate": 2e-05,
"loss": 67.6847,
"loss/crossentropy": 3.02968692779541,
"loss/hidden": 0.07080078125,
"loss/logits": 0.02540656551718712,
"loss/reg": 67.58853149414062,
"step": 205
},
{
"epoch": 0.00515,
"grad_norm": 3.082895278930664,
"grad_norm_var": 0.2724775294796672,
"learning_rate": 2e-05,
"loss": 67.4095,
"loss/crossentropy": 2.6776514053344727,
"loss/hidden": 0.07080078125,
"loss/logits": 0.020375534892082214,
"loss/reg": 67.3183364868164,
"step": 206
},
{
"epoch": 0.005175,
"grad_norm": 2.3748674392700195,
"grad_norm_var": 0.2689364182377325,
"learning_rate": 2e-05,
"loss": 67.1321,
"loss/crossentropy": 2.7331337928771973,
"loss/hidden": 0.06396484375,
"loss/logits": 0.018977412953972816,
"loss/reg": 67.04918670654297,
"step": 207
},
{
"epoch": 0.0052,
"grad_norm": 1.8422597646713257,
"grad_norm_var": 0.2941871012735906,
"learning_rate": 2e-05,
"loss": 66.8571,
"loss/crossentropy": 2.6536264419555664,
"loss/hidden": 0.0615234375,
"loss/logits": 0.01525039691478014,
"loss/reg": 66.78032684326172,
"step": 208
},
{
"epoch": 0.005225,
"grad_norm": 2.370894193649292,
"grad_norm_var": 0.29262559140895844,
"learning_rate": 2e-05,
"loss": 66.5972,
"loss/crossentropy": 2.880850076675415,
"loss/hidden": 0.06640625,
"loss/logits": 0.01811142824590206,
"loss/reg": 66.51265716552734,
"step": 209
},
{
"epoch": 0.00525,
"grad_norm": 1.8227527141571045,
"grad_norm_var": 0.3172526467082339,
"learning_rate": 2e-05,
"loss": 66.3284,
"loss/crossentropy": 2.6867871284484863,
"loss/hidden": 0.06396484375,
"loss/logits": 0.018997574225068092,
"loss/reg": 66.24544525146484,
"step": 210
},
{
"epoch": 0.005275,
"grad_norm": 1.5266575813293457,
"grad_norm_var": 0.3567894464636623,
"learning_rate": 2e-05,
"loss": 66.0535,
"loss/crossentropy": 2.711437940597534,
"loss/hidden": 0.059326171875,
"loss/logits": 0.014955010265111923,
"loss/reg": 65.97917175292969,
"step": 211
},
{
"epoch": 0.0053,
"grad_norm": 2.41679310798645,
"grad_norm_var": 0.3567280892475267,
"learning_rate": 2e-05,
"loss": 65.7917,
"loss/crossentropy": 2.9064691066741943,
"loss/hidden": 0.06396484375,
"loss/logits": 0.0197412371635437,
"loss/reg": 65.70800018310547,
"step": 212
},
{
"epoch": 0.005325,
"grad_norm": 1.972184181213379,
"grad_norm_var": 0.2695787564070026,
"learning_rate": 2e-05,
"loss": 65.5564,
"loss/crossentropy": 2.7818334102630615,
"loss/hidden": 0.0947265625,
"loss/logits": 0.019062824547290802,
"loss/reg": 65.4426040649414,
"step": 213
},
{
"epoch": 0.00535,
"grad_norm": 1.5892772674560547,
"grad_norm_var": 0.2873676086054596,
"learning_rate": 2e-05,
"loss": 65.259,
"loss/crossentropy": 2.737332344055176,
"loss/hidden": 0.0615234375,
"loss/logits": 0.019478369504213333,
"loss/reg": 65.17799377441406,
"step": 214
},
{
"epoch": 0.005375,
"grad_norm": 6.980996131896973,
"grad_norm_var": 1.6209131844775593,
"learning_rate": 2e-05,
"loss": 65.0488,
"loss/crossentropy": 2.853780746459961,
"loss/hidden": 0.095703125,
"loss/logits": 0.03903430327773094,
"loss/reg": 64.91402435302734,
"step": 215
},
{
"epoch": 0.0054,
"grad_norm": 3.549551486968994,
"grad_norm_var": 1.6651724517150548,
"learning_rate": 2e-05,
"loss": 64.7459,
"loss/crossentropy": 2.9141762256622314,
"loss/hidden": 0.07275390625,
"loss/logits": 0.022131089121103287,
"loss/reg": 64.6510238647461,
"step": 216
},
{
"epoch": 0.005425,
"grad_norm": 1.94115149974823,
"grad_norm_var": 1.6920030605329315,
"learning_rate": 2e-05,
"loss": 64.4759,
"loss/crossentropy": 2.94362473487854,
"loss/hidden": 0.068359375,
"loss/logits": 0.019080817699432373,
"loss/reg": 64.38845825195312,
"step": 217
},
{
"epoch": 0.00545,
"grad_norm": 2.473371744155884,
"grad_norm_var": 1.6923666873172403,
"learning_rate": 2e-05,
"loss": 64.2163,
"loss/crossentropy": 3.031843662261963,
"loss/hidden": 0.068359375,
"loss/logits": 0.02138814702630043,
"loss/reg": 64.12654113769531,
"step": 218
},
{
"epoch": 0.005475,
"grad_norm": 2.364009141921997,
"grad_norm_var": 1.6704069952290912,
"learning_rate": 2e-05,
"loss": 63.9948,
"loss/crossentropy": 2.869771718978882,
"loss/hidden": 0.109375,
"loss/logits": 0.020195169374346733,
"loss/reg": 63.86522674560547,
"step": 219
},
{
"epoch": 0.0055,
"grad_norm": 2.0074684619903564,
"grad_norm_var": 1.6741223453539247,
"learning_rate": 2e-05,
"loss": 63.6908,
"loss/crossentropy": 2.9813449382781982,
"loss/hidden": 0.06591796875,
"loss/logits": 0.01999806985259056,
"loss/reg": 63.60487365722656,
"step": 220
},
{
"epoch": 0.005525,
"grad_norm": 1.478585958480835,
"grad_norm_var": 1.741421135603261,
"learning_rate": 2e-05,
"loss": 63.4223,
"loss/crossentropy": 2.450387716293335,
"loss/hidden": 0.0615234375,
"loss/logits": 0.01574028469622135,
"loss/reg": 63.34505081176758,
"step": 221
},
{
"epoch": 0.00555,
"grad_norm": 1.767280101776123,
"grad_norm_var": 1.7450884712307488,
"learning_rate": 2e-05,
"loss": 63.165,
"loss/crossentropy": 2.9608945846557617,
"loss/hidden": 0.06591796875,
"loss/logits": 0.02006276696920395,
"loss/reg": 63.07899856567383,
"step": 222
},
{
"epoch": 0.005575,
"grad_norm": 2.900290012359619,
"grad_norm_var": 1.7602401150279823,
"learning_rate": 2e-05,
"loss": 62.9101,
"loss/crossentropy": 2.7069523334503174,
"loss/hidden": 0.068359375,
"loss/logits": 0.021145779639482498,
"loss/reg": 62.82056427001953,
"step": 223
},
{
"epoch": 0.0056,
"grad_norm": 2.2942614555358887,
"grad_norm_var": 1.7371226601830352,
"learning_rate": 2e-05,
"loss": 62.6463,
"loss/crossentropy": 2.82142972946167,
"loss/hidden": 0.06396484375,
"loss/logits": 0.01956920698285103,
"loss/reg": 62.56277847290039,
"step": 224
},
{
"epoch": 0.005625,
"grad_norm": 2.362708330154419,
"grad_norm_var": 1.7372306188924502,
"learning_rate": 2e-05,
"loss": 62.3916,
"loss/crossentropy": 2.910395622253418,
"loss/hidden": 0.06591796875,
"loss/logits": 0.020157648250460625,
"loss/reg": 62.30548095703125,
"step": 225
},
{
"epoch": 0.00565,
"grad_norm": 2.486630439758301,
"grad_norm_var": 1.7078860460967495,
"learning_rate": 2e-05,
"loss": 62.1414,
"loss/crossentropy": 3.0249392986297607,
"loss/hidden": 0.07080078125,
"loss/logits": 0.021398965269327164,
"loss/reg": 62.04918670654297,
"step": 226
},
{
"epoch": 0.005675,
"grad_norm": 2.1476340293884277,
"grad_norm_var": 1.6508215590496325,
"learning_rate": 2e-05,
"loss": 61.8835,
"loss/crossentropy": 3.085418462753296,
"loss/hidden": 0.068359375,
"loss/logits": 0.021710071712732315,
"loss/reg": 61.79338455200195,
"step": 227
},
{
"epoch": 0.0057,
"grad_norm": 1.5186433792114258,
"grad_norm_var": 1.7166830776668944,
"learning_rate": 2e-05,
"loss": 61.6184,
"loss/crossentropy": 2.4423468112945557,
"loss/hidden": 0.06396484375,
"loss/logits": 0.016175974160432816,
"loss/reg": 61.53825378417969,
"step": 228
},
{
"epoch": 0.005725,
"grad_norm": 2.6107263565063477,
"grad_norm_var": 1.6981119809293725,
"learning_rate": 2e-05,
"loss": 61.373,
"loss/crossentropy": 2.9523637294769287,
"loss/hidden": 0.068359375,
"loss/logits": 0.0207897387444973,
"loss/reg": 61.283809661865234,
"step": 229
},
{
"epoch": 0.00575,
"grad_norm": 1.9942364692687988,
"grad_norm_var": 1.6575925882854556,
"learning_rate": 2e-05,
"loss": 61.159,
"loss/crossentropy": 2.678663492202759,
"loss/hidden": 0.109375,
"loss/logits": 0.019178075715899467,
"loss/reg": 61.030418395996094,
"step": 230
},
{
"epoch": 0.005775,
"grad_norm": 1.4201483726501465,
"grad_norm_var": 0.30852930383016425,
"learning_rate": 2e-05,
"loss": 60.8589,
"loss/crossentropy": 2.6229424476623535,
"loss/hidden": 0.06396484375,
"loss/logits": 0.017583219334483147,
"loss/reg": 60.777366638183594,
"step": 231
},
{
"epoch": 0.0058,
"grad_norm": 1.8850197792053223,
"grad_norm_var": 0.18379847000516059,
"learning_rate": 2e-05,
"loss": 60.5975,
"loss/crossentropy": 2.717855930328369,
"loss/hidden": 0.0615234375,
"loss/logits": 0.018108602613210678,
"loss/reg": 60.51786422729492,
"step": 232
},
{
"epoch": 0.005825,
"grad_norm": 2.635830879211426,
"grad_norm_var": 0.19894452868075493,
"learning_rate": 2e-05,
"loss": 60.3529,
"loss/crossentropy": 2.717303991317749,
"loss/hidden": 0.06591796875,
"loss/logits": 0.02073700539767742,
"loss/reg": 60.26628112792969,
"step": 233
},
{
"epoch": 0.00585,
"grad_norm": 2.0215964317321777,
"grad_norm_var": 0.19202186958297318,
"learning_rate": 2e-05,
"loss": 60.1403,
"loss/crossentropy": 2.547003984451294,
"loss/hidden": 0.1044921875,
"loss/logits": 0.020180463790893555,
"loss/reg": 60.015602111816406,
"step": 234
},
{
"epoch": 0.005875,
"grad_norm": 2.0360617637634277,
"grad_norm_var": 0.18800595898524625,
"learning_rate": 2e-05,
"loss": 59.8916,
"loss/crossentropy": 2.7054409980773926,
"loss/hidden": 0.10693359375,
"loss/logits": 0.019324198365211487,
"loss/reg": 59.76530456542969,
"step": 235
},
{
"epoch": 0.0059,
"grad_norm": 1.9470698833465576,
"grad_norm_var": 0.18896257994485927,
"learning_rate": 2e-05,
"loss": 59.6092,
"loss/crossentropy": 2.5585479736328125,
"loss/hidden": 0.0703125,
"loss/logits": 0.02303473837673664,
"loss/reg": 59.515865325927734,
"step": 236
},
{
"epoch": 0.005925,
"grad_norm": 1.3322805166244507,
"grad_norm_var": 0.20230885388968142,
"learning_rate": 2e-05,
"loss": 59.3467,
"loss/crossentropy": 2.845449924468994,
"loss/hidden": 0.0615234375,
"loss/logits": 0.01773185282945633,
"loss/reg": 59.26749038696289,
"step": 237
},
{
"epoch": 0.00595,
"grad_norm": 1.921771764755249,
"grad_norm_var": 0.1972553683811412,
"learning_rate": 2e-05,
"loss": 59.1028,
"loss/crossentropy": 2.437079668045044,
"loss/hidden": 0.06591796875,
"loss/logits": 0.016928989440202713,
"loss/reg": 59.01997756958008,
"step": 238
},
{
"epoch": 0.005975,
"grad_norm": 2.337186574935913,
"grad_norm_var": 0.15658778213605107,
"learning_rate": 2e-05,
"loss": 58.8627,
"loss/crossentropy": 2.7914915084838867,
"loss/hidden": 0.068359375,
"loss/logits": 0.02136731520295143,
"loss/reg": 58.77294921875,
"step": 239
},
{
"epoch": 0.006,
"grad_norm": 1.9505559206008911,
"grad_norm_var": 0.1532120628996779,
"learning_rate": 2e-05,
"loss": 58.6135,
"loss/crossentropy": 2.8501033782958984,
"loss/hidden": 0.06591796875,
"loss/logits": 0.020856186747550964,
"loss/reg": 58.52672576904297,
"step": 240
},
{
"epoch": 0.006025,
"grad_norm": 1.3474467992782593,
"grad_norm_var": 0.17367998148026603,
"learning_rate": 2e-05,
"loss": 58.3605,
"loss/crossentropy": 2.4960248470306396,
"loss/hidden": 0.06396484375,
"loss/logits": 0.015542775392532349,
"loss/reg": 58.280967712402344,
"step": 241
},
{
"epoch": 0.00605,
"grad_norm": 2.888327121734619,
"grad_norm_var": 0.21119166554983568,
"learning_rate": 2e-05,
"loss": 58.123,
"loss/crossentropy": 2.714336395263672,
"loss/hidden": 0.07275390625,
"loss/logits": 0.02144978940486908,
"loss/reg": 58.028785705566406,
"step": 242
},
{
"epoch": 0.006075,
"grad_norm": 2.533353805541992,
"grad_norm_var": 0.22810067793098288,
"learning_rate": 2e-05,
"loss": 57.8713,
"loss/crossentropy": 2.735057830810547,
"loss/hidden": 0.06591796875,
"loss/logits": 0.020662324503064156,
"loss/reg": 57.78470993041992,
"step": 243
},
{
"epoch": 0.0061,
"grad_norm": 2.4481282234191895,
"grad_norm_var": 0.21949654966500648,
"learning_rate": 2e-05,
"loss": 57.6443,
"loss/crossentropy": 2.7660956382751465,
"loss/hidden": 0.0771484375,
"loss/logits": 0.025949764996767044,
"loss/reg": 57.541160583496094,
"step": 244
},
{
"epoch": 0.006125,
"grad_norm": 1.7993848323822021,
"grad_norm_var": 0.2034264459530992,
"learning_rate": 2e-05,
"loss": 57.3799,
"loss/crossentropy": 2.469017505645752,
"loss/hidden": 0.06396484375,
"loss/logits": 0.017876872792840004,
"loss/reg": 57.2980842590332,
"step": 245
},
{
"epoch": 0.00615,
"grad_norm": 2.93696928024292,
"grad_norm_var": 0.2543330785650771,
"learning_rate": 2e-05,
"loss": 57.1865,
"loss/crossentropy": 2.988600254058838,
"loss/hidden": 0.109375,
"loss/logits": 0.020821403712034225,
"loss/reg": 57.0562629699707,
"step": 246
},
{
"epoch": 0.006175,
"grad_norm": 1.6419404745101929,
"grad_norm_var": 0.23759642989912186,
"learning_rate": 2e-05,
"loss": 56.9006,
"loss/crossentropy": 2.8848021030426025,
"loss/hidden": 0.06591796875,
"loss/logits": 0.020043395459651947,
"loss/reg": 56.81468200683594,
"step": 247
},
{
"epoch": 0.0062,
"grad_norm": 1.7423299551010132,
"grad_norm_var": 0.24303384118528198,
"learning_rate": 2e-05,
"loss": 56.6711,
"loss/crossentropy": 2.8383066654205322,
"loss/hidden": 0.0751953125,
"loss/logits": 0.021477218717336655,
"loss/reg": 56.57439422607422,
"step": 248
},
{
"epoch": 0.006225,
"grad_norm": 1.8884074687957764,
"grad_norm_var": 0.22405312170182767,
"learning_rate": 2e-05,
"loss": 56.4196,
"loss/crossentropy": 3.002903461456299,
"loss/hidden": 0.06591796875,
"loss/logits": 0.019266938790678978,
"loss/reg": 56.33445739746094,
"step": 249
},
{
"epoch": 0.00625,
"grad_norm": 2.0407156944274902,
"grad_norm_var": 0.22400789294475007,
"learning_rate": 2e-05,
"loss": 56.1875,
"loss/crossentropy": 2.517186164855957,
"loss/hidden": 0.07080078125,
"loss/logits": 0.02101137302815914,
"loss/reg": 56.09565734863281,
"step": 250
},
{
"epoch": 0.006275,
"grad_norm": 2.173579216003418,
"grad_norm_var": 0.22494351474196836,
"learning_rate": 2e-05,
"loss": 55.945,
"loss/crossentropy": 2.853949785232544,
"loss/hidden": 0.068359375,
"loss/logits": 0.019413135945796967,
"loss/reg": 55.85718536376953,
"step": 251
},
{
"epoch": 0.0063,
"grad_norm": 10.485960960388184,
"grad_norm_var": 4.655585789275835,
"learning_rate": 2e-05,
"loss": 55.7284,
"loss/crossentropy": 2.9653632640838623,
"loss/hidden": 0.091796875,
"loss/logits": 0.025614306330680847,
"loss/reg": 55.61101150512695,
"step": 252
},
{
"epoch": 0.006325,
"grad_norm": 2.01010799407959,
"grad_norm_var": 4.570472437484742,
"learning_rate": 2e-05,
"loss": 55.461,
"loss/crossentropy": 2.8666720390319824,
"loss/hidden": 0.06591796875,
"loss/logits": 0.020927395671606064,
"loss/reg": 55.37415313720703,
"step": 253
},
{
"epoch": 0.00635,
"grad_norm": 2.971705913543701,
"grad_norm_var": 4.539645393263383,
"learning_rate": 2e-05,
"loss": 55.2367,
"loss/crossentropy": 3.0727896690368652,
"loss/hidden": 0.0771484375,
"loss/logits": 0.02156730368733406,
"loss/reg": 55.137996673583984,
"step": 254
},
{
"epoch": 0.006375,
"grad_norm": 2.025726556777954,
"grad_norm_var": 4.560765147149087,
"learning_rate": 2e-05,
"loss": 54.9904,
"loss/crossentropy": 2.8021695613861084,
"loss/hidden": 0.068359375,
"loss/logits": 0.019413193687796593,
"loss/reg": 54.90263366699219,
"step": 255
},
{
"epoch": 0.0064,
"grad_norm": 2.515801429748535,
"grad_norm_var": 4.525736863311493,
"learning_rate": 2e-05,
"loss": 54.7593,
"loss/crossentropy": 2.818976402282715,
"loss/hidden": 0.07080078125,
"loss/logits": 0.020367074757814407,
"loss/reg": 54.668121337890625,
"step": 256
},
{
"epoch": 0.006425,
"grad_norm": 6.981767177581787,
"grad_norm_var": 5.482006202620502,
"learning_rate": 2e-05,
"loss": 54.5586,
"loss/crossentropy": 2.819366693496704,
"loss/hidden": 0.10693359375,
"loss/logits": 0.01748763397336006,
"loss/reg": 54.434200286865234,
"step": 257
},
{
"epoch": 0.00645,
"grad_norm": 1.3410686254501343,
"grad_norm_var": 5.668649556530609,
"learning_rate": 2e-05,
"loss": 54.2886,
"loss/crossentropy": 2.730813503265381,
"loss/hidden": 0.068359375,
"loss/logits": 0.01936509646475315,
"loss/reg": 54.200897216796875,
"step": 258
},
{
"epoch": 0.006475,
"grad_norm": 1.8726478815078735,
"grad_norm_var": 5.734492101602018,
"learning_rate": 2e-05,
"loss": 54.0538,
"loss/crossentropy": 3.02189564704895,
"loss/hidden": 0.06591796875,
"loss/logits": 0.01970936357975006,
"loss/reg": 53.96820831298828,
"step": 259
},
{
"epoch": 0.0065,
"grad_norm": 3.454385995864868,
"grad_norm_var": 5.733156656528479,
"learning_rate": 2e-05,
"loss": 53.8456,
"loss/crossentropy": 2.5955543518066406,
"loss/hidden": 0.0751953125,
"loss/logits": 0.03384391963481903,
"loss/reg": 53.73658752441406,
"step": 260
},
{
"epoch": 0.006525,
"grad_norm": 1.4927403926849365,
"grad_norm_var": 5.787821586949178,
"learning_rate": 2e-05,
"loss": 53.59,
"loss/crossentropy": 2.7993125915527344,
"loss/hidden": 0.06591796875,
"loss/logits": 0.018266569823026657,
"loss/reg": 53.50581359863281,
"step": 261
},
{
"epoch": 0.00655,
"grad_norm": 2.202789068222046,
"grad_norm_var": 5.825085503943205,
"learning_rate": 2e-05,
"loss": 53.3542,
"loss/crossentropy": 2.6597094535827637,
"loss/hidden": 0.06591796875,
"loss/logits": 0.019392475485801697,
"loss/reg": 53.26885986328125,
"step": 262
},
{
"epoch": 0.006575,
"grad_norm": 2.4559764862060547,
"grad_norm_var": 5.72695782376293,
"learning_rate": 2e-05,
"loss": 53.1381,
"loss/crossentropy": 2.9647631645202637,
"loss/hidden": 0.0771484375,
"loss/logits": 0.02172122895717621,
"loss/reg": 53.03927993774414,
"step": 263
},
{
"epoch": 0.0066,
"grad_norm": 1.6547397375106812,
"grad_norm_var": 5.741873969332573,
"learning_rate": 2e-05,
"loss": 52.8978,
"loss/crossentropy": 2.8823652267456055,
"loss/hidden": 0.06591796875,
"loss/logits": 0.020957665517926216,
"loss/reg": 52.81092071533203,
"step": 264
},
{
"epoch": 0.006625,
"grad_norm": 2.558666706085205,
"grad_norm_var": 5.673023506020719,
"learning_rate": 2e-05,
"loss": 52.6838,
"loss/crossentropy": 2.993384599685669,
"loss/hidden": 0.0771484375,
"loss/logits": 0.023681573569774628,
"loss/reg": 52.58296203613281,
"step": 265
},
{
"epoch": 0.00665,
"grad_norm": 1.5857821702957153,
"grad_norm_var": 5.745050591443283,
"learning_rate": 2e-05,
"loss": 52.4834,
"loss/crossentropy": 2.769315242767334,
"loss/hidden": 0.10400390625,
"loss/logits": 0.023636557161808014,
"loss/reg": 52.35578918457031,
"step": 266
},
{
"epoch": 0.006675,
"grad_norm": 2.601391553878784,
"grad_norm_var": 5.710121188860965,
"learning_rate": 2e-05,
"loss": 52.2297,
"loss/crossentropy": 2.79392671585083,
"loss/hidden": 0.0751953125,
"loss/logits": 0.02494995854794979,
"loss/reg": 52.12955856323242,
"step": 267
},
{
"epoch": 0.0067,
"grad_norm": 1.7602055072784424,
"grad_norm_var": 1.7747363411260721,
"learning_rate": 2e-05,
"loss": 51.9948,
"loss/crossentropy": 2.9216957092285156,
"loss/hidden": 0.068359375,
"loss/logits": 0.022738244384527206,
"loss/reg": 51.903709411621094,
"step": 268
},
{
"epoch": 0.006725,
"grad_norm": 1.9448193311691284,
"grad_norm_var": 1.7789874166887367,
"learning_rate": 2e-05,
"loss": 51.7677,
"loss/crossentropy": 2.7518603801727295,
"loss/hidden": 0.068359375,
"loss/logits": 0.020492523908615112,
"loss/reg": 51.67887878417969,
"step": 269
},
{
"epoch": 0.00675,
"grad_norm": 5.49958610534668,
"grad_norm_var": 2.349575931185125,
"learning_rate": 2e-05,
"loss": 51.5654,
"loss/crossentropy": 2.48050856590271,
"loss/hidden": 0.0869140625,
"loss/logits": 0.0240333154797554,
"loss/reg": 51.454437255859375,
"step": 270
},
{
"epoch": 0.006775,
"grad_norm": 1.8433462381362915,
"grad_norm_var": 2.3661487125074525,
"learning_rate": 2e-05,
"loss": 51.34,
"loss/crossentropy": 2.705430030822754,
"loss/hidden": 0.08740234375,
"loss/logits": 0.021868586540222168,
"loss/reg": 51.230751037597656,
"step": 271
},
{
"epoch": 0.0068,
"grad_norm": 1.7715281248092651,
"grad_norm_var": 2.410153507700154,
"learning_rate": 2e-05,
"loss": 51.1084,
"loss/crossentropy": 2.9934589862823486,
"loss/hidden": 0.0849609375,
"loss/logits": 0.020941920578479767,
"loss/reg": 51.00248718261719,
"step": 272
},
{
"epoch": 0.006825,
"grad_norm": 2.3719260692596436,
"grad_norm_var": 1.0228592647114596,
"learning_rate": 2e-05,
"loss": 50.876,
"loss/crossentropy": 2.6865527629852295,
"loss/hidden": 0.07275390625,
"loss/logits": 0.022343173623085022,
"loss/reg": 50.78093338012695,
"step": 273
},
{
"epoch": 0.00685,
"grad_norm": 1.5276238918304443,
"grad_norm_var": 1.00178576807754,
"learning_rate": 2e-05,
"loss": 50.6493,
"loss/crossentropy": 2.901967763900757,
"loss/hidden": 0.068359375,
"loss/logits": 0.02106352336704731,
"loss/reg": 50.55984115600586,
"step": 274
},
{
"epoch": 0.006875,
"grad_norm": 1.6692343950271606,
"grad_norm_var": 1.0156202418807194,
"learning_rate": 2e-05,
"loss": 50.4286,
"loss/crossentropy": 2.760937213897705,
"loss/hidden": 0.068359375,
"loss/logits": 0.020720936357975006,
"loss/reg": 50.33951950073242,
"step": 275
},
{
"epoch": 0.0069,
"grad_norm": 1.5566948652267456,
"grad_norm_var": 0.9421993519803222,
"learning_rate": 2e-05,
"loss": 50.2053,
"loss/crossentropy": 2.9708621501922607,
"loss/hidden": 0.06591796875,
"loss/logits": 0.019328070804476738,
"loss/reg": 50.12004089355469,
"step": 276
},
{
"epoch": 0.006925,
"grad_norm": 2.321723222732544,
"grad_norm_var": 0.911832100377154,
"learning_rate": 2e-05,
"loss": 49.9954,
"loss/crossentropy": 2.669133186340332,
"loss/hidden": 0.07275390625,
"loss/logits": 0.021549934521317482,
"loss/reg": 49.90113830566406,
"step": 277
},
{
"epoch": 0.00695,
"grad_norm": 2.465369462966919,
"grad_norm_var": 0.9159632439841786,
"learning_rate": 2e-05,
"loss": 49.7791,
"loss/crossentropy": 2.7726550102233887,
"loss/hidden": 0.07275390625,
"loss/logits": 0.023049041628837585,
"loss/reg": 49.683250427246094,
"step": 278
},
{
"epoch": 0.006975,
"grad_norm": 2.3777921199798584,
"grad_norm_var": 0.9139300418415842,
"learning_rate": 2e-05,
"loss": 49.574,
"loss/crossentropy": 2.879091739654541,
"loss/hidden": 0.08447265625,
"loss/logits": 0.023941390216350555,
"loss/reg": 49.46562957763672,
"step": 279
},
{
"epoch": 0.007,
"grad_norm": 1.9317271709442139,
"grad_norm_var": 0.897871261555297,
"learning_rate": 2e-05,
"loss": 49.3497,
"loss/crossentropy": 2.8615469932556152,
"loss/hidden": 0.07958984375,
"loss/logits": 0.02096753567457199,
"loss/reg": 49.249176025390625,
"step": 280
},
{
"epoch": 0.007025,
"grad_norm": 1.8541237115859985,
"grad_norm_var": 0.8986510855344348,
"learning_rate": 2e-05,
"loss": 49.127,
"loss/crossentropy": 2.6375844478607178,
"loss/hidden": 0.07275390625,
"loss/logits": 0.020916959270834923,
"loss/reg": 49.03334426879883,
"step": 281
},
{
"epoch": 0.00705,
"grad_norm": 2.8278629779815674,
"grad_norm_var": 0.8945651245224343,
"learning_rate": 2e-05,
"loss": 48.9067,
"loss/crossentropy": 2.7122256755828857,
"loss/hidden": 0.0751953125,
"loss/logits": 0.020496461540460587,
"loss/reg": 48.81095886230469,
"step": 282
},
{
"epoch": 0.007075,
"grad_norm": 1.7860355377197266,
"grad_norm_var": 0.9001221835000724,
"learning_rate": 2e-05,
"loss": 48.687,
"loss/crossentropy": 2.636594772338867,
"loss/hidden": 0.07080078125,
"loss/logits": 0.019465088844299316,
"loss/reg": 48.596717834472656,
"step": 283
},
{
"epoch": 0.0071,
"grad_norm": 2.448721170425415,
"grad_norm_var": 0.8876001311071434,
"learning_rate": 2e-05,
"loss": 48.5043,
"loss/crossentropy": 2.9665403366088867,
"loss/hidden": 0.09912109375,
"loss/logits": 0.02214897982776165,
"loss/reg": 48.38304901123047,
"step": 284
},
{
"epoch": 0.007125,
"grad_norm": 1.8128719329833984,
"grad_norm_var": 0.8942751417055016,
"learning_rate": 2e-05,
"loss": 48.2649,
"loss/crossentropy": 2.886575698852539,
"loss/hidden": 0.0751953125,
"loss/logits": 0.0196707583963871,
"loss/reg": 48.17000198364258,
"step": 285
},
{
"epoch": 0.00715,
"grad_norm": 1.8738943338394165,
"grad_norm_var": 0.14694385548773212,
"learning_rate": 2e-05,
"loss": 48.093,
"loss/crossentropy": 2.9793410301208496,
"loss/hidden": 0.111328125,
"loss/logits": 0.023733166977763176,
"loss/reg": 47.957942962646484,
"step": 286
},
{
"epoch": 0.007175,
"grad_norm": 1.8546899557113647,
"grad_norm_var": 0.14667332129106872,
"learning_rate": 2e-05,
"loss": 47.8435,
"loss/crossentropy": 2.5512354373931885,
"loss/hidden": 0.0751953125,
"loss/logits": 0.021910887211561203,
"loss/reg": 47.74641036987305,
"step": 287
},
{
"epoch": 0.0072,
"grad_norm": 2.126343250274658,
"grad_norm_var": 0.1423970435536375,
"learning_rate": 2e-05,
"loss": 47.6355,
"loss/crossentropy": 2.801690101623535,
"loss/hidden": 0.0771484375,
"loss/logits": 0.022434473037719727,
"loss/reg": 47.53596496582031,
"step": 288
},
{
"epoch": 0.007225,
"grad_norm": 2.5778279304504395,
"grad_norm_var": 0.15387340759936208,
"learning_rate": 2e-05,
"loss": 47.4306,
"loss/crossentropy": 2.7154979705810547,
"loss/hidden": 0.07958984375,
"loss/logits": 0.025169990956783295,
"loss/reg": 47.32587814331055,
"step": 289
},
{
"epoch": 0.00725,
"grad_norm": 2.0250680446624756,
"grad_norm_var": 0.13381097704940614,
"learning_rate": 2e-05,
"loss": 47.2131,
"loss/crossentropy": 2.8832640647888184,
"loss/hidden": 0.0751953125,
"loss/logits": 0.021248571574687958,
"loss/reg": 47.11664962768555,
"step": 290
},
{
"epoch": 0.007275,
"grad_norm": 1.7236976623535156,
"grad_norm_var": 0.13090910370116995,
"learning_rate": 2e-05,
"loss": 47.0029,
"loss/crossentropy": 2.8528189659118652,
"loss/hidden": 0.07275390625,
"loss/logits": 0.021527128294110298,
"loss/reg": 46.90858459472656,
"step": 291
},
{
"epoch": 0.0073,
"grad_norm": 1.55807363986969,
"grad_norm_var": 0.1308097516741924,
"learning_rate": 2e-05,
"loss": 46.7919,
"loss/crossentropy": 2.664276361465454,
"loss/hidden": 0.0771484375,
"loss/logits": 0.021483037620782852,
"loss/reg": 46.69326400756836,
"step": 292
},
{
"epoch": 0.007325,
"grad_norm": 2.162757396697998,
"grad_norm_var": 0.12764433705414124,
"learning_rate": 2e-05,
"loss": 46.6093,
"loss/crossentropy": 2.663728952407837,
"loss/hidden": 0.09375,
"loss/logits": 0.028836317360401154,
"loss/reg": 46.486732482910156,
"step": 293
},
{
"epoch": 0.00735,
"grad_norm": 3.1162450313568115,
"grad_norm_var": 0.18687738678810295,
"learning_rate": 2e-05,
"loss": 46.3789,
"loss/crossentropy": 2.6197454929351807,
"loss/hidden": 0.0751953125,
"loss/logits": 0.02272343635559082,
"loss/reg": 46.28093719482422,
"step": 294
},
{
"epoch": 0.007375,
"grad_norm": 1.504614233970642,
"grad_norm_var": 0.20551894946539725,
"learning_rate": 2e-05,
"loss": 46.1705,
"loss/crossentropy": 2.695830821990967,
"loss/hidden": 0.07275390625,
"loss/logits": 0.02187586948275566,
"loss/reg": 46.075836181640625,
"step": 295
},
{
"epoch": 0.0074,
"grad_norm": 2.918551206588745,
"grad_norm_var": 0.24765848230744034,
"learning_rate": 2e-05,
"loss": 45.9933,
"loss/crossentropy": 2.9710702896118164,
"loss/hidden": 0.09326171875,
"loss/logits": 0.028397060930728912,
"loss/reg": 45.87164306640625,
"step": 296
},
{
"epoch": 0.007425,
"grad_norm": 2.2940146923065186,
"grad_norm_var": 0.24323678513681368,
"learning_rate": 2e-05,
"loss": 45.7702,
"loss/crossentropy": 2.80556058883667,
"loss/hidden": 0.0771484375,
"loss/logits": 0.025129806250333786,
"loss/reg": 45.66794204711914,
"step": 297
},
{
"epoch": 0.00745,
"grad_norm": 3.3687565326690674,
"grad_norm_var": 0.3094567617568638,
"learning_rate": 2e-05,
"loss": 45.5908,
"loss/crossentropy": 3.0187761783599854,
"loss/hidden": 0.09521484375,
"loss/logits": 0.03083011880517006,
"loss/reg": 45.46473693847656,
"step": 298
},
{
"epoch": 0.007475,
"grad_norm": 1.961309790611267,
"grad_norm_var": 0.3017723922679678,
"learning_rate": 2e-05,
"loss": 45.3716,
"loss/crossentropy": 2.801523447036743,
"loss/hidden": 0.083984375,
"loss/logits": 0.025450080633163452,
"loss/reg": 45.26218032836914,
"step": 299
},
{
"epoch": 0.0075,
"grad_norm": 2.01747989654541,
"grad_norm_var": 0.2995522458701202,
"learning_rate": 2e-05,
"loss": 45.1653,
"loss/crossentropy": 2.8323147296905518,
"loss/hidden": 0.07958984375,
"loss/logits": 0.02548621967434883,
"loss/reg": 45.06023406982422,
"step": 300
},
{
"epoch": 0.007525,
"grad_norm": 2.3311493396759033,
"grad_norm_var": 0.2909006236155415,
"learning_rate": 2e-05,
"loss": 44.9708,
"loss/crossentropy": 2.7468738555908203,
"loss/hidden": 0.083984375,
"loss/logits": 0.027555497363209724,
"loss/reg": 44.859291076660156,
"step": 301
},
{
"epoch": 0.00755,
"grad_norm": 2.1381912231445312,
"grad_norm_var": 0.2833022269660657,
"learning_rate": 2e-05,
"loss": 44.7534,
"loss/crossentropy": 2.472642421722412,
"loss/hidden": 0.0771484375,
"loss/logits": 0.022814348340034485,
"loss/reg": 44.65345001220703,
"step": 302
},
{
"epoch": 0.007575,
"grad_norm": 2.22025203704834,
"grad_norm_var": 0.2733649855868098,
"learning_rate": 2e-05,
"loss": 44.5945,
"loss/crossentropy": 2.7359273433685303,
"loss/hidden": 0.10791015625,
"loss/logits": 0.03234729915857315,
"loss/reg": 44.45427703857422,
"step": 303
},
{
"epoch": 0.0076,
"grad_norm": 1.7129307985305786,
"grad_norm_var": 0.29101574490811494,
"learning_rate": 2e-05,
"loss": 44.3613,
"loss/crossentropy": 2.7747325897216797,
"loss/hidden": 0.07958984375,
"loss/logits": 0.025640204548835754,
"loss/reg": 44.25602340698242,
"step": 304
},
{
"epoch": 0.007625,
"grad_norm": 1.9883134365081787,
"grad_norm_var": 0.2851551419506978,
"learning_rate": 2e-05,
"loss": 44.1729,
"loss/crossentropy": 2.8054683208465576,
"loss/hidden": 0.0888671875,
"loss/logits": 0.02528468333184719,
"loss/reg": 44.05876159667969,
"step": 305
},
{
"epoch": 0.00765,
"grad_norm": 1.863852858543396,
"grad_norm_var": 0.29032669692645163,
"learning_rate": 2e-05,
"loss": 43.9655,
"loss/crossentropy": 2.888584852218628,
"loss/hidden": 0.0771484375,
"loss/logits": 0.02664921060204506,
"loss/reg": 43.861751556396484,
"step": 306
},
{
"epoch": 0.007675,
"grad_norm": 2.4267005920410156,
"grad_norm_var": 0.27844298773661547,
"learning_rate": 2e-05,
"loss": 43.768,
"loss/crossentropy": 2.7776331901550293,
"loss/hidden": 0.0791015625,
"loss/logits": 0.023454634472727776,
"loss/reg": 43.6654052734375,
"step": 307
},
{
"epoch": 0.0077,
"grad_norm": 1.952818512916565,
"grad_norm_var": 0.25313514417236144,
"learning_rate": 2e-05,
"loss": 43.5914,
"loss/crossentropy": 2.8431732654571533,
"loss/hidden": 0.09375,
"loss/logits": 0.027756892144680023,
"loss/reg": 43.46989440917969,
"step": 308
},
{
"epoch": 0.007725,
"grad_norm": 1.5612258911132812,
"grad_norm_var": 0.2826367832773698,
"learning_rate": 2e-05,
"loss": 43.3769,
"loss/crossentropy": 2.791144847869873,
"loss/hidden": 0.07958984375,
"loss/logits": 0.02249729633331299,
"loss/reg": 43.274803161621094,
"step": 309
},
{
"epoch": 0.00775,
"grad_norm": 1.583145022392273,
"grad_norm_var": 0.24449754312349892,
"learning_rate": 2e-05,
"loss": 43.1754,
"loss/crossentropy": 2.503157615661621,
"loss/hidden": 0.07275390625,
"loss/logits": 0.021705985069274902,
"loss/reg": 43.08091354370117,
"step": 310
},
{
"epoch": 0.007775,
"grad_norm": 4.015507221221924,
"grad_norm_var": 0.4341163960002968,
"learning_rate": 2e-05,
"loss": 43.029,
"loss/crossentropy": 2.5782852172851562,
"loss/hidden": 0.11279296875,
"loss/logits": 0.028784021735191345,
"loss/reg": 42.88740921020508,
"step": 311
},
{
"epoch": 0.0078,
"grad_norm": 8.164732933044434,
"grad_norm_var": 2.6064283600035822,
"learning_rate": 2e-05,
"loss": 42.834,
"loss/crossentropy": 2.8343613147735596,
"loss/hidden": 0.10693359375,
"loss/logits": 0.03768664225935936,
"loss/reg": 42.68938446044922,
"step": 312
},
{
"epoch": 0.007825,
"grad_norm": 1.5896762609481812,
"grad_norm_var": 2.666172012313534,
"learning_rate": 2e-05,
"loss": 42.6196,
"loss/crossentropy": 2.701037883758545,
"loss/hidden": 0.09619140625,
"loss/logits": 0.0258449986577034,
"loss/reg": 42.497581481933594,
"step": 313
},
{
"epoch": 0.00785,
"grad_norm": 3.047536611557007,
"grad_norm_var": 2.6378112036190617,
"learning_rate": 2e-05,
"loss": 42.4423,
"loss/crossentropy": 3.052617073059082,
"loss/hidden": 0.10302734375,
"loss/logits": 0.032618869096040726,
"loss/reg": 42.30662536621094,
"step": 314
},
{
"epoch": 0.007875,
"grad_norm": 3.6461970806121826,
"grad_norm_var": 2.686150464102299,
"learning_rate": 2e-05,
"loss": 42.2375,
"loss/crossentropy": 2.7235336303710938,
"loss/hidden": 0.09375,
"loss/logits": 0.027443446218967438,
"loss/reg": 42.11628723144531,
"step": 315
},
{
"epoch": 0.0079,
"grad_norm": 2.1187753677368164,
"grad_norm_var": 2.6783673292138563,
"learning_rate": 2e-05,
"loss": 42.0521,
"loss/crossentropy": 3.022662401199341,
"loss/hidden": 0.0986328125,
"loss/logits": 0.02708546072244644,
"loss/reg": 41.926414489746094,
"step": 316
},
{
"epoch": 0.007925,
"grad_norm": 3.020916700363159,
"grad_norm_var": 2.6790032915758424,
"learning_rate": 2e-05,
"loss": 41.8743,
"loss/crossentropy": 3.040327787399292,
"loss/hidden": 0.10498046875,
"loss/logits": 0.03200379014015198,
"loss/reg": 41.737281799316406,
"step": 317
},
{
"epoch": 0.00795,
"grad_norm": 1.2161192893981934,
"grad_norm_var": 2.800065592587199,
"learning_rate": 2e-05,
"loss": 41.6418,
"loss/crossentropy": 2.455482006072998,
"loss/hidden": 0.07275390625,
"loss/logits": 0.01976654678583145,
"loss/reg": 41.54926300048828,
"step": 318
},
{
"epoch": 0.007975,
"grad_norm": 1.7299896478652954,
"grad_norm_var": 2.842071420926294,
"learning_rate": 2e-05,
"loss": 41.4748,
"loss/crossentropy": 2.58695650100708,
"loss/hidden": 0.08935546875,
"loss/logits": 0.023488441482186317,
"loss/reg": 41.36199951171875,
"step": 319
},
{
"epoch": 0.008,
"grad_norm": 1.4372546672821045,
"grad_norm_var": 2.8795153989340108,
"learning_rate": 2e-05,
"loss": 41.2741,
"loss/crossentropy": 2.516500473022461,
"loss/hidden": 0.0771484375,
"loss/logits": 0.021874364465475082,
"loss/reg": 41.175071716308594,
"step": 320
},
{
"epoch": 0.008025,
"grad_norm": 1.4945651292800903,
"grad_norm_var": 2.934045206445749,
"learning_rate": 2e-05,
"loss": 41.094,
"loss/crossentropy": 2.5344505310058594,
"loss/hidden": 0.08154296875,
"loss/logits": 0.023724202066659927,
"loss/reg": 40.98875427246094,
"step": 321
},
{
"epoch": 0.00805,
"grad_norm": 2.882326602935791,
"grad_norm_var": 2.9051136352536324,
"learning_rate": 2e-05,
"loss": 40.9209,
"loss/crossentropy": 2.9912266731262207,
"loss/hidden": 0.09375,
"loss/logits": 0.028768811374902725,
"loss/reg": 40.798377990722656,
"step": 322
},
{
"epoch": 0.008075,
"grad_norm": 2.139209270477295,
"grad_norm_var": 2.917611033862551,
"learning_rate": 2e-05,
"loss": 40.7388,
"loss/crossentropy": 3.03482985496521,
"loss/hidden": 0.0986328125,
"loss/logits": 0.02618756890296936,
"loss/reg": 40.61396789550781,
"step": 323
},
{
"epoch": 0.0081,
"grad_norm": 1.5827956199645996,
"grad_norm_var": 2.9580979264872167,
"learning_rate": 2e-05,
"loss": 40.5364,
"loss/crossentropy": 3.321441411972046,
"loss/hidden": 0.0771484375,
"loss/logits": 0.02882487140595913,
"loss/reg": 40.43039321899414,
"step": 324
},
{
"epoch": 0.008125,
"grad_norm": 3.9542465209960938,
"grad_norm_var": 2.9919451226799567,
"learning_rate": 2e-05,
"loss": 40.3997,
"loss/crossentropy": 2.867952585220337,
"loss/hidden": 0.11669921875,
"loss/logits": 0.03544551879167557,
"loss/reg": 40.24754333496094,
"step": 325
},
{
"epoch": 0.00815,
"grad_norm": 1.3926026821136475,
"grad_norm_var": 3.023260345272987,
"learning_rate": 2e-05,
"loss": 40.1766,
"loss/crossentropy": 2.859464645385742,
"loss/hidden": 0.08642578125,
"loss/logits": 0.02500808984041214,
"loss/reg": 40.06516647338867,
"step": 326
},
{
"epoch": 0.008175,
"grad_norm": 1.7433418035507202,
"grad_norm_var": 2.9517927278006177,
"learning_rate": 2e-05,
"loss": 39.9911,
"loss/crossentropy": 2.5700864791870117,
"loss/hidden": 0.08447265625,
"loss/logits": 0.02326771430671215,
"loss/reg": 39.883399963378906,
"step": 327
},
{
"epoch": 0.0082,
"grad_norm": 1.7338348627090454,
"grad_norm_var": 0.7415088588988434,
"learning_rate": 2e-05,
"loss": 39.823,
"loss/crossentropy": 3.0460493564605713,
"loss/hidden": 0.09375,
"loss/logits": 0.026862995699048042,
"loss/reg": 39.70237731933594,
"step": 328
},
{
"epoch": 0.008225,
"grad_norm": 2.071286916732788,
"grad_norm_var": 0.7187026357001777,
"learning_rate": 2e-05,
"loss": 39.6565,
"loss/crossentropy": 2.8687374591827393,
"loss/hidden": 0.10546875,
"loss/logits": 0.028884585946798325,
"loss/reg": 39.52215576171875,
"step": 329
},
{
"epoch": 0.00825,
"grad_norm": 3.727687120437622,
"grad_norm_var": 0.8244134178781979,
"learning_rate": 2e-05,
"loss": 39.4632,
"loss/crossentropy": 3.016721248626709,
"loss/hidden": 0.091796875,
"loss/logits": 0.02862412855029106,
"loss/reg": 39.34282684326172,
"step": 330
},
{
"epoch": 0.008275,
"grad_norm": 1.4077656269073486,
"grad_norm_var": 0.7188383933289703,
"learning_rate": 2e-05,
"loss": 39.2819,
"loss/crossentropy": 2.6854705810546875,
"loss/hidden": 0.09375,
"loss/logits": 0.02421317994594574,
"loss/reg": 39.163978576660156,
"step": 331
},
{
"epoch": 0.0083,
"grad_norm": 2.4871857166290283,
"grad_norm_var": 0.7280817035919266,
"learning_rate": 2e-05,
"loss": 39.1068,
"loss/crossentropy": 2.6913976669311523,
"loss/hidden": 0.09814453125,
"loss/logits": 0.030065573751926422,
"loss/reg": 38.9786262512207,
"step": 332
},
{
"epoch": 0.008325,
"grad_norm": 2.2533462047576904,
"grad_norm_var": 0.6733490639853538,
"learning_rate": 2e-05,
"loss": 38.9318,
"loss/crossentropy": 2.960143804550171,
"loss/hidden": 0.09814453125,
"loss/logits": 0.032106850296258926,
"loss/reg": 38.80150604248047,
"step": 333
},
{
"epoch": 0.00835,
"grad_norm": 9.487396240234375,
"grad_norm_var": 3.99832851363437,
"learning_rate": 2e-05,
"loss": 38.7644,
"loss/crossentropy": 2.7557554244995117,
"loss/hidden": 0.10986328125,
"loss/logits": 0.029345765709877014,
"loss/reg": 38.62519454956055,
"step": 334
},
{
"epoch": 0.008375,
"grad_norm": 2.2959508895874023,
"grad_norm_var": 3.9530502420570093,
"learning_rate": 2e-05,
"loss": 38.5785,
"loss/crossentropy": 2.8349740505218506,
"loss/hidden": 0.10009765625,
"loss/logits": 0.028996147215366364,
"loss/reg": 38.449363708496094,
"step": 335
},
{
"epoch": 0.0084,
"grad_norm": 2.147952079772949,
"grad_norm_var": 3.8715303432503254,
"learning_rate": 2e-05,
"loss": 38.4335,
"loss/crossentropy": 2.791761636734009,
"loss/hidden": 0.1259765625,
"loss/logits": 0.033383168280124664,
"loss/reg": 38.274166107177734,
"step": 336
},
{
"epoch": 0.008425,
"grad_norm": 2.610886573791504,
"grad_norm_var": 3.7737029983097736,
"learning_rate": 2e-05,
"loss": 38.2484,
"loss/crossentropy": 2.7260618209838867,
"loss/hidden": 0.109375,
"loss/logits": 0.03957097977399826,
"loss/reg": 38.09947204589844,
"step": 337
},
{
"epoch": 0.00845,
"grad_norm": 2.4849438667297363,
"grad_norm_var": 3.776289163852828,
"learning_rate": 2e-05,
"loss": 38.116,
"loss/crossentropy": 2.946385145187378,
"loss/hidden": 0.16015625,
"loss/logits": 0.03063117153942585,
"loss/reg": 37.9251708984375,
"step": 338
},
{
"epoch": 0.008475,
"grad_norm": 2.560534954071045,
"grad_norm_var": 3.754755415172888,
"learning_rate": 2e-05,
"loss": 37.9078,
"loss/crossentropy": 2.9236810207366943,
"loss/hidden": 0.1240234375,
"loss/logits": 0.03169061243534088,
"loss/reg": 37.75209045410156,
"step": 339
},
{
"epoch": 0.0085,
"grad_norm": 2.5575692653656006,
"grad_norm_var": 3.662913660444159,
"learning_rate": 2e-05,
"loss": 37.7183,
"loss/crossentropy": 3.100344181060791,
"loss/hidden": 0.10693359375,
"loss/logits": 0.03185933083295822,
"loss/reg": 37.57951354980469,
"step": 340
},
{
"epoch": 0.008525,
"grad_norm": 2.165684461593628,
"grad_norm_var": 3.5893262917243276,
"learning_rate": 2e-05,
"loss": 37.5408,
"loss/crossentropy": 2.663774251937866,
"loss/hidden": 0.10595703125,
"loss/logits": 0.027293076738715172,
"loss/reg": 37.407588958740234,
"step": 341
},
{
"epoch": 0.00855,
"grad_norm": 1.7527039051055908,
"grad_norm_var": 3.534874283949813,
"learning_rate": 2e-05,
"loss": 37.3585,
"loss/crossentropy": 2.862217903137207,
"loss/hidden": 0.10107421875,
"loss/logits": 0.025328852236270905,
"loss/reg": 37.23208236694336,
"step": 342
},
{
"epoch": 0.008575,
"grad_norm": 1.9454602003097534,
"grad_norm_var": 3.5111612253458673,
"learning_rate": 2e-05,
"loss": 37.1783,
"loss/crossentropy": 2.7598438262939453,
"loss/hidden": 0.09375,
"loss/logits": 0.023144034668803215,
"loss/reg": 37.061397552490234,
"step": 343
},
{
"epoch": 0.0086,
"grad_norm": 1.8846484422683716,
"grad_norm_var": 3.4925386021223233,
"learning_rate": 2e-05,
"loss": 37.0198,
"loss/crossentropy": 2.5700979232788086,
"loss/hidden": 0.1005859375,
"loss/logits": 0.027288008481264114,
"loss/reg": 36.89188003540039,
"step": 344
},
{
"epoch": 0.008625,
"grad_norm": 2.6235435009002686,
"grad_norm_var": 3.4623555366449352,
"learning_rate": 2e-05,
"loss": 36.8804,
"loss/crossentropy": 2.765270471572876,
"loss/hidden": 0.1259765625,
"loss/logits": 0.03164152801036835,
"loss/reg": 36.72273254394531,
"step": 345
},
{
"epoch": 0.00865,
"grad_norm": 1.9208968877792358,
"grad_norm_var": 3.4367772871203273,
"learning_rate": 2e-05,
"loss": 36.6959,
"loss/crossentropy": 3.0344183444976807,
"loss/hidden": 0.11572265625,
"loss/logits": 0.0260776337236166,
"loss/reg": 36.55405807495117,
"step": 346
},
{
"epoch": 0.008675,
"grad_norm": 2.031083583831787,
"grad_norm_var": 3.356850606629226,
"learning_rate": 2e-05,
"loss": 36.5175,
"loss/crossentropy": 2.8031561374664307,
"loss/hidden": 0.1005859375,
"loss/logits": 0.030941586941480637,
"loss/reg": 36.38599395751953,
"step": 347
},
{
"epoch": 0.0087,
"grad_norm": 2.3042380809783936,
"grad_norm_var": 3.3641485746872584,
"learning_rate": 2e-05,
"loss": 36.3562,
"loss/crossentropy": 2.727903366088867,
"loss/hidden": 0.107421875,
"loss/logits": 0.029567444697022438,
"loss/reg": 36.219173431396484,
"step": 348
},
{
"epoch": 0.008725,
"grad_norm": 2.40014910697937,
"grad_norm_var": 3.356964679129802,
"learning_rate": 2e-05,
"loss": 36.2139,
"loss/crossentropy": 3.2525317668914795,
"loss/hidden": 0.1298828125,
"loss/logits": 0.030907781794667244,
"loss/reg": 36.0531005859375,
"step": 349
},
{
"epoch": 0.00875,
"grad_norm": 2.395150661468506,
"grad_norm_var": 0.08076944890508499,
"learning_rate": 2e-05,
"loss": 36.0546,
"loss/crossentropy": 2.973644256591797,
"loss/hidden": 0.130859375,
"loss/logits": 0.03634234890341759,
"loss/reg": 35.887351989746094,
"step": 350
},
{
"epoch": 0.008775,
"grad_norm": 2.5341644287109375,
"grad_norm_var": 0.0856139565907459,
"learning_rate": 2e-05,
"loss": 35.8731,
"loss/crossentropy": 2.683912754058838,
"loss/hidden": 0.11962890625,
"loss/logits": 0.03108149953186512,
"loss/reg": 35.722354888916016,
"step": 351
},
{
"epoch": 0.0088,
"grad_norm": 2.0005195140838623,
"grad_norm_var": 0.08937117842573526,
"learning_rate": 2e-05,
"loss": 35.6898,
"loss/crossentropy": 3.0332581996917725,
"loss/hidden": 0.10498046875,
"loss/logits": 0.03197905793786049,
"loss/reg": 35.552879333496094,
"step": 352
},
{
"epoch": 0.008825,
"grad_norm": 2.4667744636535645,
"grad_norm_var": 0.08394155421554826,
"learning_rate": 2e-05,
"loss": 35.5462,
"loss/crossentropy": 2.7992653846740723,
"loss/hidden": 0.1279296875,
"loss/logits": 0.02878217026591301,
"loss/reg": 35.38950729370117,
"step": 353
},
{
"epoch": 0.00885,
"grad_norm": 1.8536417484283447,
"grad_norm_var": 0.0892220247213709,
"learning_rate": 2e-05,
"loss": 35.3542,
"loss/crossentropy": 2.6289925575256348,
"loss/hidden": 0.09814453125,
"loss/logits": 0.029167503118515015,
"loss/reg": 35.226863861083984,
"step": 354
},
{
"epoch": 0.008875,
"grad_norm": 2.0940122604370117,
"grad_norm_var": 0.08116332781685945,
"learning_rate": 2e-05,
"loss": 35.2059,
"loss/crossentropy": 2.63510799407959,
"loss/hidden": 0.1123046875,
"loss/logits": 0.028763707727193832,
"loss/reg": 35.06479263305664,
"step": 355
},
{
"epoch": 0.0089,
"grad_norm": 2.8152213096618652,
"grad_norm_var": 0.098175358135505,
"learning_rate": 2e-05,
"loss": 35.0467,
"loss/crossentropy": 2.7934632301330566,
"loss/hidden": 0.109375,
"loss/logits": 0.03379831463098526,
"loss/reg": 34.903560638427734,
"step": 356
},
{
"epoch": 0.008925,
"grad_norm": 2.6323490142822266,
"grad_norm_var": 0.10969825083783077,
"learning_rate": 2e-05,
"loss": 34.8794,
"loss/crossentropy": 2.75350284576416,
"loss/hidden": 0.10546875,
"loss/logits": 0.031125463545322418,
"loss/reg": 34.74283218383789,
"step": 357
},
{
"epoch": 0.00895,
"grad_norm": 1.6894372701644897,
"grad_norm_var": 0.11396125918644202,
"learning_rate": 2e-05,
"loss": 34.7202,
"loss/crossentropy": 2.6154894828796387,
"loss/hidden": 0.1103515625,
"loss/logits": 0.027205005288124084,
"loss/reg": 34.58259963989258,
"step": 358
},
{
"epoch": 0.008975,
"grad_norm": 1.4869511127471924,
"grad_norm_var": 0.14415693080625397,
"learning_rate": 2e-05,
"loss": 34.5381,
"loss/crossentropy": 2.6061604022979736,
"loss/hidden": 0.0888671875,
"loss/logits": 0.026232272386550903,
"loss/reg": 34.4229621887207,
"step": 359
},
{
"epoch": 0.009,
"grad_norm": 1.736647367477417,
"grad_norm_var": 0.1516660297286099,
"learning_rate": 2e-05,
"loss": 34.4049,
"loss/crossentropy": 2.993236780166626,
"loss/hidden": 0.107421875,
"loss/logits": 0.033529821783304214,
"loss/reg": 34.263916015625,
"step": 360
},
{
"epoch": 0.009025,
"grad_norm": 1.8005791902542114,
"grad_norm_var": 0.146044611712044,
"learning_rate": 2e-05,
"loss": 34.2384,
"loss/crossentropy": 2.6941475868225098,
"loss/hidden": 0.1025390625,
"loss/logits": 0.029957549646496773,
"loss/reg": 34.105873107910156,
"step": 361
},
{
"epoch": 0.00905,
"grad_norm": 2.542261838912964,
"grad_norm_var": 0.1524279525586619,
"learning_rate": 2e-05,
"loss": 34.0847,
"loss/crossentropy": 2.9499382972717285,
"loss/hidden": 0.10888671875,
"loss/logits": 0.03367718309164047,
"loss/reg": 33.942176818847656,
"step": 362
},
{
"epoch": 0.009075,
"grad_norm": 2.2937710285186768,
"grad_norm_var": 0.15173689243287544,
"learning_rate": 2e-05,
"loss": 33.9802,
"loss/crossentropy": 2.7008249759674072,
"loss/hidden": 0.1611328125,
"loss/logits": 0.03337787464261055,
"loss/reg": 33.78564453125,
"step": 363
},
{
"epoch": 0.0091,
"grad_norm": 2.6358730792999268,
"grad_norm_var": 0.16364591458108368,
"learning_rate": 2e-05,
"loss": 33.7769,
"loss/crossentropy": 2.773253917694092,
"loss/hidden": 0.1142578125,
"loss/logits": 0.033234164118766785,
"loss/reg": 33.6294059753418,
"step": 364
},
{
"epoch": 0.009125,
"grad_norm": 2.1610424518585205,
"grad_norm_var": 0.16119191833999386,
"learning_rate": 2e-05,
"loss": 33.6193,
"loss/crossentropy": 3.2032077312469482,
"loss/hidden": 0.109375,
"loss/logits": 0.03599859029054642,
"loss/reg": 33.4738883972168,
"step": 365
},
{
"epoch": 0.00915,
"grad_norm": 1.6486601829528809,
"grad_norm_var": 0.17621295368330056,
"learning_rate": 2e-05,
"loss": 33.4549,
"loss/crossentropy": 2.7943410873413086,
"loss/hidden": 0.10986328125,
"loss/logits": 0.0257001630961895,
"loss/reg": 33.31929397583008,
"step": 366
},
{
"epoch": 0.009175,
"grad_norm": 1.789995551109314,
"grad_norm_var": 0.17265670694948476,
"learning_rate": 2e-05,
"loss": 33.3092,
"loss/crossentropy": 2.974548101425171,
"loss/hidden": 0.11181640625,
"loss/logits": 0.032162584364414215,
"loss/reg": 33.16526794433594,
"step": 367
},
{
"epoch": 0.0092,
"grad_norm": 2.0894887447357178,
"grad_norm_var": 0.17193594057413286,
"learning_rate": 2e-05,
"loss": 33.1586,
"loss/crossentropy": 2.8424057960510254,
"loss/hidden": 0.11376953125,
"loss/logits": 0.032967403531074524,
"loss/reg": 33.01184844970703,
"step": 368
},
{
"epoch": 0.009225,
"grad_norm": 2.1341543197631836,
"grad_norm_var": 0.1629634187131666,
"learning_rate": 2e-05,
"loss": 32.9976,
"loss/crossentropy": 2.5876452922821045,
"loss/hidden": 0.10693359375,
"loss/logits": 0.031542833894491196,
"loss/reg": 32.859153747558594,
"step": 369
},
{
"epoch": 0.00925,
"grad_norm": 1.9316920042037964,
"grad_norm_var": 0.16090780810081545,
"learning_rate": 2e-05,
"loss": 32.8487,
"loss/crossentropy": 2.521399974822998,
"loss/hidden": 0.11181640625,
"loss/logits": 0.029829028993844986,
"loss/reg": 32.70704650878906,
"step": 370
},
{
"epoch": 0.009275,
"grad_norm": 1.597779631614685,
"grad_norm_var": 0.1762070126850133,
"learning_rate": 2e-05,
"loss": 32.7608,
"loss/crossentropy": 2.5190658569335938,
"loss/hidden": 0.1708984375,
"loss/logits": 0.03421615809202194,
"loss/reg": 32.55567932128906,
"step": 371
},
{
"epoch": 0.0093,
"grad_norm": 1.9022934436798096,
"grad_norm_var": 0.13656557084311008,
"learning_rate": 2e-05,
"loss": 32.5483,
"loss/crossentropy": 2.6691269874572754,
"loss/hidden": 0.119140625,
"loss/logits": 0.02910671941936016,
"loss/reg": 32.40000534057617,
"step": 372
},
{
"epoch": 0.009325,
"grad_norm": 1.3260115385055542,
"grad_norm_var": 0.1338758554160241,
"learning_rate": 2e-05,
"loss": 32.3699,
"loss/crossentropy": 2.8744542598724365,
"loss/hidden": 0.09326171875,
"loss/logits": 0.026506535708904266,
"loss/reg": 32.2501220703125,
"step": 373
},
{
"epoch": 0.00935,
"grad_norm": 2.3821325302124023,
"grad_norm_var": 0.14230117723275787,
"learning_rate": 2e-05,
"loss": 32.2718,
"loss/crossentropy": 2.809650421142578,
"loss/hidden": 0.13671875,
"loss/logits": 0.03423428162932396,
"loss/reg": 32.1008415222168,
"step": 374
},
{
"epoch": 0.009375,
"grad_norm": 2.7512526512145996,
"grad_norm_var": 0.16141473329477468,
"learning_rate": 2e-05,
"loss": 32.106,
"loss/crossentropy": 3.2307400703430176,
"loss/hidden": 0.12255859375,
"loss/logits": 0.03138088434934616,
"loss/reg": 31.952072143554688,
"step": 375
},
{
"epoch": 0.0094,
"grad_norm": 3.3555221557617188,
"grad_norm_var": 0.2586050041345979,
"learning_rate": 2e-05,
"loss": 31.9938,
"loss/crossentropy": 3.005693197250366,
"loss/hidden": 0.1533203125,
"loss/logits": 0.036577533930540085,
"loss/reg": 31.803926467895508,
"step": 376
},
{
"epoch": 0.009425,
"grad_norm": 3.3253965377807617,
"grad_norm_var": 0.3336118725112632,
"learning_rate": 2e-05,
"loss": 31.8469,
"loss/crossentropy": 3.547780990600586,
"loss/hidden": 0.1533203125,
"loss/logits": 0.03727081045508385,
"loss/reg": 31.656269073486328,
"step": 377
},
{
"epoch": 0.00945,
"grad_norm": 1.6626850366592407,
"grad_norm_var": 0.3467173050765467,
"learning_rate": 2e-05,
"loss": 31.6552,
"loss/crossentropy": 2.612881898880005,
"loss/hidden": 0.1162109375,
"loss/logits": 0.029575737193226814,
"loss/reg": 31.50940704345703,
"step": 378
},
{
"epoch": 0.009475,
"grad_norm": 3.107280969619751,
"grad_norm_var": 0.39968975146762575,
"learning_rate": 2e-05,
"loss": 31.5286,
"loss/crossentropy": 2.9622950553894043,
"loss/hidden": 0.1298828125,
"loss/logits": 0.03570305183529854,
"loss/reg": 31.363008499145508,
"step": 379
},
{
"epoch": 0.0095,
"grad_norm": 2.105592727661133,
"grad_norm_var": 0.38910356240631966,
"learning_rate": 2e-05,
"loss": 31.3657,
"loss/crossentropy": 2.7373220920562744,
"loss/hidden": 0.11865234375,
"loss/logits": 0.02971404604613781,
"loss/reg": 31.217342376708984,
"step": 380
},
{
"epoch": 0.009525,
"grad_norm": 6.658995628356934,
"grad_norm_var": 1.6275530318456253,
"learning_rate": 2e-05,
"loss": 31.2326,
"loss/crossentropy": 2.7819790840148926,
"loss/hidden": 0.1259765625,
"loss/logits": 0.03413031995296478,
"loss/reg": 31.072481155395508,
"step": 381
},
{
"epoch": 0.00955,
"grad_norm": 17.702425003051758,
"grad_norm_var": 15.943881150548076,
"learning_rate": 2e-05,
"loss": 31.1402,
"loss/crossentropy": 2.9263079166412354,
"loss/hidden": 0.1787109375,
"loss/logits": 0.037496551871299744,
"loss/reg": 30.92399787902832,
"step": 382
},
{
"epoch": 0.009575,
"grad_norm": 9.755738258361816,
"grad_norm_var": 18.105272629168468,
"learning_rate": 2e-05,
"loss": 30.9569,
"loss/crossentropy": 2.610964298248291,
"loss/hidden": 0.146484375,
"loss/logits": 0.03007764369249344,
"loss/reg": 30.780288696289062,
"step": 383
},
{
"epoch": 0.0096,
"grad_norm": 1.4994475841522217,
"grad_norm_var": 18.276295715224364,
"learning_rate": 2e-05,
"loss": 30.7982,
"loss/crossentropy": 2.600332498550415,
"loss/hidden": 0.1298828125,
"loss/logits": 0.031154906377196312,
"loss/reg": 30.637126922607422,
"step": 384
},
{
"epoch": 0.009625,
"grad_norm": 1.7103643417358398,
"grad_norm_var": 18.39011989648016,
"learning_rate": 2e-05,
"loss": 30.6776,
"loss/crossentropy": 3.032283306121826,
"loss/hidden": 0.1396484375,
"loss/logits": 0.04348542541265488,
"loss/reg": 30.494457244873047,
"step": 385
},
{
"epoch": 0.00965,
"grad_norm": 2.2822492122650146,
"grad_norm_var": 18.304705584234476,
"learning_rate": 2e-05,
"loss": 30.5332,
"loss/crossentropy": 2.367107629776001,
"loss/hidden": 0.1484375,
"loss/logits": 0.03226654976606369,
"loss/reg": 30.352514266967773,
"step": 386
},
{
"epoch": 0.009675,
"grad_norm": 2.002864122390747,
"grad_norm_var": 18.188167639061543,
"learning_rate": 2e-05,
"loss": 30.3683,
"loss/crossentropy": 2.7061312198638916,
"loss/hidden": 0.12353515625,
"loss/logits": 0.03370695561170578,
"loss/reg": 30.211055755615234,
"step": 387
},
{
"epoch": 0.0097,
"grad_norm": 5.602196216583252,
"grad_norm_var": 18.023389822229017,
"learning_rate": 2e-05,
"loss": 30.2394,
"loss/crossentropy": 2.397667169570923,
"loss/hidden": 0.138671875,
"loss/logits": 0.03049187734723091,
"loss/reg": 30.070276260375977,
"step": 388
},
{
"epoch": 0.009725,
"grad_norm": 1.5408035516738892,
"grad_norm_var": 17.943911354217352,
"learning_rate": 2e-05,
"loss": 30.0907,
"loss/crossentropy": 2.834867000579834,
"loss/hidden": 0.1298828125,
"loss/logits": 0.03077739104628563,
"loss/reg": 29.930038452148438,
"step": 389
},
{
"epoch": 0.00975,
"grad_norm": 6.018068313598633,
"grad_norm_var": 17.881454834941156,
"learning_rate": 2e-05,
"loss": 29.9551,
"loss/crossentropy": 2.8220815658569336,
"loss/hidden": 0.1328125,
"loss/logits": 0.031766001135110855,
"loss/reg": 29.790555953979492,
"step": 390
},
{
"epoch": 0.009775,
"grad_norm": 10.923192977905273,
"grad_norm_var": 20.212413139020512,
"learning_rate": 2e-05,
"loss": 29.8563,
"loss/crossentropy": 2.811016321182251,
"loss/hidden": 0.1689453125,
"loss/logits": 0.03583759814500809,
"loss/reg": 29.651498794555664,
"step": 391
},
{
"epoch": 0.0098,
"grad_norm": 1.9290435314178467,
"grad_norm_var": 20.64348377939481,
"learning_rate": 2e-05,
"loss": 29.6891,
"loss/crossentropy": 2.863645076751709,
"loss/hidden": 0.1435546875,
"loss/logits": 0.036813415586948395,
"loss/reg": 29.50878143310547,
"step": 392
},
{
"epoch": 0.009825,
"grad_norm": 3.237116813659668,
"grad_norm_var": 20.662082917547686,
"learning_rate": 2e-05,
"loss": 29.5663,
"loss/crossentropy": 2.7489192485809326,
"loss/hidden": 0.158203125,
"loss/logits": 0.03718395531177521,
"loss/reg": 29.370952606201172,
"step": 393
},
{
"epoch": 0.00985,
"grad_norm": 1.6290416717529297,
"grad_norm_var": 20.676489967742715,
"learning_rate": 2e-05,
"loss": 29.3821,
"loss/crossentropy": 2.96504282951355,
"loss/hidden": 0.11181640625,
"loss/logits": 0.03646458685398102,
"loss/reg": 29.2337703704834,
"step": 394
},
{
"epoch": 0.009875,
"grad_norm": 2.237454652786255,
"grad_norm_var": 20.926649282212274,
"learning_rate": 2e-05,
"loss": 29.269,
"loss/crossentropy": 2.8521690368652344,
"loss/hidden": 0.1376953125,
"loss/logits": 0.03403034061193466,
"loss/reg": 29.097240447998047,
"step": 395
},
{
"epoch": 0.0099,
"grad_norm": 1.9784162044525146,
"grad_norm_var": 20.97338552568199,
"learning_rate": 2e-05,
"loss": 29.1487,
"loss/crossentropy": 2.748701572418213,
"loss/hidden": 0.1533203125,
"loss/logits": 0.034278061240911484,
"loss/reg": 28.961105346679688,
"step": 396
},
{
"epoch": 0.009925,
"grad_norm": 2.4967384338378906,
"grad_norm_var": 21.021265946892495,
"learning_rate": 2e-05,
"loss": 29.011,
"loss/crossentropy": 3.1087985038757324,
"loss/hidden": 0.1455078125,
"loss/logits": 0.03961321711540222,
"loss/reg": 28.825891494750977,
"step": 397
},
{
"epoch": 0.00995,
"grad_norm": 2.17684006690979,
"grad_norm_var": 8.826986086885336,
"learning_rate": 2e-05,
"loss": 28.8691,
"loss/crossentropy": 2.919372320175171,
"loss/hidden": 0.1416015625,
"loss/logits": 0.03645738214254379,
"loss/reg": 28.69107437133789,
"step": 398
},
{
"epoch": 0.009975,
"grad_norm": 2.1314728260040283,
"grad_norm_var": 6.165466501526623,
"learning_rate": 2e-05,
"loss": 28.7475,
"loss/crossentropy": 2.795255422592163,
"loss/hidden": 0.1533203125,
"loss/logits": 0.037487540394067764,
"loss/reg": 28.556673049926758,
"step": 399
},
{
"epoch": 0.01,
"grad_norm": 1.821183204650879,
"grad_norm_var": 6.103824283803898,
"learning_rate": 2e-05,
"loss": 28.5859,
"loss/crossentropy": 2.6662023067474365,
"loss/hidden": 0.1279296875,
"loss/logits": 0.034927770495414734,
"loss/reg": 28.423030853271484,
"step": 400
},
{
"epoch": 0.010025,
"grad_norm": 1.682054042816162,
"grad_norm_var": 6.109147456078416,
"learning_rate": 2e-05,
"loss": 28.4353,
"loss/crossentropy": 3.024109125137329,
"loss/hidden": 0.11181640625,
"loss/logits": 0.03378160297870636,
"loss/reg": 28.28972625732422,
"step": 401
},
{
"epoch": 0.01005,
"grad_norm": 2.1850829124450684,
"grad_norm_var": 6.120403763567225,
"learning_rate": 2e-05,
"loss": 28.3327,
"loss/crossentropy": 3.1913986206054688,
"loss/hidden": 0.1435546875,
"loss/logits": 0.03758620843291283,
"loss/reg": 28.151596069335938,
"step": 402
},
{
"epoch": 0.010075,
"grad_norm": 2.2850840091705322,
"grad_norm_var": 6.084117119532081,
"learning_rate": 2e-05,
"loss": 28.2094,
"loss/crossentropy": 2.844414710998535,
"loss/hidden": 0.1484375,
"loss/logits": 0.04106660932302475,
"loss/reg": 28.019901275634766,
"step": 403
},
{
"epoch": 0.0101,
"grad_norm": 3.0755233764648438,
"grad_norm_var": 5.6459224869144,
"learning_rate": 2e-05,
"loss": 28.1054,
"loss/crossentropy": 2.7236874103546143,
"loss/hidden": 0.169921875,
"loss/logits": 0.04699024558067322,
"loss/reg": 27.88851547241211,
"step": 404
},
{
"epoch": 0.010125,
"grad_norm": 2.632650852203369,
"grad_norm_var": 5.513941759398255,
"learning_rate": 2e-05,
"loss": 27.9764,
"loss/crossentropy": 2.971388578414917,
"loss/hidden": 0.1787109375,
"loss/logits": 0.039863601326942444,
"loss/reg": 27.757869720458984,
"step": 405
},
{
"epoch": 0.01015,
"grad_norm": 2.490246057510376,
"grad_norm_var": 4.8850644555352325,
"learning_rate": 2e-05,
"loss": 27.8153,
"loss/crossentropy": 3.0535378456115723,
"loss/hidden": 0.1513671875,
"loss/logits": 0.03618897125124931,
"loss/reg": 27.62775421142578,
"step": 406
},
{
"epoch": 0.010175,
"grad_norm": 3.116687536239624,
"grad_norm_var": 0.24597344619940412,
"learning_rate": 2e-05,
"loss": 27.723,
"loss/crossentropy": 3.0417606830596924,
"loss/hidden": 0.1748046875,
"loss/logits": 0.049759428948163986,
"loss/reg": 27.49844741821289,
"step": 407
},
{
"epoch": 0.0102,
"grad_norm": 2.1385886669158936,
"grad_norm_var": 0.2378215272062145,
"learning_rate": 2e-05,
"loss": 27.5528,
"loss/crossentropy": 3.0288665294647217,
"loss/hidden": 0.1455078125,
"loss/logits": 0.03794676810503006,
"loss/reg": 27.369342803955078,
"step": 408
},
{
"epoch": 0.010225,
"grad_norm": 2.3155360221862793,
"grad_norm_var": 0.1797017907210517,
"learning_rate": 2e-05,
"loss": 27.4314,
"loss/crossentropy": 2.7471296787261963,
"loss/hidden": 0.1484375,
"loss/logits": 0.041661813855171204,
"loss/reg": 27.241291046142578,
"step": 409
},
{
"epoch": 0.01025,
"grad_norm": 2.1539342403411865,
"grad_norm_var": 0.15174583963641694,
"learning_rate": 2e-05,
"loss": 27.3053,
"loss/crossentropy": 3.1815056800842285,
"loss/hidden": 0.1484375,
"loss/logits": 0.0430469810962677,
"loss/reg": 27.11379051208496,
"step": 410
},
{
"epoch": 0.010275,
"grad_norm": 3.004754066467285,
"grad_norm_var": 0.1813925430056429,
"learning_rate": 2e-05,
"loss": 27.1907,
"loss/crossentropy": 2.7185699939727783,
"loss/hidden": 0.166015625,
"loss/logits": 0.038052164018154144,
"loss/reg": 26.986587524414062,
"step": 411
},
{
"epoch": 0.0103,
"grad_norm": 1.7727258205413818,
"grad_norm_var": 0.19437299657180346,
"learning_rate": 2e-05,
"loss": 27.031,
"loss/crossentropy": 2.5481350421905518,
"loss/hidden": 0.1416015625,
"loss/logits": 0.03450712189078331,
"loss/reg": 26.85489273071289,
"step": 412
},
{
"epoch": 0.010325,
"grad_norm": 2.942481756210327,
"grad_norm_var": 0.21596104298213442,
"learning_rate": 2e-05,
"loss": 26.9305,
"loss/crossentropy": 3.015340566635132,
"loss/hidden": 0.1630859375,
"loss/logits": 0.038723111152648926,
"loss/reg": 26.728679656982422,
"step": 413
},
{
"epoch": 0.01035,
"grad_norm": 6.163323879241943,
"grad_norm_var": 1.1063828714438992,
"learning_rate": 2e-05,
"loss": 26.8133,
"loss/crossentropy": 2.55928373336792,
"loss/hidden": 0.1728515625,
"loss/logits": 0.03731565177440643,
"loss/reg": 26.603172302246094,
"step": 414
},
{
"epoch": 0.010375,
"grad_norm": 1.9960532188415527,
"grad_norm_var": 1.1163400619934953,
"learning_rate": 2e-05,
"loss": 26.6719,
"loss/crossentropy": 2.80226731300354,
"loss/hidden": 0.15625,
"loss/logits": 0.03723851963877678,
"loss/reg": 26.47838592529297,
"step": 415
},
{
"epoch": 0.0104,
"grad_norm": 3.301450729370117,
"grad_norm_var": 1.0974053192222506,
"learning_rate": 2e-05,
"loss": 26.5614,
"loss/crossentropy": 2.791954517364502,
"loss/hidden": 0.166015625,
"loss/logits": 0.04129061847925186,
"loss/reg": 26.354053497314453,
"step": 416
},
{
"epoch": 0.010425,
"grad_norm": 1.9505646228790283,
"grad_norm_var": 1.0653418372806414,
"learning_rate": 2e-05,
"loss": 26.4048,
"loss/crossentropy": 2.7842905521392822,
"loss/hidden": 0.13671875,
"loss/logits": 0.0375644825398922,
"loss/reg": 26.230472564697266,
"step": 417
},
{
"epoch": 0.01045,
"grad_norm": 2.055081605911255,
"grad_norm_var": 1.075675176346497,
"learning_rate": 2e-05,
"loss": 26.296,
"loss/crossentropy": 2.7261054515838623,
"loss/hidden": 0.1513671875,
"loss/logits": 0.03734801337122917,
"loss/reg": 26.1073055267334,
"step": 418
},
{
"epoch": 0.010475,
"grad_norm": 3.4958529472351074,
"grad_norm_var": 1.09835111004395,
"learning_rate": 2e-05,
"loss": 26.1998,
"loss/crossentropy": 3.256852626800537,
"loss/hidden": 0.1689453125,
"loss/logits": 0.04601683095097542,
"loss/reg": 25.984798431396484,
"step": 419
},
{
"epoch": 0.0105,
"grad_norm": 2.4832944869995117,
"grad_norm_var": 1.097555539592861,
"learning_rate": 2e-05,
"loss": 26.0394,
"loss/crossentropy": 2.6658496856689453,
"loss/hidden": 0.13671875,
"loss/logits": 0.03990530967712402,
"loss/reg": 25.862812042236328,
"step": 420
},
{
"epoch": 0.010525,
"grad_norm": 1.888179898262024,
"grad_norm_var": 1.143925812704361,
"learning_rate": 2e-05,
"loss": 25.9507,
"loss/crossentropy": 2.893690347671509,
"loss/hidden": 0.166015625,
"loss/logits": 0.04318728297948837,
"loss/reg": 25.74148941040039,
"step": 421
},
{
"epoch": 0.01055,
"grad_norm": 1.8328545093536377,
"grad_norm_var": 1.1896980975503562,
"learning_rate": 2e-05,
"loss": 25.8021,
"loss/crossentropy": 2.862842559814453,
"loss/hidden": 0.1484375,
"loss/logits": 0.037468597292900085,
"loss/reg": 25.616188049316406,
"step": 422
},
{
"epoch": 0.010575,
"grad_norm": 2.6049251556396484,
"grad_norm_var": 1.1751238780603264,
"learning_rate": 2e-05,
"loss": 25.7268,
"loss/crossentropy": 2.9560859203338623,
"loss/hidden": 0.1767578125,
"loss/logits": 0.054243914783000946,
"loss/reg": 25.49581527709961,
"step": 423
},
{
"epoch": 0.0106,
"grad_norm": 2.8561511039733887,
"grad_norm_var": 1.1601718819937763,
"learning_rate": 2e-05,
"loss": 25.5684,
"loss/crossentropy": 2.6629557609558105,
"loss/hidden": 0.1533203125,
"loss/logits": 0.039182234555482864,
"loss/reg": 25.375917434692383,
"step": 424
},
{
"epoch": 0.010625,
"grad_norm": 1.8477426767349243,
"grad_norm_var": 1.1963363532274474,
"learning_rate": 2e-05,
"loss": 25.4692,
"loss/crossentropy": 3.0463385581970215,
"loss/hidden": 0.166015625,
"loss/logits": 0.046355295926332474,
"loss/reg": 25.256792068481445,
"step": 425
},
{
"epoch": 0.01065,
"grad_norm": 2.3839778900146484,
"grad_norm_var": 1.1845253457707392,
"learning_rate": 2e-05,
"loss": 25.3301,
"loss/crossentropy": 2.812811851501465,
"loss/hidden": 0.1533203125,
"loss/logits": 0.03877168148756027,
"loss/reg": 25.13796615600586,
"step": 426
},
{
"epoch": 0.010675,
"grad_norm": 3.649106979370117,
"grad_norm_var": 1.2399896204357086,
"learning_rate": 2e-05,
"loss": 25.2656,
"loss/crossentropy": 2.817918300628662,
"loss/hidden": 0.201171875,
"loss/logits": 0.04450376331806183,
"loss/reg": 25.01987648010254,
"step": 427
},
{
"epoch": 0.0107,
"grad_norm": 2.212465524673462,
"grad_norm_var": 1.197620310146587,
"learning_rate": 2e-05,
"loss": 25.1043,
"loss/crossentropy": 3.138951301574707,
"loss/hidden": 0.1611328125,
"loss/logits": 0.04081891477108002,
"loss/reg": 24.902353286743164,
"step": 428
},
{
"epoch": 0.010725,
"grad_norm": 2.2249510288238525,
"grad_norm_var": 1.2093715461039343,
"learning_rate": 2e-05,
"loss": 25.003,
"loss/crossentropy": 2.6012675762176514,
"loss/hidden": 0.177734375,
"loss/logits": 0.04012144356966019,
"loss/reg": 24.785186767578125,
"step": 429
},
{
"epoch": 0.01075,
"grad_norm": 2.0124897956848145,
"grad_norm_var": 0.36066606030676096,
"learning_rate": 2e-05,
"loss": 24.862,
"loss/crossentropy": 2.6634442806243896,
"loss/hidden": 0.15625,
"loss/logits": 0.03708156198263168,
"loss/reg": 24.668659210205078,
"step": 430
},
{
"epoch": 0.010775,
"grad_norm": 5.937109470367432,
"grad_norm_var": 1.1061704228386418,
"learning_rate": 2e-05,
"loss": 24.7883,
"loss/crossentropy": 2.98293137550354,
"loss/hidden": 0.1904296875,
"loss/logits": 0.04528553783893585,
"loss/reg": 24.552602767944336,
"step": 431
},
{
"epoch": 0.0108,
"grad_norm": 2.0511343479156494,
"grad_norm_var": 1.098776464532748,
"learning_rate": 2e-05,
"loss": 24.6483,
"loss/crossentropy": 3.482454299926758,
"loss/hidden": 0.173828125,
"loss/logits": 0.04099034145474434,
"loss/reg": 24.43346405029297,
"step": 432
},
{
"epoch": 0.010825,
"grad_norm": 2.0217044353485107,
"grad_norm_var": 1.0930003270482167,
"learning_rate": 2e-05,
"loss": 24.4952,
"loss/crossentropy": 2.766650915145874,
"loss/hidden": 0.134765625,
"loss/logits": 0.041929375380277634,
"loss/reg": 24.31853485107422,
"step": 433
},
{
"epoch": 0.01085,
"grad_norm": 2.1080517768859863,
"grad_norm_var": 1.0893460739164103,
"learning_rate": 2e-05,
"loss": 24.3975,
"loss/crossentropy": 2.829714059829712,
"loss/hidden": 0.15625,
"loss/logits": 0.03715629130601883,
"loss/reg": 24.20409393310547,
"step": 434
},
{
"epoch": 0.010875,
"grad_norm": 2.590503215789795,
"grad_norm_var": 1.0325087297316353,
"learning_rate": 2e-05,
"loss": 24.2874,
"loss/crossentropy": 2.812587261199951,
"loss/hidden": 0.158203125,
"loss/logits": 0.03894827514886856,
"loss/reg": 24.090213775634766,
"step": 435
},
{
"epoch": 0.0109,
"grad_norm": 2.045022964477539,
"grad_norm_var": 1.0480635957117082,
"learning_rate": 2e-05,
"loss": 24.1827,
"loss/crossentropy": 3.086454391479492,
"loss/hidden": 0.166015625,
"loss/logits": 0.040063485503196716,
"loss/reg": 23.976669311523438,
"step": 436
},
{
"epoch": 0.010925,
"grad_norm": 2.0887551307678223,
"grad_norm_var": 1.033770641152501,
"learning_rate": 2e-05,
"loss": 24.0595,
"loss/crossentropy": 2.839022636413574,
"loss/hidden": 0.1533203125,
"loss/logits": 0.042500950396060944,
"loss/reg": 23.863662719726562,
"step": 437
},
{
"epoch": 0.01095,
"grad_norm": 2.314814567565918,
"grad_norm_var": 1.0035414097905393,
"learning_rate": 2e-05,
"loss": 23.9523,
"loss/crossentropy": 2.9472670555114746,
"loss/hidden": 0.15625,
"loss/logits": 0.044826384633779526,
"loss/reg": 23.751253128051758,
"step": 438
},
{
"epoch": 0.010975,
"grad_norm": 3.2724692821502686,
"grad_norm_var": 1.0354526746118948,
"learning_rate": 2e-05,
"loss": 23.853,
"loss/crossentropy": 2.7853267192840576,
"loss/hidden": 0.173828125,
"loss/logits": 0.03991552069783211,
"loss/reg": 23.63922882080078,
"step": 439
},
{
"epoch": 0.011,
"grad_norm": 1.8264756202697754,
"grad_norm_var": 1.0666913011185395,
"learning_rate": 2e-05,
"loss": 23.7176,
"loss/crossentropy": 2.5802700519561768,
"loss/hidden": 0.1513671875,
"loss/logits": 0.03871052712202072,
"loss/reg": 23.527509689331055,
"step": 440
},
{
"epoch": 0.011025,
"grad_norm": 2.1698477268218994,
"grad_norm_var": 1.043588037234136,
"learning_rate": 2e-05,
"loss": 23.6232,
"loss/crossentropy": 2.9302892684936523,
"loss/hidden": 0.1630859375,
"loss/logits": 0.043618083000183105,
"loss/reg": 23.41650390625,
"step": 441
},
{
"epoch": 0.01105,
"grad_norm": 2.231228828430176,
"grad_norm_var": 1.0485661946787095,
"learning_rate": 2e-05,
"loss": 23.4972,
"loss/crossentropy": 2.867483615875244,
"loss/hidden": 0.1513671875,
"loss/logits": 0.04387916252017021,
"loss/reg": 23.30193328857422,
"step": 442
},
{
"epoch": 0.011075,
"grad_norm": 1.5965532064437866,
"grad_norm_var": 1.0103300653968241,
"learning_rate": 2e-05,
"loss": 23.3734,
"loss/crossentropy": 2.6820480823516846,
"loss/hidden": 0.1455078125,
"loss/logits": 0.03568604588508606,
"loss/reg": 23.192241668701172,
"step": 443
},
{
"epoch": 0.0111,
"grad_norm": 1.9236520528793335,
"grad_norm_var": 1.0234956986676196,
"learning_rate": 2e-05,
"loss": 23.299,
"loss/crossentropy": 2.791132688522339,
"loss/hidden": 0.1728515625,
"loss/logits": 0.04304938763380051,
"loss/reg": 23.083120346069336,
"step": 444
},
{
"epoch": 0.011125,
"grad_norm": 2.615415096282959,
"grad_norm_var": 1.0238631693360027,
"learning_rate": 2e-05,
"loss": 23.2208,
"loss/crossentropy": 2.7200093269348145,
"loss/hidden": 0.20703125,
"loss/logits": 0.039411187171936035,
"loss/reg": 22.974393844604492,
"step": 445
},
{
"epoch": 0.01115,
"grad_norm": 3.7206993103027344,
"grad_norm_var": 1.11220864186432,
"learning_rate": 2e-05,
"loss": 23.1272,
"loss/crossentropy": 2.934048652648926,
"loss/hidden": 0.2109375,
"loss/logits": 0.050051335245370865,
"loss/reg": 22.866172790527344,
"step": 446
},
{
"epoch": 0.011175,
"grad_norm": 2.2452712059020996,
"grad_norm_var": 0.2879587549997645,
"learning_rate": 2e-05,
"loss": 22.9826,
"loss/crossentropy": 3.0915427207946777,
"loss/hidden": 0.177734375,
"loss/logits": 0.04651949554681778,
"loss/reg": 22.75834846496582,
"step": 447
},
{
"epoch": 0.0112,
"grad_norm": 1.6305818557739258,
"grad_norm_var": 0.3130432844059186,
"learning_rate": 2e-05,
"loss": 22.8382,
"loss/crossentropy": 2.8177685737609863,
"loss/hidden": 0.1484375,
"loss/logits": 0.038578279316425323,
"loss/reg": 22.65113639831543,
"step": 448
},
{
"epoch": 0.011225,
"grad_norm": 2.833962917327881,
"grad_norm_var": 0.3268392463359208,
"learning_rate": 2e-05,
"loss": 22.7767,
"loss/crossentropy": 2.2847812175750732,
"loss/hidden": 0.189453125,
"loss/logits": 0.04296421259641647,
"loss/reg": 22.544269561767578,
"step": 449
},
{
"epoch": 0.01125,
"grad_norm": 1.8361362218856812,
"grad_norm_var": 0.33935606993982687,
"learning_rate": 2e-05,
"loss": 22.621,
"loss/crossentropy": 3.0569963455200195,
"loss/hidden": 0.138671875,
"loss/logits": 0.044218260794878006,
"loss/reg": 22.43810272216797,
"step": 450
},
{
"epoch": 0.011275,
"grad_norm": 2.2452023029327393,
"grad_norm_var": 0.3338401600601277,
"learning_rate": 2e-05,
"loss": 22.5931,
"loss/crossentropy": 2.7638001441955566,
"loss/hidden": 0.21484375,
"loss/logits": 0.046010829508304596,
"loss/reg": 22.332237243652344,
"step": 451
},
{
"epoch": 0.0113,
"grad_norm": 2.50168776512146,
"grad_norm_var": 0.3321248705423902,
"learning_rate": 2e-05,
"loss": 22.4375,
"loss/crossentropy": 2.8486430644989014,
"loss/hidden": 0.169921875,
"loss/logits": 0.04485378414392471,
"loss/reg": 22.22277069091797,
"step": 452
},
{
"epoch": 0.011325,
"grad_norm": 2.492642402648926,
"grad_norm_var": 0.3300935987394846,
"learning_rate": 2e-05,
"loss": 22.3492,
"loss/crossentropy": 2.71463942527771,
"loss/hidden": 0.1796875,
"loss/logits": 0.05168076232075691,
"loss/reg": 22.117813110351562,
"step": 453
},
{
"epoch": 0.01135,
"grad_norm": 1.9064103364944458,
"grad_norm_var": 0.341946302980373,
"learning_rate": 2e-05,
"loss": 22.2137,
"loss/crossentropy": 2.8734917640686035,
"loss/hidden": 0.158203125,
"loss/logits": 0.042219504714012146,
"loss/reg": 22.013267517089844,
"step": 454
},
{
"epoch": 0.011375,
"grad_norm": 2.5244898796081543,
"grad_norm_var": 0.2814757407362549,
"learning_rate": 2e-05,
"loss": 22.1323,
"loss/crossentropy": 2.5757384300231934,
"loss/hidden": 0.1796875,
"loss/logits": 0.043334104120731354,
"loss/reg": 21.90932273864746,
"step": 455
},
{
"epoch": 0.0114,
"grad_norm": 1.49411940574646,
"grad_norm_var": 0.3079792610992302,
"learning_rate": 2e-05,
"loss": 21.9885,
"loss/crossentropy": 2.7647297382354736,
"loss/hidden": 0.1416015625,
"loss/logits": 0.04113258421421051,
"loss/reg": 21.80577278137207,
"step": 456
},
{
"epoch": 0.011425,
"grad_norm": 2.4946882724761963,
"grad_norm_var": 0.31118967972119144,
"learning_rate": 2e-05,
"loss": 21.9178,
"loss/crossentropy": 2.653733968734741,
"loss/hidden": 0.169921875,
"loss/logits": 0.04473033547401428,
"loss/reg": 21.7031192779541,
"step": 457
},
{
"epoch": 0.01145,
"grad_norm": 1.9149430990219116,
"grad_norm_var": 0.31900516012573277,
"learning_rate": 2e-05,
"loss": 21.825,
"loss/crossentropy": 2.6428985595703125,
"loss/hidden": 0.1826171875,
"loss/logits": 0.04166660085320473,
"loss/reg": 21.600685119628906,
"step": 458
},
{
"epoch": 0.011475,
"grad_norm": 1.7886494398117065,
"grad_norm_var": 0.3046125382910664,
"learning_rate": 2e-05,
"loss": 21.7068,
"loss/crossentropy": 2.5072784423828125,
"loss/hidden": 0.169921875,
"loss/logits": 0.03796003758907318,
"loss/reg": 21.498947143554688,
"step": 459
},
{
"epoch": 0.0115,
"grad_norm": 2.5755653381347656,
"grad_norm_var": 0.30189205483048126,
"learning_rate": 2e-05,
"loss": 21.6383,
"loss/crossentropy": 2.8383917808532715,
"loss/hidden": 0.19140625,
"loss/logits": 0.0494539812207222,
"loss/reg": 21.397401809692383,
"step": 460
},
{
"epoch": 0.011525,
"grad_norm": 3.0548369884490967,
"grad_norm_var": 0.33236538038720104,
"learning_rate": 2e-05,
"loss": 21.4977,
"loss/crossentropy": 2.717653751373291,
"loss/hidden": 0.1630859375,
"loss/logits": 0.03829260170459747,
"loss/reg": 21.29631805419922,
"step": 461
},
{
"epoch": 0.01155,
"grad_norm": 2.220721483230591,
"grad_norm_var": 0.19459906037036465,
"learning_rate": 2e-05,
"loss": 21.4131,
"loss/crossentropy": 2.909445285797119,
"loss/hidden": 0.177734375,
"loss/logits": 0.04297621548175812,
"loss/reg": 21.192386627197266,
"step": 462
},
{
"epoch": 0.011575,
"grad_norm": 2.5778627395629883,
"grad_norm_var": 0.2019683654886343,
"learning_rate": 2e-05,
"loss": 21.3024,
"loss/crossentropy": 2.7236719131469727,
"loss/hidden": 0.1650390625,
"loss/logits": 0.044891439378261566,
"loss/reg": 21.092477798461914,
"step": 463
},
{
"epoch": 0.0116,
"grad_norm": 2.302924156188965,
"grad_norm_var": 0.17417472367764025,
"learning_rate": 2e-05,
"loss": 21.2079,
"loss/crossentropy": 2.661893129348755,
"loss/hidden": 0.1728515625,
"loss/logits": 0.042085111141204834,
"loss/reg": 20.99297523498535,
"step": 464
},
{
"epoch": 0.011625,
"grad_norm": 3.1231939792633057,
"grad_norm_var": 0.2000796962515011,
"learning_rate": 2e-05,
"loss": 21.158,
"loss/crossentropy": 2.9192662239074707,
"loss/hidden": 0.1923828125,
"loss/logits": 0.07160253822803497,
"loss/reg": 20.89402961730957,
"step": 465
},
{
"epoch": 0.01165,
"grad_norm": 2.0219297409057617,
"grad_norm_var": 0.19035271984780966,
"learning_rate": 2e-05,
"loss": 21.0286,
"loss/crossentropy": 3.011322021484375,
"loss/hidden": 0.1826171875,
"loss/logits": 0.05049794167280197,
"loss/reg": 20.7955265045166,
"step": 466
},
{
"epoch": 0.011675,
"grad_norm": 2.3812074661254883,
"grad_norm_var": 0.1900165697599144,
"learning_rate": 2e-05,
"loss": 20.9126,
"loss/crossentropy": 2.594330072402954,
"loss/hidden": 0.1728515625,
"loss/logits": 0.04238292574882507,
"loss/reg": 20.69733428955078,
"step": 467
},
{
"epoch": 0.0117,
"grad_norm": 3.212700128555298,
"grad_norm_var": 0.23732095982581564,
"learning_rate": 2e-05,
"loss": 20.8329,
"loss/crossentropy": 3.0930912494659424,
"loss/hidden": 0.1796875,
"loss/logits": 0.05329480022192001,
"loss/reg": 20.599882125854492,
"step": 468
},
{
"epoch": 0.011725,
"grad_norm": 2.802736759185791,
"grad_norm_var": 0.24797037852165235,
"learning_rate": 2e-05,
"loss": 20.7412,
"loss/crossentropy": 2.682535409927368,
"loss/hidden": 0.1865234375,
"loss/logits": 0.05204048752784729,
"loss/reg": 20.502676010131836,
"step": 469
},
{
"epoch": 0.01175,
"grad_norm": 6.470301151275635,
"grad_norm_var": 1.249545399419666,
"learning_rate": 2e-05,
"loss": 20.6501,
"loss/crossentropy": 3.1839163303375244,
"loss/hidden": 0.1923828125,
"loss/logits": 0.052045367658138275,
"loss/reg": 20.405675888061523,
"step": 470
},
{
"epoch": 0.011775,
"grad_norm": 2.094099283218384,
"grad_norm_var": 1.2703367134671737,
"learning_rate": 2e-05,
"loss": 20.5315,
"loss/crossentropy": 2.884645462036133,
"loss/hidden": 0.1728515625,
"loss/logits": 0.0492381826043129,
"loss/reg": 20.309377670288086,
"step": 471
},
{
"epoch": 0.0118,
"grad_norm": 2.2063822746276855,
"grad_norm_var": 1.1914975389137128,
"learning_rate": 2e-05,
"loss": 20.4294,
"loss/crossentropy": 3.0197482109069824,
"loss/hidden": 0.1728515625,
"loss/logits": 0.0465642511844635,
"loss/reg": 20.20997428894043,
"step": 472
},
{
"epoch": 0.011825,
"grad_norm": 2.1600499153137207,
"grad_norm_var": 1.2077763497968268,
"learning_rate": 2e-05,
"loss": 20.3465,
"loss/crossentropy": 2.7459301948547363,
"loss/hidden": 0.1845703125,
"loss/logits": 0.04744536057114601,
"loss/reg": 20.114482879638672,
"step": 473
},
{
"epoch": 0.01185,
"grad_norm": 2.352337121963501,
"grad_norm_var": 1.1750134992717125,
"learning_rate": 2e-05,
"loss": 20.2671,
"loss/crossentropy": 3.246368646621704,
"loss/hidden": 0.19921875,
"loss/logits": 0.048278048634529114,
"loss/reg": 20.019636154174805,
"step": 474
},
{
"epoch": 0.011875,
"grad_norm": 1.7839800119400024,
"grad_norm_var": 1.1755879216903617,
"learning_rate": 2e-05,
"loss": 20.1489,
"loss/crossentropy": 2.9967947006225586,
"loss/hidden": 0.1748046875,
"loss/logits": 0.048982493579387665,
"loss/reg": 19.92508888244629,
"step": 475
},
{
"epoch": 0.0119,
"grad_norm": 2.4864652156829834,
"grad_norm_var": 1.1776669498268881,
"learning_rate": 2e-05,
"loss": 20.0598,
"loss/crossentropy": 2.60671329498291,
"loss/hidden": 0.1767578125,
"loss/logits": 0.052169423550367355,
"loss/reg": 19.830875396728516,
"step": 476
},
{
"epoch": 0.011925,
"grad_norm": 2.363190174102783,
"grad_norm_var": 1.175140638989229,
"learning_rate": 2e-05,
"loss": 20.0069,
"loss/crossentropy": 2.858886480331421,
"loss/hidden": 0.220703125,
"loss/logits": 0.04880058765411377,
"loss/reg": 19.737428665161133,
"step": 477
},
{
"epoch": 0.01195,
"grad_norm": 2.021653652191162,
"grad_norm_var": 1.189277020649322,
"learning_rate": 2e-05,
"loss": 19.8696,
"loss/crossentropy": 2.7821409702301025,
"loss/hidden": 0.1796875,
"loss/logits": 0.04575974866747856,
"loss/reg": 19.64413833618164,
"step": 478
},
{
"epoch": 0.011975,
"grad_norm": 1.9459588527679443,
"grad_norm_var": 1.2201059740655162,
"learning_rate": 2e-05,
"loss": 19.763,
"loss/crossentropy": 2.772062301635742,
"loss/hidden": 0.169921875,
"loss/logits": 0.041855379939079285,
"loss/reg": 19.551227569580078,
"step": 479
},
{
"epoch": 0.012,
"grad_norm": 2.5396687984466553,
"grad_norm_var": 1.2139767764762164,
"learning_rate": 2e-05,
"loss": 19.705,
"loss/crossentropy": 2.853855848312378,
"loss/hidden": 0.197265625,
"loss/logits": 0.0488080270588398,
"loss/reg": 19.45893096923828,
"step": 480
},
{
"epoch": 0.012025,
"grad_norm": 2.137702226638794,
"grad_norm_var": 1.208933842733709,
"learning_rate": 2e-05,
"loss": 19.6405,
"loss/crossentropy": 2.817518949508667,
"loss/hidden": 0.224609375,
"loss/logits": 0.04868399724364281,
"loss/reg": 19.367197036743164,
"step": 481
},
{
"epoch": 0.01205,
"grad_norm": 2.0161988735198975,
"grad_norm_var": 1.2093480157874186,
"learning_rate": 2e-05,
"loss": 19.4942,
"loss/crossentropy": 3.041217565536499,
"loss/hidden": 0.1728515625,
"loss/logits": 0.049482930451631546,
"loss/reg": 19.271909713745117,
"step": 482
},
{
"epoch": 0.012075,
"grad_norm": 2.4689245223999023,
"grad_norm_var": 1.2077271255452753,
"learning_rate": 2e-05,
"loss": 19.4371,
"loss/crossentropy": 2.8089406490325928,
"loss/hidden": 0.205078125,
"loss/logits": 0.05119004100561142,
"loss/reg": 19.180824279785156,
"step": 483
},
{
"epoch": 0.0121,
"grad_norm": 2.2316370010375977,
"grad_norm_var": 1.1833405153175727,
"learning_rate": 2e-05,
"loss": 19.3215,
"loss/crossentropy": 2.9401211738586426,
"loss/hidden": 0.1875,
"loss/logits": 0.04366878420114517,
"loss/reg": 19.090309143066406,
"step": 484
},
{
"epoch": 0.012125,
"grad_norm": 2.047006607055664,
"grad_norm_var": 1.1890429642677636,
"learning_rate": 2e-05,
"loss": 19.2222,
"loss/crossentropy": 2.878319501876831,
"loss/hidden": 0.1748046875,
"loss/logits": 0.04722724109888077,
"loss/reg": 19.000150680541992,
"step": 485
},
{
"epoch": 0.01215,
"grad_norm": 1.9564520120620728,
"grad_norm_var": 0.047588535415659786,
"learning_rate": 2e-05,
"loss": 19.1429,
"loss/crossentropy": 2.7814621925354004,
"loss/hidden": 0.1845703125,
"loss/logits": 0.04784928262233734,
"loss/reg": 18.91046142578125,
"step": 486
},
{
"epoch": 0.012175,
"grad_norm": 2.660238265991211,
"grad_norm_var": 0.06145858363412131,
"learning_rate": 2e-05,
"loss": 19.042,
"loss/crossentropy": 2.7573916912078857,
"loss/hidden": 0.171875,
"loss/logits": 0.04883112385869026,
"loss/reg": 18.821319580078125,
"step": 487
},
{
"epoch": 0.0122,
"grad_norm": 3.515303611755371,
"grad_norm_var": 0.16771224633902548,
"learning_rate": 2e-05,
"loss": 18.9393,
"loss/crossentropy": 2.9051125049591064,
"loss/hidden": 0.162109375,
"loss/logits": 0.04467906057834625,
"loss/reg": 18.73251724243164,
"step": 488
},
{
"epoch": 0.012225,
"grad_norm": 2.64762806892395,
"grad_norm_var": 0.17393239434562133,
"learning_rate": 2e-05,
"loss": 18.9105,
"loss/crossentropy": 2.8703958988189697,
"loss/hidden": 0.21875,
"loss/logits": 0.04748106747865677,
"loss/reg": 18.644243240356445,
"step": 489
},
{
"epoch": 0.01225,
"grad_norm": 2.6073079109191895,
"grad_norm_var": 0.17897939206579636,
"learning_rate": 2e-05,
"loss": 18.8279,
"loss/crossentropy": 3.221273422241211,
"loss/hidden": 0.2158203125,
"loss/logits": 0.05585157871246338,
"loss/reg": 18.556241989135742,
"step": 490
},
{
"epoch": 0.012275,
"grad_norm": 1.5472153425216675,
"grad_norm_var": 0.20001469118204512,
"learning_rate": 2e-05,
"loss": 18.679,
"loss/crossentropy": 2.915916681289673,
"loss/hidden": 0.1669921875,
"loss/logits": 0.043436288833618164,
"loss/reg": 18.468524932861328,
"step": 491
},
{
"epoch": 0.0123,
"grad_norm": 1.7620818614959717,
"grad_norm_var": 0.21717040004255922,
"learning_rate": 2e-05,
"loss": 18.6017,
"loss/crossentropy": 2.867325782775879,
"loss/hidden": 0.1728515625,
"loss/logits": 0.051028743386268616,
"loss/reg": 18.377832412719727,
"step": 492
},
{
"epoch": 0.012325,
"grad_norm": 2.20041561126709,
"grad_norm_var": 0.21700482333929852,
"learning_rate": 2e-05,
"loss": 18.5379,
"loss/crossentropy": 2.9628453254699707,
"loss/hidden": 0.19921875,
"loss/logits": 0.04741261899471283,
"loss/reg": 18.291229248046875,
"step": 493
},
{
"epoch": 0.01235,
"grad_norm": 1.9837777614593506,
"grad_norm_var": 0.2183440529603916,
"learning_rate": 2e-05,
"loss": 18.4331,
"loss/crossentropy": 2.825575113296509,
"loss/hidden": 0.1796875,
"loss/logits": 0.048515524715185165,
"loss/reg": 18.204925537109375,
"step": 494
},
{
"epoch": 0.012375,
"grad_norm": 2.4184892177581787,
"grad_norm_var": 0.21209012166573088,
"learning_rate": 2e-05,
"loss": 18.4085,
"loss/crossentropy": 2.6464486122131348,
"loss/hidden": 0.2451171875,
"loss/logits": 0.04423639923334122,
"loss/reg": 18.119152069091797,
"step": 495
},
{
"epoch": 0.0124,
"grad_norm": 1.9086065292358398,
"grad_norm_var": 0.21649869079500092,
"learning_rate": 2e-05,
"loss": 18.2784,
"loss/crossentropy": 3.1577796936035156,
"loss/hidden": 0.1953125,
"loss/logits": 0.04937343671917915,
"loss/reg": 18.033668518066406,
"step": 496
},
{
"epoch": 0.012425,
"grad_norm": 1.8744006156921387,
"grad_norm_var": 0.22501323270567444,
"learning_rate": 2e-05,
"loss": 18.1669,
"loss/crossentropy": 2.6969730854034424,
"loss/hidden": 0.1748046875,
"loss/logits": 0.0435866042971611,
"loss/reg": 17.94854736328125,
"step": 497
},
{
"epoch": 0.01245,
"grad_norm": 3.1248912811279297,
"grad_norm_var": 0.26870209982110016,
"learning_rate": 2e-05,
"loss": 18.1115,
"loss/crossentropy": 2.8137989044189453,
"loss/hidden": 0.20703125,
"loss/logits": 0.04066001996397972,
"loss/reg": 17.863773345947266,
"step": 498
},
{
"epoch": 0.012475,
"grad_norm": 3.302873373031616,
"grad_norm_var": 0.32987942357099176,
"learning_rate": 2e-05,
"loss": 18.0577,
"loss/crossentropy": 3.0537073612213135,
"loss/hidden": 0.2197265625,
"loss/logits": 0.05833124369382858,
"loss/reg": 17.779624938964844,
"step": 499
},
{
"epoch": 0.0125,
"grad_norm": 2.7058842182159424,
"grad_norm_var": 0.335707609950227,
"learning_rate": 2e-05,
"loss": 17.982,
"loss/crossentropy": 3.436614751815796,
"loss/hidden": 0.23046875,
"loss/logits": 0.055682770907878876,
"loss/reg": 17.69584846496582,
"step": 500
}
],
"logging_steps": 1,
"max_steps": 40000,
"num_input_tokens_seen": 0,
"num_train_epochs": 9223372036854775807,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": true,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 3.227844083712e+16,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}