| { | |
| "best_global_step": 1250, | |
| "best_metric": 0.7589222033223469, | |
| "best_model_checkpoint": "voice_emotion_classification/checkpoint-1250", | |
| "epoch": 1.0, | |
| "eval_steps": 500, | |
| "global_step": 1250, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0008, | |
| "grad_norm": 1.4005043506622314, | |
| "learning_rate": 6.000000000000001e-07, | |
| "loss": 2.0903, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.008, | |
| "grad_norm": 1.3180536031723022, | |
| "learning_rate": 6e-06, | |
| "loss": 2.0826, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.016, | |
| "grad_norm": 1.0437175035476685, | |
| "learning_rate": 1.2e-05, | |
| "loss": 2.0693, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.024, | |
| "grad_norm": 1.371071457862854, | |
| "learning_rate": 1.8e-05, | |
| "loss": 2.0463, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.032, | |
| "grad_norm": 1.51685631275177, | |
| "learning_rate": 2.4e-05, | |
| "loss": 2.0423, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 1.4282890558242798, | |
| "learning_rate": 3e-05, | |
| "loss": 1.996, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.048, | |
| "grad_norm": 1.860026478767395, | |
| "learning_rate": 2.991891891891892e-05, | |
| "loss": 1.9639, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.056, | |
| "grad_norm": 2.121481418609619, | |
| "learning_rate": 2.983783783783784e-05, | |
| "loss": 1.9517, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.064, | |
| "grad_norm": 2.056445837020874, | |
| "learning_rate": 2.9756756756756758e-05, | |
| "loss": 1.8765, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.072, | |
| "grad_norm": 2.3928184509277344, | |
| "learning_rate": 2.9675675675675678e-05, | |
| "loss": 1.8703, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 3.0733420848846436, | |
| "learning_rate": 2.9594594594594598e-05, | |
| "loss": 1.853, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.088, | |
| "grad_norm": 2.88864803314209, | |
| "learning_rate": 2.9513513513513514e-05, | |
| "loss": 1.7652, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.096, | |
| "grad_norm": 2.9048268795013428, | |
| "learning_rate": 2.943243243243243e-05, | |
| "loss": 1.7984, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.104, | |
| "grad_norm": 2.6991426944732666, | |
| "learning_rate": 2.935135135135135e-05, | |
| "loss": 1.6869, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.112, | |
| "grad_norm": 2.4570231437683105, | |
| "learning_rate": 2.927027027027027e-05, | |
| "loss": 1.7812, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 4.502678871154785, | |
| "learning_rate": 2.918918918918919e-05, | |
| "loss": 1.8444, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.128, | |
| "grad_norm": 3.117838144302368, | |
| "learning_rate": 2.9108108108108108e-05, | |
| "loss": 1.7544, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.136, | |
| "grad_norm": 2.812086343765259, | |
| "learning_rate": 2.9027027027027028e-05, | |
| "loss": 1.638, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.144, | |
| "grad_norm": 3.23271107673645, | |
| "learning_rate": 2.8954054054054057e-05, | |
| "loss": 1.7026, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.152, | |
| "grad_norm": 3.975172996520996, | |
| "learning_rate": 2.8872972972972977e-05, | |
| "loss": 1.6715, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 3.8965938091278076, | |
| "learning_rate": 2.879189189189189e-05, | |
| "loss": 1.6804, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.168, | |
| "grad_norm": 4.267274856567383, | |
| "learning_rate": 2.871081081081081e-05, | |
| "loss": 1.6587, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.176, | |
| "grad_norm": 3.524360179901123, | |
| "learning_rate": 2.862972972972973e-05, | |
| "loss": 1.4811, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.184, | |
| "grad_norm": 3.266697883605957, | |
| "learning_rate": 2.854864864864865e-05, | |
| "loss": 1.672, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.192, | |
| "grad_norm": 5.3684186935424805, | |
| "learning_rate": 2.8467567567567567e-05, | |
| "loss": 1.5284, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 3.898176431655884, | |
| "learning_rate": 2.8386486486486487e-05, | |
| "loss": 1.5774, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.208, | |
| "grad_norm": 3.189732074737549, | |
| "learning_rate": 2.8305405405405407e-05, | |
| "loss": 1.4874, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.216, | |
| "grad_norm": 3.274244785308838, | |
| "learning_rate": 2.8224324324324327e-05, | |
| "loss": 1.5098, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.224, | |
| "grad_norm": 5.691224098205566, | |
| "learning_rate": 2.8143243243243244e-05, | |
| "loss": 1.509, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.232, | |
| "grad_norm": 6.856773376464844, | |
| "learning_rate": 2.8062162162162164e-05, | |
| "loss": 1.4558, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 7.078716278076172, | |
| "learning_rate": 2.7981081081081084e-05, | |
| "loss": 1.5298, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.248, | |
| "grad_norm": 4.4305100440979, | |
| "learning_rate": 2.79e-05, | |
| "loss": 1.3387, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.256, | |
| "grad_norm": 10.400449752807617, | |
| "learning_rate": 2.7818918918918917e-05, | |
| "loss": 1.4501, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.264, | |
| "grad_norm": 5.316948890686035, | |
| "learning_rate": 2.7737837837837837e-05, | |
| "loss": 1.367, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.272, | |
| "grad_norm": 9.753177642822266, | |
| "learning_rate": 2.7656756756756757e-05, | |
| "loss": 1.4684, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 8.100529670715332, | |
| "learning_rate": 2.7575675675675677e-05, | |
| "loss": 1.4175, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.288, | |
| "grad_norm": 9.878854751586914, | |
| "learning_rate": 2.7494594594594594e-05, | |
| "loss": 1.308, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.296, | |
| "grad_norm": 5.865877151489258, | |
| "learning_rate": 2.7413513513513514e-05, | |
| "loss": 1.3035, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.304, | |
| "grad_norm": 7.870754241943359, | |
| "learning_rate": 2.7332432432432434e-05, | |
| "loss": 1.2915, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.312, | |
| "grad_norm": 8.517908096313477, | |
| "learning_rate": 2.7251351351351354e-05, | |
| "loss": 1.4318, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 4.7960309982299805, | |
| "learning_rate": 2.717027027027027e-05, | |
| "loss": 1.3154, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.328, | |
| "grad_norm": 5.629390716552734, | |
| "learning_rate": 2.708918918918919e-05, | |
| "loss": 1.3433, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.336, | |
| "grad_norm": 8.473249435424805, | |
| "learning_rate": 2.700810810810811e-05, | |
| "loss": 1.1474, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.344, | |
| "grad_norm": 3.652617931365967, | |
| "learning_rate": 2.6927027027027028e-05, | |
| "loss": 1.3247, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.352, | |
| "grad_norm": 4.9890055656433105, | |
| "learning_rate": 2.6845945945945944e-05, | |
| "loss": 1.3347, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 5.2355055809021, | |
| "learning_rate": 2.6764864864864864e-05, | |
| "loss": 1.0932, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.368, | |
| "grad_norm": 6.325026512145996, | |
| "learning_rate": 2.6683783783783785e-05, | |
| "loss": 1.4873, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.376, | |
| "grad_norm": 6.78115701675415, | |
| "learning_rate": 2.6602702702702705e-05, | |
| "loss": 1.2311, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.384, | |
| "grad_norm": 4.194353103637695, | |
| "learning_rate": 2.652162162162162e-05, | |
| "loss": 1.2493, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.392, | |
| "grad_norm": 3.8817057609558105, | |
| "learning_rate": 2.644054054054054e-05, | |
| "loss": 1.1237, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 6.7539520263671875, | |
| "learning_rate": 2.635945945945946e-05, | |
| "loss": 1.1135, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.408, | |
| "grad_norm": 9.044737815856934, | |
| "learning_rate": 2.627837837837838e-05, | |
| "loss": 1.2459, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.416, | |
| "grad_norm": 15.829017639160156, | |
| "learning_rate": 2.6197297297297298e-05, | |
| "loss": 1.2803, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.424, | |
| "grad_norm": 10.789520263671875, | |
| "learning_rate": 2.6116216216216218e-05, | |
| "loss": 1.1912, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.432, | |
| "grad_norm": 5.011368274688721, | |
| "learning_rate": 2.6035135135135135e-05, | |
| "loss": 1.0143, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 8.985868453979492, | |
| "learning_rate": 2.5954054054054055e-05, | |
| "loss": 1.118, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.448, | |
| "grad_norm": 6.862995147705078, | |
| "learning_rate": 2.587297297297297e-05, | |
| "loss": 1.1269, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.456, | |
| "grad_norm": 10.972336769104004, | |
| "learning_rate": 2.579189189189189e-05, | |
| "loss": 1.1591, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.464, | |
| "grad_norm": 8.179327011108398, | |
| "learning_rate": 2.5710810810810812e-05, | |
| "loss": 1.154, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.472, | |
| "grad_norm": 11.713990211486816, | |
| "learning_rate": 2.5629729729729732e-05, | |
| "loss": 1.0995, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 10.86710262298584, | |
| "learning_rate": 2.554864864864865e-05, | |
| "loss": 1.1544, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.488, | |
| "grad_norm": 6.228063106536865, | |
| "learning_rate": 2.546756756756757e-05, | |
| "loss": 1.2395, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.496, | |
| "grad_norm": 12.631518363952637, | |
| "learning_rate": 2.538648648648649e-05, | |
| "loss": 1.0992, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.504, | |
| "grad_norm": 7.058006763458252, | |
| "learning_rate": 2.530540540540541e-05, | |
| "loss": 1.194, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.512, | |
| "grad_norm": 5.026750087738037, | |
| "learning_rate": 2.5224324324324325e-05, | |
| "loss": 1.103, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 7.1134843826293945, | |
| "learning_rate": 2.5143243243243242e-05, | |
| "loss": 0.9427, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.528, | |
| "grad_norm": 7.147433280944824, | |
| "learning_rate": 2.5062162162162162e-05, | |
| "loss": 0.9881, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.536, | |
| "grad_norm": 6.535639762878418, | |
| "learning_rate": 2.4981081081081082e-05, | |
| "loss": 1.1143, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.544, | |
| "grad_norm": 10.878937721252441, | |
| "learning_rate": 2.49e-05, | |
| "loss": 0.8909, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.552, | |
| "grad_norm": 5.79094934463501, | |
| "learning_rate": 2.481891891891892e-05, | |
| "loss": 0.9728, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 6.935592174530029, | |
| "learning_rate": 2.473783783783784e-05, | |
| "loss": 1.0735, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.568, | |
| "grad_norm": 5.661824703216553, | |
| "learning_rate": 2.465675675675676e-05, | |
| "loss": 1.0012, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.576, | |
| "grad_norm": 13.233421325683594, | |
| "learning_rate": 2.4575675675675676e-05, | |
| "loss": 1.0315, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.584, | |
| "grad_norm": 9.292459487915039, | |
| "learning_rate": 2.4494594594594596e-05, | |
| "loss": 0.9547, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.592, | |
| "grad_norm": 13.138367652893066, | |
| "learning_rate": 2.442162162162162e-05, | |
| "loss": 0.9379, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 13.352531433105469, | |
| "learning_rate": 2.434054054054054e-05, | |
| "loss": 0.9484, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.608, | |
| "grad_norm": 11.993139266967773, | |
| "learning_rate": 2.4259459459459458e-05, | |
| "loss": 1.1064, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.616, | |
| "grad_norm": 12.132452011108398, | |
| "learning_rate": 2.4178378378378378e-05, | |
| "loss": 1.1363, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.624, | |
| "grad_norm": 13.944737434387207, | |
| "learning_rate": 2.4097297297297298e-05, | |
| "loss": 0.9835, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.632, | |
| "grad_norm": 6.077609062194824, | |
| "learning_rate": 2.4016216216216218e-05, | |
| "loss": 0.8391, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 7.873855113983154, | |
| "learning_rate": 2.3935135135135135e-05, | |
| "loss": 0.7772, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.648, | |
| "grad_norm": 13.312115669250488, | |
| "learning_rate": 2.3854054054054055e-05, | |
| "loss": 1.0117, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.656, | |
| "grad_norm": 9.016510963439941, | |
| "learning_rate": 2.3772972972972975e-05, | |
| "loss": 0.9353, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.664, | |
| "grad_norm": 8.618375778198242, | |
| "learning_rate": 2.3691891891891895e-05, | |
| "loss": 0.9598, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.672, | |
| "grad_norm": 10.867205619812012, | |
| "learning_rate": 2.361081081081081e-05, | |
| "loss": 0.8726, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 13.182415962219238, | |
| "learning_rate": 2.3529729729729728e-05, | |
| "loss": 0.9202, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.688, | |
| "grad_norm": 12.405129432678223, | |
| "learning_rate": 2.3448648648648648e-05, | |
| "loss": 0.8795, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.696, | |
| "grad_norm": 8.207524299621582, | |
| "learning_rate": 2.3367567567567568e-05, | |
| "loss": 0.8015, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.704, | |
| "grad_norm": 15.442817687988281, | |
| "learning_rate": 2.3286486486486485e-05, | |
| "loss": 0.9932, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.712, | |
| "grad_norm": 13.388226509094238, | |
| "learning_rate": 2.3205405405405405e-05, | |
| "loss": 0.87, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 8.635920524597168, | |
| "learning_rate": 2.3124324324324325e-05, | |
| "loss": 0.7842, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.728, | |
| "grad_norm": 11.66073989868164, | |
| "learning_rate": 2.3043243243243245e-05, | |
| "loss": 0.9023, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.736, | |
| "grad_norm": 12.954612731933594, | |
| "learning_rate": 2.2962162162162162e-05, | |
| "loss": 1.0076, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.744, | |
| "grad_norm": 11.18680191040039, | |
| "learning_rate": 2.2881081081081082e-05, | |
| "loss": 1.1349, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.752, | |
| "grad_norm": 8.514711380004883, | |
| "learning_rate": 2.2800000000000002e-05, | |
| "loss": 0.9604, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 4.436418056488037, | |
| "learning_rate": 2.2718918918918922e-05, | |
| "loss": 0.9632, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.768, | |
| "grad_norm": 10.213781356811523, | |
| "learning_rate": 2.263783783783784e-05, | |
| "loss": 0.8139, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.776, | |
| "grad_norm": 9.987252235412598, | |
| "learning_rate": 2.2556756756756755e-05, | |
| "loss": 0.8276, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.784, | |
| "grad_norm": 12.511467933654785, | |
| "learning_rate": 2.2475675675675675e-05, | |
| "loss": 0.8709, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.792, | |
| "grad_norm": 8.908098220825195, | |
| "learning_rate": 2.2394594594594595e-05, | |
| "loss": 0.8812, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 10.62246322631836, | |
| "learning_rate": 2.2313513513513512e-05, | |
| "loss": 0.9733, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.808, | |
| "grad_norm": 14.651544570922852, | |
| "learning_rate": 2.2232432432432432e-05, | |
| "loss": 1.0309, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.816, | |
| "grad_norm": 19.1525936126709, | |
| "learning_rate": 2.2151351351351352e-05, | |
| "loss": 0.8808, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.824, | |
| "grad_norm": 7.289106369018555, | |
| "learning_rate": 2.2070270270270272e-05, | |
| "loss": 0.9126, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.832, | |
| "grad_norm": 5.375001907348633, | |
| "learning_rate": 2.198918918918919e-05, | |
| "loss": 0.847, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 8.623431205749512, | |
| "learning_rate": 2.190810810810811e-05, | |
| "loss": 0.9139, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.848, | |
| "grad_norm": 6.639071941375732, | |
| "learning_rate": 2.182702702702703e-05, | |
| "loss": 0.9345, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.856, | |
| "grad_norm": 7.635943412780762, | |
| "learning_rate": 2.174594594594595e-05, | |
| "loss": 0.8134, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.864, | |
| "grad_norm": 12.048315048217773, | |
| "learning_rate": 2.1664864864864862e-05, | |
| "loss": 0.6728, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.872, | |
| "grad_norm": 13.869949340820312, | |
| "learning_rate": 2.1583783783783783e-05, | |
| "loss": 0.8256, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 26.233325958251953, | |
| "learning_rate": 2.1502702702702703e-05, | |
| "loss": 0.7044, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.888, | |
| "grad_norm": 7.98716926574707, | |
| "learning_rate": 2.1421621621621623e-05, | |
| "loss": 0.7398, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.896, | |
| "grad_norm": 13.682205200195312, | |
| "learning_rate": 2.134054054054054e-05, | |
| "loss": 0.7522, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.904, | |
| "grad_norm": 9.086796760559082, | |
| "learning_rate": 2.125945945945946e-05, | |
| "loss": 0.8574, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.912, | |
| "grad_norm": 10.3043851852417, | |
| "learning_rate": 2.117837837837838e-05, | |
| "loss": 0.8005, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 12.477950096130371, | |
| "learning_rate": 2.10972972972973e-05, | |
| "loss": 0.8436, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.928, | |
| "grad_norm": 16.634178161621094, | |
| "learning_rate": 2.1016216216216216e-05, | |
| "loss": 0.6515, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.936, | |
| "grad_norm": 11.066425323486328, | |
| "learning_rate": 2.0935135135135136e-05, | |
| "loss": 0.8689, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.944, | |
| "grad_norm": 19.47179412841797, | |
| "learning_rate": 2.0854054054054056e-05, | |
| "loss": 0.6605, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.952, | |
| "grad_norm": 3.917236804962158, | |
| "learning_rate": 2.0772972972972973e-05, | |
| "loss": 0.6826, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 16.43979263305664, | |
| "learning_rate": 2.069189189189189e-05, | |
| "loss": 0.7731, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.968, | |
| "grad_norm": 6.7848711013793945, | |
| "learning_rate": 2.061081081081081e-05, | |
| "loss": 0.705, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 0.976, | |
| "grad_norm": 7.472936153411865, | |
| "learning_rate": 2.052972972972973e-05, | |
| "loss": 0.7663, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.984, | |
| "grad_norm": 5.729743957519531, | |
| "learning_rate": 2.044864864864865e-05, | |
| "loss": 0.9337, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.992, | |
| "grad_norm": 6.306894302368164, | |
| "learning_rate": 2.0367567567567567e-05, | |
| "loss": 0.7655, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 17.98261260986328, | |
| "learning_rate": 2.0286486486486487e-05, | |
| "loss": 0.8202, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_accuracy": 0.7668834417208604, | |
| "eval_f1": 0.7589222033223469, | |
| "eval_loss": 0.7338727712631226, | |
| "eval_model_preparation_time": 0.0029, | |
| "eval_runtime": 59.9403, | |
| "eval_samples_per_second": 33.35, | |
| "eval_steps_per_second": 4.171, | |
| "step": 1250 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 3750, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 4.4103920295511066e+17, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |