| { | |
| "best_metric": 0.011560924351215363, | |
| "best_model_checkpoint": "/media/metricspace/Data1/LLamaT5/TLAN-T5/checkpoint-4078", | |
| "epoch": 14.964143426294822, | |
| "global_step": 4695, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 4.989350372736955e-05, | |
| "loss": 39.1826, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 4.9787007454739086e-05, | |
| "loss": 29.1838, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.968051118210863e-05, | |
| "loss": 20.9138, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 4.957401490947817e-05, | |
| "loss": 8.3513, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 4.9467518636847715e-05, | |
| "loss": 4.4736, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 4.936102236421725e-05, | |
| "loss": 3.3849, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 4.925452609158679e-05, | |
| "loss": 2.2067, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.9148029818956336e-05, | |
| "loss": 0.5818, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 4.904153354632588e-05, | |
| "loss": 0.1404, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 4.893503727369542e-05, | |
| "loss": 0.0717, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 4.8828541001064965e-05, | |
| "loss": 0.0578, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.872204472843451e-05, | |
| "loss": 0.0405, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.861554845580405e-05, | |
| "loss": 0.0468, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.8509052183173594e-05, | |
| "loss": 0.0277, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 4.840255591054313e-05, | |
| "loss": 0.0327, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 4.829605963791268e-05, | |
| "loss": 0.0297, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.818956336528222e-05, | |
| "loss": 0.0242, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.8083067092651754e-05, | |
| "loss": 0.0453, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.79765708200213e-05, | |
| "loss": 0.0239, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 4.7870074547390844e-05, | |
| "loss": 0.0451, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 4.776357827476038e-05, | |
| "loss": 0.0186, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.765708200212993e-05, | |
| "loss": 0.0367, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 4.755058572949947e-05, | |
| "loss": 0.0152, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 4.744408945686901e-05, | |
| "loss": 0.0207, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 4.7337593184238556e-05, | |
| "loss": 0.016, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 4.7231096911608095e-05, | |
| "loss": 0.0235, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.712460063897764e-05, | |
| "loss": 0.0196, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 4.7018104366347185e-05, | |
| "loss": 0.0216, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.691160809371672e-05, | |
| "loss": 0.0278, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 4.680511182108626e-05, | |
| "loss": 0.0297, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 4.669861554845581e-05, | |
| "loss": 0.0215, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_gen_len": 19.0, | |
| "eval_loss": 0.015359184704720974, | |
| "eval_rouge1": 31.9011, | |
| "eval_rouge2": 22.9103, | |
| "eval_rougeL": 31.716, | |
| "eval_rougeLsum": 31.8299, | |
| "eval_runtime": 717.4413, | |
| "eval_samples_per_second": 0.389, | |
| "eval_steps_per_second": 0.195, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 4.6592119275825345e-05, | |
| "loss": 0.0171, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 4.648562300319489e-05, | |
| "loss": 0.0242, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 4.637912673056443e-05, | |
| "loss": 0.0216, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 4.6272630457933974e-05, | |
| "loss": 0.0135, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 4.616613418530352e-05, | |
| "loss": 0.0183, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 4.605963791267306e-05, | |
| "loss": 0.0234, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 4.59531416400426e-05, | |
| "loss": 0.03, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 4.584664536741215e-05, | |
| "loss": 0.021, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 4.5740149094781686e-05, | |
| "loss": 0.0212, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 4.563365282215123e-05, | |
| "loss": 0.0121, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 4.552715654952077e-05, | |
| "loss": 0.0148, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 4.542066027689031e-05, | |
| "loss": 0.0155, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 4.531416400425985e-05, | |
| "loss": 0.0213, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 4.520766773162939e-05, | |
| "loss": 0.0193, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 4.5101171458998936e-05, | |
| "loss": 0.0191, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 4.499467518636848e-05, | |
| "loss": 0.0314, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 4.488817891373802e-05, | |
| "loss": 0.0185, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 4.4781682641107565e-05, | |
| "loss": 0.0152, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 4.46751863684771e-05, | |
| "loss": 0.0361, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 4.456869009584665e-05, | |
| "loss": 0.0189, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 4.4462193823216194e-05, | |
| "loss": 0.0149, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 4.435569755058573e-05, | |
| "loss": 0.0236, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 4.424920127795527e-05, | |
| "loss": 0.0328, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 4.4142705005324815e-05, | |
| "loss": 0.0182, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 4.4036208732694354e-05, | |
| "loss": 0.015, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 4.39297124600639e-05, | |
| "loss": 0.0123, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 4.3823216187433444e-05, | |
| "loss": 0.0175, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 4.371671991480298e-05, | |
| "loss": 0.0164, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 4.361022364217253e-05, | |
| "loss": 0.0156, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 4.3503727369542066e-05, | |
| "loss": 0.0109, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 4.339723109691161e-05, | |
| "loss": 0.0186, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_gen_len": 17.795698924731184, | |
| "eval_loss": 0.013456961140036583, | |
| "eval_rouge1": 39.398, | |
| "eval_rouge2": 31.4448, | |
| "eval_rougeL": 39.0922, | |
| "eval_rougeLsum": 39.237, | |
| "eval_runtime": 716.8614, | |
| "eval_samples_per_second": 0.389, | |
| "eval_steps_per_second": 0.195, | |
| "step": 627 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 4.3290734824281156e-05, | |
| "loss": 0.0119, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 4.3184238551650695e-05, | |
| "loss": 0.0169, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 4.307774227902024e-05, | |
| "loss": 0.0157, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 4.297124600638978e-05, | |
| "loss": 0.0246, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 4.2864749733759316e-05, | |
| "loss": 0.0389, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 4.275825346112886e-05, | |
| "loss": 0.02, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 4.265175718849841e-05, | |
| "loss": 0.0176, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 4.2545260915867945e-05, | |
| "loss": 0.0147, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 4.243876464323749e-05, | |
| "loss": 0.0153, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 4.233226837060703e-05, | |
| "loss": 0.0134, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 4.2225772097976574e-05, | |
| "loss": 0.0119, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 4.211927582534612e-05, | |
| "loss": 0.0256, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 4.201277955271566e-05, | |
| "loss": 0.0196, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 4.19062832800852e-05, | |
| "loss": 0.0149, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 4.179978700745474e-05, | |
| "loss": 0.0204, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 4.169329073482428e-05, | |
| "loss": 0.0182, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 4.1586794462193824e-05, | |
| "loss": 0.0142, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 4.148029818956336e-05, | |
| "loss": 0.0095, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 4.137380191693291e-05, | |
| "loss": 0.0108, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 4.126730564430245e-05, | |
| "loss": 0.0123, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 4.116080937167199e-05, | |
| "loss": 0.0226, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 4.1054313099041536e-05, | |
| "loss": 0.0224, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 4.094781682641108e-05, | |
| "loss": 0.0147, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 4.084132055378062e-05, | |
| "loss": 0.0158, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 4.0734824281150165e-05, | |
| "loss": 0.0118, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 4.06283280085197e-05, | |
| "loss": 0.0272, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 4.052183173588925e-05, | |
| "loss": 0.0148, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 4.041533546325879e-05, | |
| "loss": 0.0314, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 4.0308839190628325e-05, | |
| "loss": 0.016, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 4.020234291799787e-05, | |
| "loss": 0.0146, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 4.0095846645367415e-05, | |
| "loss": 0.0159, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 3.9989350372736954e-05, | |
| "loss": 0.0166, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_gen_len": 16.892473118279568, | |
| "eval_loss": 0.01282462291419506, | |
| "eval_rouge1": 44.2185, | |
| "eval_rouge2": 36.802, | |
| "eval_rougeL": 44.0029, | |
| "eval_rougeLsum": 44.0496, | |
| "eval_runtime": 709.8988, | |
| "eval_samples_per_second": 0.393, | |
| "eval_steps_per_second": 0.197, | |
| "step": 941 | |
| }, | |
| { | |
| "epoch": 3.03, | |
| "learning_rate": 3.98828541001065e-05, | |
| "loss": 0.011, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 3.06, | |
| "learning_rate": 3.9776357827476044e-05, | |
| "loss": 0.0136, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 3.09, | |
| "learning_rate": 3.966986155484558e-05, | |
| "loss": 0.0165, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 3.12, | |
| "learning_rate": 3.956336528221513e-05, | |
| "loss": 0.0132, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 3.16, | |
| "learning_rate": 3.9456869009584666e-05, | |
| "loss": 0.0184, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 3.19, | |
| "learning_rate": 3.935037273695421e-05, | |
| "loss": 0.0303, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 3.22, | |
| "learning_rate": 3.9243876464323756e-05, | |
| "loss": 0.0202, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 3.25, | |
| "learning_rate": 3.913738019169329e-05, | |
| "loss": 0.0159, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 3.28, | |
| "learning_rate": 3.903088391906283e-05, | |
| "loss": 0.0246, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 3.31, | |
| "learning_rate": 3.892438764643238e-05, | |
| "loss": 0.0108, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 3.35, | |
| "learning_rate": 3.8817891373801916e-05, | |
| "loss": 0.0167, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 3.38, | |
| "learning_rate": 3.871139510117146e-05, | |
| "loss": 0.0134, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 3.41, | |
| "learning_rate": 3.8604898828541e-05, | |
| "loss": 0.0122, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 3.44, | |
| "learning_rate": 3.8498402555910545e-05, | |
| "loss": 0.0101, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 3.47, | |
| "learning_rate": 3.839190628328009e-05, | |
| "loss": 0.0118, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 3.51, | |
| "learning_rate": 3.828541001064963e-05, | |
| "loss": 0.0139, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 3.54, | |
| "learning_rate": 3.8178913738019174e-05, | |
| "loss": 0.0175, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 3.57, | |
| "learning_rate": 3.807241746538872e-05, | |
| "loss": 0.0156, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 3.6, | |
| "learning_rate": 3.796592119275825e-05, | |
| "loss": 0.0235, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 3.63, | |
| "learning_rate": 3.7859424920127795e-05, | |
| "loss": 0.0124, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 3.67, | |
| "learning_rate": 3.775292864749734e-05, | |
| "loss": 0.0109, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 3.7, | |
| "learning_rate": 3.764643237486688e-05, | |
| "loss": 0.0189, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 3.73, | |
| "learning_rate": 3.7539936102236424e-05, | |
| "loss": 0.0134, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 3.76, | |
| "learning_rate": 3.743343982960596e-05, | |
| "loss": 0.0314, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 3.79, | |
| "learning_rate": 3.732694355697551e-05, | |
| "loss": 0.0377, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 3.82, | |
| "learning_rate": 3.722044728434505e-05, | |
| "loss": 0.015, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 3.86, | |
| "learning_rate": 3.711395101171459e-05, | |
| "loss": 0.0129, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 3.89, | |
| "learning_rate": 3.7007454739084136e-05, | |
| "loss": 0.0154, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 3.92, | |
| "learning_rate": 3.6900958466453675e-05, | |
| "loss": 0.0117, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 3.95, | |
| "learning_rate": 3.679446219382322e-05, | |
| "loss": 0.0125, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 3.98, | |
| "learning_rate": 3.668796592119276e-05, | |
| "loss": 0.0103, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "eval_gen_len": 14.734767025089607, | |
| "eval_loss": 0.012266965582966805, | |
| "eval_rouge1": 54.8551, | |
| "eval_rouge2": 49.2036, | |
| "eval_rougeL": 54.5865, | |
| "eval_rougeLsum": 54.8397, | |
| "eval_runtime": 674.9391, | |
| "eval_samples_per_second": 0.413, | |
| "eval_steps_per_second": 0.207, | |
| "step": 1255 | |
| }, | |
| { | |
| "epoch": 4.02, | |
| "learning_rate": 3.65814696485623e-05, | |
| "loss": 0.0177, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 4.05, | |
| "learning_rate": 3.647497337593184e-05, | |
| "loss": 0.0189, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 4.08, | |
| "learning_rate": 3.636847710330139e-05, | |
| "loss": 0.0113, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 4.11, | |
| "learning_rate": 3.6261980830670925e-05, | |
| "loss": 0.0277, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 4.14, | |
| "learning_rate": 3.615548455804047e-05, | |
| "loss": 0.0236, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 4.18, | |
| "learning_rate": 3.6048988285410015e-05, | |
| "loss": 0.0198, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 4.21, | |
| "learning_rate": 3.5942492012779554e-05, | |
| "loss": 0.0155, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 4.24, | |
| "learning_rate": 3.58359957401491e-05, | |
| "loss": 0.0138, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 4.27, | |
| "learning_rate": 3.572949946751864e-05, | |
| "loss": 0.0154, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 4.3, | |
| "learning_rate": 3.562300319488818e-05, | |
| "loss": 0.0106, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 4.33, | |
| "learning_rate": 3.551650692225773e-05, | |
| "loss": 0.0168, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 4.37, | |
| "learning_rate": 3.541001064962726e-05, | |
| "loss": 0.0114, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 4.4, | |
| "learning_rate": 3.5303514376996804e-05, | |
| "loss": 0.0119, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 4.43, | |
| "learning_rate": 3.519701810436635e-05, | |
| "loss": 0.0168, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 4.46, | |
| "learning_rate": 3.509052183173589e-05, | |
| "loss": 0.012, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 4.49, | |
| "learning_rate": 3.498402555910543e-05, | |
| "loss": 0.011, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 4.53, | |
| "learning_rate": 3.487752928647498e-05, | |
| "loss": 0.014, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 4.56, | |
| "learning_rate": 3.4771033013844516e-05, | |
| "loss": 0.0151, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 4.59, | |
| "learning_rate": 3.466453674121406e-05, | |
| "loss": 0.0148, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 4.62, | |
| "learning_rate": 3.45580404685836e-05, | |
| "loss": 0.0142, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 4.65, | |
| "learning_rate": 3.4451544195953145e-05, | |
| "loss": 0.0183, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 4.69, | |
| "learning_rate": 3.434504792332269e-05, | |
| "loss": 0.0249, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 4.72, | |
| "learning_rate": 3.423855165069223e-05, | |
| "loss": 0.0147, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 4.75, | |
| "learning_rate": 3.413205537806177e-05, | |
| "loss": 0.0205, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 4.78, | |
| "learning_rate": 3.402555910543131e-05, | |
| "loss": 0.0298, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 4.81, | |
| "learning_rate": 3.391906283280085e-05, | |
| "loss": 0.009, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 4.84, | |
| "learning_rate": 3.3812566560170395e-05, | |
| "loss": 0.0167, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 4.88, | |
| "learning_rate": 3.3706070287539934e-05, | |
| "loss": 0.0108, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 4.91, | |
| "learning_rate": 3.359957401490948e-05, | |
| "loss": 0.0126, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 4.94, | |
| "learning_rate": 3.3493077742279024e-05, | |
| "loss": 0.0201, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 4.97, | |
| "learning_rate": 3.338658146964856e-05, | |
| "loss": 0.0132, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "eval_gen_len": 13.17921146953405, | |
| "eval_loss": 0.012017679400742054, | |
| "eval_rouge1": 61.0563, | |
| "eval_rouge2": 56.3505, | |
| "eval_rougeL": 60.775, | |
| "eval_rougeLsum": 60.928, | |
| "eval_runtime": 625.9017, | |
| "eval_samples_per_second": 0.446, | |
| "eval_steps_per_second": 0.224, | |
| "step": 1568 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "learning_rate": 3.328008519701811e-05, | |
| "loss": 0.0088, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 5.04, | |
| "learning_rate": 3.317358892438765e-05, | |
| "loss": 0.012, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 5.07, | |
| "learning_rate": 3.306709265175719e-05, | |
| "loss": 0.0094, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 5.1, | |
| "learning_rate": 3.2960596379126736e-05, | |
| "loss": 0.0263, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 5.13, | |
| "learning_rate": 3.2854100106496274e-05, | |
| "loss": 0.0139, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 5.16, | |
| "learning_rate": 3.274760383386581e-05, | |
| "loss": 0.0111, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 5.2, | |
| "learning_rate": 3.264110756123536e-05, | |
| "loss": 0.0099, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 5.23, | |
| "learning_rate": 3.2534611288604896e-05, | |
| "loss": 0.0284, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 5.26, | |
| "learning_rate": 3.242811501597444e-05, | |
| "loss": 0.023, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 5.29, | |
| "learning_rate": 3.2321618743343987e-05, | |
| "loss": 0.0077, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 5.32, | |
| "learning_rate": 3.2215122470713525e-05, | |
| "loss": 0.0167, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 5.35, | |
| "learning_rate": 3.210862619808307e-05, | |
| "loss": 0.0096, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 5.39, | |
| "learning_rate": 3.2002129925452615e-05, | |
| "loss": 0.0197, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 5.42, | |
| "learning_rate": 3.1895633652822154e-05, | |
| "loss": 0.0105, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 5.45, | |
| "learning_rate": 3.17891373801917e-05, | |
| "loss": 0.0308, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 5.48, | |
| "learning_rate": 3.168264110756124e-05, | |
| "loss": 0.0118, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 5.51, | |
| "learning_rate": 3.1576144834930775e-05, | |
| "loss": 0.0128, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 5.55, | |
| "learning_rate": 3.146964856230032e-05, | |
| "loss": 0.0142, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 5.58, | |
| "learning_rate": 3.136315228966986e-05, | |
| "loss": 0.0086, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 5.61, | |
| "learning_rate": 3.1256656017039404e-05, | |
| "loss": 0.0189, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 5.64, | |
| "learning_rate": 3.115015974440895e-05, | |
| "loss": 0.0137, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 5.67, | |
| "learning_rate": 3.104366347177849e-05, | |
| "loss": 0.0117, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 5.71, | |
| "learning_rate": 3.093716719914803e-05, | |
| "loss": 0.0112, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 5.74, | |
| "learning_rate": 3.083067092651757e-05, | |
| "loss": 0.017, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 5.77, | |
| "learning_rate": 3.0724174653887116e-05, | |
| "loss": 0.0089, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 5.8, | |
| "learning_rate": 3.061767838125666e-05, | |
| "loss": 0.0354, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 5.83, | |
| "learning_rate": 3.0511182108626203e-05, | |
| "loss": 0.0173, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 5.86, | |
| "learning_rate": 3.0404685835995745e-05, | |
| "loss": 0.0132, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 5.9, | |
| "learning_rate": 3.029818956336528e-05, | |
| "loss": 0.0101, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 5.93, | |
| "learning_rate": 3.0191693290734825e-05, | |
| "loss": 0.0172, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 5.96, | |
| "learning_rate": 3.0085197018104367e-05, | |
| "loss": 0.0109, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 5.99, | |
| "learning_rate": 2.997870074547391e-05, | |
| "loss": 0.0179, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "eval_gen_len": 12.727598566308243, | |
| "eval_loss": 0.011885586194694042, | |
| "eval_rouge1": 63.4215, | |
| "eval_rouge2": 59.2435, | |
| "eval_rougeL": 63.1198, | |
| "eval_rougeLsum": 63.3059, | |
| "eval_runtime": 611.8658, | |
| "eval_samples_per_second": 0.456, | |
| "eval_steps_per_second": 0.229, | |
| "step": 1882 | |
| }, | |
| { | |
| "epoch": 6.02, | |
| "learning_rate": 2.987220447284345e-05, | |
| "loss": 0.0157, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 6.06, | |
| "learning_rate": 2.9765708200212995e-05, | |
| "loss": 0.0173, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 6.09, | |
| "learning_rate": 2.9659211927582537e-05, | |
| "loss": 0.0132, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 6.12, | |
| "learning_rate": 2.955271565495208e-05, | |
| "loss": 0.0253, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 6.15, | |
| "learning_rate": 2.944621938232162e-05, | |
| "loss": 0.0158, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 6.18, | |
| "learning_rate": 2.9339723109691166e-05, | |
| "loss": 0.0112, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 6.22, | |
| "learning_rate": 2.9233226837060707e-05, | |
| "loss": 0.0097, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 6.25, | |
| "learning_rate": 2.912673056443025e-05, | |
| "loss": 0.0101, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 6.28, | |
| "learning_rate": 2.9020234291799788e-05, | |
| "loss": 0.0197, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 6.31, | |
| "learning_rate": 2.891373801916933e-05, | |
| "loss": 0.0138, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 6.34, | |
| "learning_rate": 2.880724174653887e-05, | |
| "loss": 0.0123, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 6.37, | |
| "learning_rate": 2.8700745473908413e-05, | |
| "loss": 0.0187, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 6.41, | |
| "learning_rate": 2.8594249201277955e-05, | |
| "loss": 0.0164, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 6.44, | |
| "learning_rate": 2.84877529286475e-05, | |
| "loss": 0.0097, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 6.47, | |
| "learning_rate": 2.838125665601704e-05, | |
| "loss": 0.017, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 6.5, | |
| "learning_rate": 2.8274760383386583e-05, | |
| "loss": 0.0072, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 6.53, | |
| "learning_rate": 2.8168264110756125e-05, | |
| "loss": 0.0173, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 6.57, | |
| "learning_rate": 2.806176783812567e-05, | |
| "loss": 0.0196, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 6.6, | |
| "learning_rate": 2.7955271565495212e-05, | |
| "loss": 0.0255, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 6.63, | |
| "learning_rate": 2.7848775292864754e-05, | |
| "loss": 0.019, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 6.66, | |
| "learning_rate": 2.7742279020234292e-05, | |
| "loss": 0.0103, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 6.69, | |
| "learning_rate": 2.7635782747603834e-05, | |
| "loss": 0.0173, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 6.73, | |
| "learning_rate": 2.7529286474973375e-05, | |
| "loss": 0.0113, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 6.76, | |
| "learning_rate": 2.7422790202342917e-05, | |
| "loss": 0.0189, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 6.79, | |
| "learning_rate": 2.7316293929712462e-05, | |
| "loss": 0.0149, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 6.82, | |
| "learning_rate": 2.7209797657082004e-05, | |
| "loss": 0.0128, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 6.85, | |
| "learning_rate": 2.7103301384451546e-05, | |
| "loss": 0.0181, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 6.88, | |
| "learning_rate": 2.6996805111821088e-05, | |
| "loss": 0.0097, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 6.92, | |
| "learning_rate": 2.6890308839190633e-05, | |
| "loss": 0.013, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 6.95, | |
| "learning_rate": 2.6783812566560174e-05, | |
| "loss": 0.014, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 6.98, | |
| "learning_rate": 2.6677316293929716e-05, | |
| "loss": 0.0115, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 7.0, | |
| "eval_gen_len": 12.92831541218638, | |
| "eval_loss": 0.011721653863787651, | |
| "eval_rouge1": 62.5874, | |
| "eval_rouge2": 58.2782, | |
| "eval_rougeL": 62.2116, | |
| "eval_rougeLsum": 62.3846, | |
| "eval_runtime": 618.8742, | |
| "eval_samples_per_second": 0.451, | |
| "eval_steps_per_second": 0.226, | |
| "step": 2196 | |
| }, | |
| { | |
| "epoch": 7.01, | |
| "learning_rate": 2.6570820021299255e-05, | |
| "loss": 0.0124, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 7.04, | |
| "learning_rate": 2.6464323748668796e-05, | |
| "loss": 0.0166, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 7.08, | |
| "learning_rate": 2.6357827476038338e-05, | |
| "loss": 0.0099, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 7.11, | |
| "learning_rate": 2.625133120340788e-05, | |
| "loss": 0.0113, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 7.14, | |
| "learning_rate": 2.6144834930777425e-05, | |
| "loss": 0.0174, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 7.17, | |
| "learning_rate": 2.6038338658146967e-05, | |
| "loss": 0.0095, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 7.2, | |
| "learning_rate": 2.593184238551651e-05, | |
| "loss": 0.0068, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 7.24, | |
| "learning_rate": 2.582534611288605e-05, | |
| "loss": 0.0291, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 7.27, | |
| "learning_rate": 2.5718849840255592e-05, | |
| "loss": 0.0171, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 7.3, | |
| "learning_rate": 2.5612353567625137e-05, | |
| "loss": 0.0183, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 7.33, | |
| "learning_rate": 2.550585729499468e-05, | |
| "loss": 0.0105, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 7.36, | |
| "learning_rate": 2.539936102236422e-05, | |
| "loss": 0.0227, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 7.39, | |
| "learning_rate": 2.529286474973376e-05, | |
| "loss": 0.0311, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 7.43, | |
| "learning_rate": 2.51863684771033e-05, | |
| "loss": 0.0133, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 7.46, | |
| "learning_rate": 2.5079872204472842e-05, | |
| "loss": 0.0102, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 7.49, | |
| "learning_rate": 2.4973375931842384e-05, | |
| "loss": 0.0189, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 7.52, | |
| "learning_rate": 2.486687965921193e-05, | |
| "loss": 0.0157, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 7.55, | |
| "learning_rate": 2.476038338658147e-05, | |
| "loss": 0.0114, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 7.59, | |
| "learning_rate": 2.4653887113951013e-05, | |
| "loss": 0.0131, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 7.62, | |
| "learning_rate": 2.4547390841320554e-05, | |
| "loss": 0.0144, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 7.65, | |
| "learning_rate": 2.44408945686901e-05, | |
| "loss": 0.0107, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 7.68, | |
| "learning_rate": 2.4334398296059638e-05, | |
| "loss": 0.0089, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 7.71, | |
| "learning_rate": 2.422790202342918e-05, | |
| "loss": 0.0119, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 7.75, | |
| "learning_rate": 2.412140575079872e-05, | |
| "loss": 0.0136, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 7.78, | |
| "learning_rate": 2.4014909478168267e-05, | |
| "loss": 0.026, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 7.81, | |
| "learning_rate": 2.390841320553781e-05, | |
| "loss": 0.0147, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 7.84, | |
| "learning_rate": 2.380191693290735e-05, | |
| "loss": 0.0114, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 7.87, | |
| "learning_rate": 2.3695420660276892e-05, | |
| "loss": 0.0122, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 7.9, | |
| "learning_rate": 2.3588924387646434e-05, | |
| "loss": 0.0115, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 7.94, | |
| "learning_rate": 2.3482428115015975e-05, | |
| "loss": 0.016, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 7.97, | |
| "learning_rate": 2.3375931842385517e-05, | |
| "loss": 0.0177, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "learning_rate": 2.326943556975506e-05, | |
| "loss": 0.0097, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "eval_gen_len": 12.827956989247312, | |
| "eval_loss": 0.0116655882447958, | |
| "eval_rouge1": 63.06, | |
| "eval_rouge2": 58.8818, | |
| "eval_rougeL": 62.7484, | |
| "eval_rougeLsum": 62.8736, | |
| "eval_runtime": 618.883, | |
| "eval_samples_per_second": 0.451, | |
| "eval_steps_per_second": 0.226, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 8.03, | |
| "learning_rate": 2.3162939297124604e-05, | |
| "loss": 0.0122, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 8.06, | |
| "learning_rate": 2.3056443024494142e-05, | |
| "loss": 0.0125, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 8.1, | |
| "learning_rate": 2.2949946751863684e-05, | |
| "loss": 0.013, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 8.13, | |
| "learning_rate": 2.284345047923323e-05, | |
| "loss": 0.0138, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 8.16, | |
| "learning_rate": 2.273695420660277e-05, | |
| "loss": 0.017, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 8.19, | |
| "learning_rate": 2.2630457933972313e-05, | |
| "loss": 0.0199, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 8.22, | |
| "learning_rate": 2.2523961661341854e-05, | |
| "loss": 0.0157, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 8.25, | |
| "learning_rate": 2.2417465388711396e-05, | |
| "loss": 0.0092, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 8.29, | |
| "learning_rate": 2.2310969116080938e-05, | |
| "loss": 0.011, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 8.32, | |
| "learning_rate": 2.220447284345048e-05, | |
| "loss": 0.012, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 8.35, | |
| "learning_rate": 2.209797657082002e-05, | |
| "loss": 0.0155, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 8.38, | |
| "learning_rate": 2.1991480298189567e-05, | |
| "loss": 0.0132, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 8.41, | |
| "learning_rate": 2.188498402555911e-05, | |
| "loss": 0.0119, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 8.45, | |
| "learning_rate": 2.1778487752928647e-05, | |
| "loss": 0.0147, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 8.48, | |
| "learning_rate": 2.167199148029819e-05, | |
| "loss": 0.0152, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 8.51, | |
| "learning_rate": 2.1565495207667734e-05, | |
| "loss": 0.0158, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 8.54, | |
| "learning_rate": 2.1458998935037275e-05, | |
| "loss": 0.017, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 8.57, | |
| "learning_rate": 2.1352502662406817e-05, | |
| "loss": 0.0143, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 8.61, | |
| "learning_rate": 2.124600638977636e-05, | |
| "loss": 0.015, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 8.64, | |
| "learning_rate": 2.11395101171459e-05, | |
| "loss": 0.0113, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 8.67, | |
| "learning_rate": 2.1033013844515442e-05, | |
| "loss": 0.0131, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 8.7, | |
| "learning_rate": 2.0926517571884984e-05, | |
| "loss": 0.0109, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 8.73, | |
| "learning_rate": 2.0820021299254526e-05, | |
| "loss": 0.0132, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 8.76, | |
| "learning_rate": 2.071352502662407e-05, | |
| "loss": 0.0219, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 8.8, | |
| "learning_rate": 2.0607028753993613e-05, | |
| "loss": 0.0108, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 8.83, | |
| "learning_rate": 2.050053248136315e-05, | |
| "loss": 0.0143, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 8.86, | |
| "learning_rate": 2.0394036208732696e-05, | |
| "loss": 0.0114, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 8.89, | |
| "learning_rate": 2.0287539936102238e-05, | |
| "loss": 0.0136, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 8.92, | |
| "learning_rate": 2.018104366347178e-05, | |
| "loss": 0.0114, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 8.96, | |
| "learning_rate": 2.007454739084132e-05, | |
| "loss": 0.0379, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 8.99, | |
| "learning_rate": 1.9968051118210863e-05, | |
| "loss": 0.0172, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 9.0, | |
| "eval_gen_len": 12.827956989247312, | |
| "eval_loss": 0.011623449623584747, | |
| "eval_rouge1": 63.4646, | |
| "eval_rouge2": 59.358, | |
| "eval_rougeL": 63.1834, | |
| "eval_rougeLsum": 63.3962, | |
| "eval_runtime": 618.8791, | |
| "eval_samples_per_second": 0.451, | |
| "eval_steps_per_second": 0.226, | |
| "step": 2823 | |
| }, | |
| { | |
| "epoch": 9.02, | |
| "learning_rate": 1.9861554845580405e-05, | |
| "loss": 0.0179, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 9.05, | |
| "learning_rate": 1.9755058572949947e-05, | |
| "loss": 0.0116, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 9.08, | |
| "learning_rate": 1.964856230031949e-05, | |
| "loss": 0.0108, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 9.12, | |
| "learning_rate": 1.9542066027689034e-05, | |
| "loss": 0.0126, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 9.15, | |
| "learning_rate": 1.9435569755058575e-05, | |
| "loss": 0.0092, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 9.18, | |
| "learning_rate": 1.9329073482428117e-05, | |
| "loss": 0.0163, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 9.21, | |
| "learning_rate": 1.9222577209797655e-05, | |
| "loss": 0.0162, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 9.24, | |
| "learning_rate": 1.91160809371672e-05, | |
| "loss": 0.0161, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 9.27, | |
| "learning_rate": 1.9009584664536742e-05, | |
| "loss": 0.0335, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 9.31, | |
| "learning_rate": 1.8903088391906284e-05, | |
| "loss": 0.0087, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 9.34, | |
| "learning_rate": 1.8796592119275826e-05, | |
| "loss": 0.0163, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 9.37, | |
| "learning_rate": 1.869009584664537e-05, | |
| "loss": 0.01, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 9.4, | |
| "learning_rate": 1.858359957401491e-05, | |
| "loss": 0.0215, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 9.43, | |
| "learning_rate": 1.847710330138445e-05, | |
| "loss": 0.0137, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 9.47, | |
| "learning_rate": 1.8370607028753993e-05, | |
| "loss": 0.0208, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 9.5, | |
| "learning_rate": 1.8264110756123538e-05, | |
| "loss": 0.026, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 9.53, | |
| "learning_rate": 1.815761448349308e-05, | |
| "loss": 0.0109, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 9.56, | |
| "learning_rate": 1.805111821086262e-05, | |
| "loss": 0.0122, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 9.59, | |
| "learning_rate": 1.7944621938232163e-05, | |
| "loss": 0.0109, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 9.63, | |
| "learning_rate": 1.7838125665601705e-05, | |
| "loss": 0.0203, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 9.66, | |
| "learning_rate": 1.7731629392971247e-05, | |
| "loss": 0.0081, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 9.69, | |
| "learning_rate": 1.762513312034079e-05, | |
| "loss": 0.0166, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 9.72, | |
| "learning_rate": 1.7518636847710333e-05, | |
| "loss": 0.0131, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 9.75, | |
| "learning_rate": 1.7412140575079875e-05, | |
| "loss": 0.0159, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 9.78, | |
| "learning_rate": 1.7305644302449414e-05, | |
| "loss": 0.0089, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 9.82, | |
| "learning_rate": 1.7199148029818955e-05, | |
| "loss": 0.0112, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 9.85, | |
| "learning_rate": 1.70926517571885e-05, | |
| "loss": 0.0165, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 9.88, | |
| "learning_rate": 1.6986155484558042e-05, | |
| "loss": 0.0094, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 9.91, | |
| "learning_rate": 1.6879659211927584e-05, | |
| "loss": 0.0118, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 9.94, | |
| "learning_rate": 1.6773162939297126e-05, | |
| "loss": 0.0125, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 9.98, | |
| "learning_rate": 1.6666666666666667e-05, | |
| "loss": 0.0122, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "eval_gen_len": 12.526881720430108, | |
| "eval_loss": 0.011583242565393448, | |
| "eval_rouge1": 63.9777, | |
| "eval_rouge2": 60.058, | |
| "eval_rougeL": 63.7162, | |
| "eval_rougeLsum": 63.8393, | |
| "eval_runtime": 615.3638, | |
| "eval_samples_per_second": 0.453, | |
| "eval_steps_per_second": 0.228, | |
| "step": 3137 | |
| }, | |
| { | |
| "epoch": 10.01, | |
| "learning_rate": 1.656017039403621e-05, | |
| "loss": 0.0102, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 10.04, | |
| "learning_rate": 1.645367412140575e-05, | |
| "loss": 0.0139, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 10.07, | |
| "learning_rate": 1.6347177848775293e-05, | |
| "loss": 0.0146, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 10.1, | |
| "learning_rate": 1.6240681576144838e-05, | |
| "loss": 0.022, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 10.14, | |
| "learning_rate": 1.6134185303514376e-05, | |
| "loss": 0.0159, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 10.17, | |
| "learning_rate": 1.6027689030883918e-05, | |
| "loss": 0.0124, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 10.2, | |
| "learning_rate": 1.5921192758253463e-05, | |
| "loss": 0.0144, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 10.23, | |
| "learning_rate": 1.5814696485623005e-05, | |
| "loss": 0.0145, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 10.26, | |
| "learning_rate": 1.5708200212992547e-05, | |
| "loss": 0.0134, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 10.29, | |
| "learning_rate": 1.560170394036209e-05, | |
| "loss": 0.0152, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 10.33, | |
| "learning_rate": 1.549520766773163e-05, | |
| "loss": 0.0105, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 10.36, | |
| "learning_rate": 1.5388711395101172e-05, | |
| "loss": 0.0124, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 10.39, | |
| "learning_rate": 1.5282215122470714e-05, | |
| "loss": 0.0179, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 10.42, | |
| "learning_rate": 1.5175718849840257e-05, | |
| "loss": 0.0236, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 10.45, | |
| "learning_rate": 1.5069222577209799e-05, | |
| "loss": 0.0083, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 10.49, | |
| "learning_rate": 1.4962726304579342e-05, | |
| "loss": 0.0232, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 10.52, | |
| "learning_rate": 1.485623003194888e-05, | |
| "loss": 0.0089, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 10.55, | |
| "learning_rate": 1.4749733759318424e-05, | |
| "loss": 0.0116, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 10.58, | |
| "learning_rate": 1.4643237486687966e-05, | |
| "loss": 0.0106, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 10.61, | |
| "learning_rate": 1.453674121405751e-05, | |
| "loss": 0.0327, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 10.65, | |
| "learning_rate": 1.4430244941427051e-05, | |
| "loss": 0.0074, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 10.68, | |
| "learning_rate": 1.4323748668796594e-05, | |
| "loss": 0.0113, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 10.71, | |
| "learning_rate": 1.4217252396166134e-05, | |
| "loss": 0.0139, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 10.74, | |
| "learning_rate": 1.4110756123535676e-05, | |
| "loss": 0.0106, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 10.77, | |
| "learning_rate": 1.4004259850905218e-05, | |
| "loss": 0.011, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 10.8, | |
| "learning_rate": 1.3897763578274761e-05, | |
| "loss": 0.0161, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 10.84, | |
| "learning_rate": 1.3791267305644303e-05, | |
| "loss": 0.0156, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 10.87, | |
| "learning_rate": 1.3684771033013847e-05, | |
| "loss": 0.0076, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 10.9, | |
| "learning_rate": 1.3578274760383387e-05, | |
| "loss": 0.0111, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 10.93, | |
| "learning_rate": 1.3471778487752928e-05, | |
| "loss": 0.0159, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 10.96, | |
| "learning_rate": 1.3365282215122472e-05, | |
| "loss": 0.0138, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 11.0, | |
| "learning_rate": 1.3258785942492014e-05, | |
| "loss": 0.0132, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 11.0, | |
| "eval_gen_len": 12.577060931899641, | |
| "eval_loss": 0.011564880609512329, | |
| "eval_rouge1": 63.7793, | |
| "eval_rouge2": 59.7984, | |
| "eval_rougeL": 63.5653, | |
| "eval_rougeLsum": 63.7034, | |
| "eval_runtime": 615.3689, | |
| "eval_samples_per_second": 0.453, | |
| "eval_steps_per_second": 0.228, | |
| "step": 3451 | |
| }, | |
| { | |
| "epoch": 11.03, | |
| "learning_rate": 1.3152289669861555e-05, | |
| "loss": 0.0129, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 11.06, | |
| "learning_rate": 1.3045793397231099e-05, | |
| "loss": 0.0227, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 11.09, | |
| "learning_rate": 1.2939297124600639e-05, | |
| "loss": 0.0112, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 11.12, | |
| "learning_rate": 1.283280085197018e-05, | |
| "loss": 0.0147, | |
| "step": 3490 | |
| }, | |
| { | |
| "epoch": 11.16, | |
| "learning_rate": 1.2726304579339724e-05, | |
| "loss": 0.0113, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 11.19, | |
| "learning_rate": 1.2619808306709266e-05, | |
| "loss": 0.0123, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 11.22, | |
| "learning_rate": 1.2513312034078809e-05, | |
| "loss": 0.0191, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 11.25, | |
| "learning_rate": 1.240681576144835e-05, | |
| "loss": 0.0128, | |
| "step": 3530 | |
| }, | |
| { | |
| "epoch": 11.28, | |
| "learning_rate": 1.2300319488817893e-05, | |
| "loss": 0.0101, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 11.31, | |
| "learning_rate": 1.2193823216187434e-05, | |
| "loss": 0.0129, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 11.35, | |
| "learning_rate": 1.2087326943556976e-05, | |
| "loss": 0.0093, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 11.38, | |
| "learning_rate": 1.1980830670926518e-05, | |
| "loss": 0.0137, | |
| "step": 3570 | |
| }, | |
| { | |
| "epoch": 11.41, | |
| "learning_rate": 1.1874334398296061e-05, | |
| "loss": 0.0114, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 11.44, | |
| "learning_rate": 1.1767838125665601e-05, | |
| "loss": 0.0164, | |
| "step": 3590 | |
| }, | |
| { | |
| "epoch": 11.47, | |
| "learning_rate": 1.1661341853035145e-05, | |
| "loss": 0.0207, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 11.51, | |
| "learning_rate": 1.1554845580404687e-05, | |
| "loss": 0.0109, | |
| "step": 3610 | |
| }, | |
| { | |
| "epoch": 11.54, | |
| "learning_rate": 1.1448349307774228e-05, | |
| "loss": 0.0169, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 11.57, | |
| "learning_rate": 1.134185303514377e-05, | |
| "loss": 0.0102, | |
| "step": 3630 | |
| }, | |
| { | |
| "epoch": 11.6, | |
| "learning_rate": 1.1235356762513314e-05, | |
| "loss": 0.0248, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 11.63, | |
| "learning_rate": 1.1128860489882854e-05, | |
| "loss": 0.0112, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 11.67, | |
| "learning_rate": 1.1022364217252397e-05, | |
| "loss": 0.02, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 11.7, | |
| "learning_rate": 1.0915867944621939e-05, | |
| "loss": 0.0127, | |
| "step": 3670 | |
| }, | |
| { | |
| "epoch": 11.73, | |
| "learning_rate": 1.080937167199148e-05, | |
| "loss": 0.0147, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 11.76, | |
| "learning_rate": 1.0702875399361024e-05, | |
| "loss": 0.014, | |
| "step": 3690 | |
| }, | |
| { | |
| "epoch": 11.79, | |
| "learning_rate": 1.0596379126730564e-05, | |
| "loss": 0.0079, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 11.82, | |
| "learning_rate": 1.0489882854100107e-05, | |
| "loss": 0.0134, | |
| "step": 3710 | |
| }, | |
| { | |
| "epoch": 11.86, | |
| "learning_rate": 1.038338658146965e-05, | |
| "loss": 0.0188, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 11.89, | |
| "learning_rate": 1.0276890308839191e-05, | |
| "loss": 0.0087, | |
| "step": 3730 | |
| }, | |
| { | |
| "epoch": 11.92, | |
| "learning_rate": 1.0170394036208733e-05, | |
| "loss": 0.0104, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 11.95, | |
| "learning_rate": 1.0063897763578276e-05, | |
| "loss": 0.0132, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 11.98, | |
| "learning_rate": 9.957401490947816e-06, | |
| "loss": 0.0242, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 12.0, | |
| "eval_gen_len": 12.67741935483871, | |
| "eval_loss": 0.011580224148929119, | |
| "eval_rouge1": 63.3071, | |
| "eval_rouge2": 59.216, | |
| "eval_rougeL": 62.9553, | |
| "eval_rougeLsum": 63.1625, | |
| "eval_runtime": 618.8937, | |
| "eval_samples_per_second": 0.451, | |
| "eval_steps_per_second": 0.226, | |
| "step": 3765 | |
| }, | |
| { | |
| "epoch": 12.02, | |
| "learning_rate": 9.85090521831736e-06, | |
| "loss": 0.0084, | |
| "step": 3770 | |
| }, | |
| { | |
| "epoch": 12.05, | |
| "learning_rate": 9.744408945686901e-06, | |
| "loss": 0.0106, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 12.08, | |
| "learning_rate": 9.637912673056443e-06, | |
| "loss": 0.0106, | |
| "step": 3790 | |
| }, | |
| { | |
| "epoch": 12.11, | |
| "learning_rate": 9.531416400425985e-06, | |
| "loss": 0.0111, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 12.14, | |
| "learning_rate": 9.424920127795528e-06, | |
| "loss": 0.0118, | |
| "step": 3810 | |
| }, | |
| { | |
| "epoch": 12.18, | |
| "learning_rate": 9.318423855165068e-06, | |
| "loss": 0.0114, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 12.21, | |
| "learning_rate": 9.211927582534612e-06, | |
| "loss": 0.0134, | |
| "step": 3830 | |
| }, | |
| { | |
| "epoch": 12.24, | |
| "learning_rate": 9.105431309904154e-06, | |
| "loss": 0.0119, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 12.27, | |
| "learning_rate": 8.998935037273695e-06, | |
| "loss": 0.0097, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 12.3, | |
| "learning_rate": 8.892438764643237e-06, | |
| "loss": 0.0167, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 12.33, | |
| "learning_rate": 8.78594249201278e-06, | |
| "loss": 0.0281, | |
| "step": 3870 | |
| }, | |
| { | |
| "epoch": 12.37, | |
| "learning_rate": 8.679446219382322e-06, | |
| "loss": 0.0084, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 12.4, | |
| "learning_rate": 8.572949946751864e-06, | |
| "loss": 0.0105, | |
| "step": 3890 | |
| }, | |
| { | |
| "epoch": 12.43, | |
| "learning_rate": 8.466453674121406e-06, | |
| "loss": 0.0146, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 12.46, | |
| "learning_rate": 8.359957401490947e-06, | |
| "loss": 0.011, | |
| "step": 3910 | |
| }, | |
| { | |
| "epoch": 12.49, | |
| "learning_rate": 8.253461128860491e-06, | |
| "loss": 0.0077, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 12.53, | |
| "learning_rate": 8.146964856230033e-06, | |
| "loss": 0.0278, | |
| "step": 3930 | |
| }, | |
| { | |
| "epoch": 12.56, | |
| "learning_rate": 8.040468583599574e-06, | |
| "loss": 0.0151, | |
| "step": 3940 | |
| }, | |
| { | |
| "epoch": 12.59, | |
| "learning_rate": 7.933972310969116e-06, | |
| "loss": 0.0124, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 12.62, | |
| "learning_rate": 7.82747603833866e-06, | |
| "loss": 0.0265, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 12.65, | |
| "learning_rate": 7.7209797657082e-06, | |
| "loss": 0.014, | |
| "step": 3970 | |
| }, | |
| { | |
| "epoch": 12.69, | |
| "learning_rate": 7.614483493077742e-06, | |
| "loss": 0.0118, | |
| "step": 3980 | |
| }, | |
| { | |
| "epoch": 12.72, | |
| "learning_rate": 7.507987220447285e-06, | |
| "loss": 0.0163, | |
| "step": 3990 | |
| }, | |
| { | |
| "epoch": 12.75, | |
| "learning_rate": 7.4014909478168266e-06, | |
| "loss": 0.0199, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 12.78, | |
| "learning_rate": 7.294994675186369e-06, | |
| "loss": 0.0082, | |
| "step": 4010 | |
| }, | |
| { | |
| "epoch": 12.81, | |
| "learning_rate": 7.188498402555911e-06, | |
| "loss": 0.0067, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 12.84, | |
| "learning_rate": 7.082002129925453e-06, | |
| "loss": 0.015, | |
| "step": 4030 | |
| }, | |
| { | |
| "epoch": 12.88, | |
| "learning_rate": 6.975505857294995e-06, | |
| "loss": 0.0156, | |
| "step": 4040 | |
| }, | |
| { | |
| "epoch": 12.91, | |
| "learning_rate": 6.869009584664538e-06, | |
| "loss": 0.0218, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 12.94, | |
| "learning_rate": 6.762513312034079e-06, | |
| "loss": 0.0165, | |
| "step": 4060 | |
| }, | |
| { | |
| "epoch": 12.97, | |
| "learning_rate": 6.656017039403621e-06, | |
| "loss": 0.0178, | |
| "step": 4070 | |
| }, | |
| { | |
| "epoch": 13.0, | |
| "eval_gen_len": 12.67741935483871, | |
| "eval_loss": 0.011560924351215363, | |
| "eval_rouge1": 63.3607, | |
| "eval_rouge2": 59.1699, | |
| "eval_rougeL": 63.0136, | |
| "eval_rougeLsum": 63.1274, | |
| "eval_runtime": 618.877, | |
| "eval_samples_per_second": 0.451, | |
| "eval_steps_per_second": 0.226, | |
| "step": 4078 | |
| }, | |
| { | |
| "epoch": 13.0, | |
| "learning_rate": 6.549520766773164e-06, | |
| "loss": 0.0092, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 13.04, | |
| "learning_rate": 6.443024494142705e-06, | |
| "loss": 0.0084, | |
| "step": 4090 | |
| }, | |
| { | |
| "epoch": 13.07, | |
| "learning_rate": 6.336528221512247e-06, | |
| "loss": 0.0149, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 13.1, | |
| "learning_rate": 6.230031948881789e-06, | |
| "loss": 0.0138, | |
| "step": 4110 | |
| }, | |
| { | |
| "epoch": 13.13, | |
| "learning_rate": 6.123535676251332e-06, | |
| "loss": 0.0128, | |
| "step": 4120 | |
| }, | |
| { | |
| "epoch": 13.16, | |
| "learning_rate": 6.0170394036208735e-06, | |
| "loss": 0.0124, | |
| "step": 4130 | |
| }, | |
| { | |
| "epoch": 13.2, | |
| "learning_rate": 5.910543130990415e-06, | |
| "loss": 0.0122, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 13.23, | |
| "learning_rate": 5.804046858359958e-06, | |
| "loss": 0.0175, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 13.26, | |
| "learning_rate": 5.6975505857295e-06, | |
| "loss": 0.0142, | |
| "step": 4160 | |
| }, | |
| { | |
| "epoch": 13.29, | |
| "learning_rate": 5.591054313099041e-06, | |
| "loss": 0.009, | |
| "step": 4170 | |
| }, | |
| { | |
| "epoch": 13.32, | |
| "learning_rate": 5.484558040468584e-06, | |
| "loss": 0.0129, | |
| "step": 4180 | |
| }, | |
| { | |
| "epoch": 13.35, | |
| "learning_rate": 5.378061767838126e-06, | |
| "loss": 0.0086, | |
| "step": 4190 | |
| }, | |
| { | |
| "epoch": 13.39, | |
| "learning_rate": 5.2715654952076674e-06, | |
| "loss": 0.0109, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 13.42, | |
| "learning_rate": 5.16506922257721e-06, | |
| "loss": 0.0127, | |
| "step": 4210 | |
| }, | |
| { | |
| "epoch": 13.45, | |
| "learning_rate": 5.058572949946752e-06, | |
| "loss": 0.0235, | |
| "step": 4220 | |
| }, | |
| { | |
| "epoch": 13.48, | |
| "learning_rate": 4.952076677316294e-06, | |
| "loss": 0.0101, | |
| "step": 4230 | |
| }, | |
| { | |
| "epoch": 13.51, | |
| "learning_rate": 4.845580404685836e-06, | |
| "loss": 0.0177, | |
| "step": 4240 | |
| }, | |
| { | |
| "epoch": 13.55, | |
| "learning_rate": 4.739084132055379e-06, | |
| "loss": 0.0242, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 13.58, | |
| "learning_rate": 4.6325878594249205e-06, | |
| "loss": 0.0157, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 13.61, | |
| "learning_rate": 4.526091586794463e-06, | |
| "loss": 0.015, | |
| "step": 4270 | |
| }, | |
| { | |
| "epoch": 13.64, | |
| "learning_rate": 4.419595314164005e-06, | |
| "loss": 0.0139, | |
| "step": 4280 | |
| }, | |
| { | |
| "epoch": 13.67, | |
| "learning_rate": 4.3130990415335465e-06, | |
| "loss": 0.0126, | |
| "step": 4290 | |
| }, | |
| { | |
| "epoch": 13.71, | |
| "learning_rate": 4.206602768903089e-06, | |
| "loss": 0.0137, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 13.74, | |
| "learning_rate": 4.100106496272631e-06, | |
| "loss": 0.0071, | |
| "step": 4310 | |
| }, | |
| { | |
| "epoch": 13.77, | |
| "learning_rate": 3.993610223642173e-06, | |
| "loss": 0.0201, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 13.8, | |
| "learning_rate": 3.887113951011715e-06, | |
| "loss": 0.0188, | |
| "step": 4330 | |
| }, | |
| { | |
| "epoch": 13.83, | |
| "learning_rate": 3.780617678381257e-06, | |
| "loss": 0.0081, | |
| "step": 4340 | |
| }, | |
| { | |
| "epoch": 13.86, | |
| "learning_rate": 3.6741214057507987e-06, | |
| "loss": 0.0154, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 13.9, | |
| "learning_rate": 3.5676251331203413e-06, | |
| "loss": 0.0159, | |
| "step": 4360 | |
| }, | |
| { | |
| "epoch": 13.93, | |
| "learning_rate": 3.461128860489883e-06, | |
| "loss": 0.0128, | |
| "step": 4370 | |
| }, | |
| { | |
| "epoch": 13.96, | |
| "learning_rate": 3.354632587859425e-06, | |
| "loss": 0.0144, | |
| "step": 4380 | |
| }, | |
| { | |
| "epoch": 13.99, | |
| "learning_rate": 3.2481363152289674e-06, | |
| "loss": 0.0225, | |
| "step": 4390 | |
| }, | |
| { | |
| "epoch": 14.0, | |
| "eval_gen_len": 12.627240143369175, | |
| "eval_loss": 0.011567563749849796, | |
| "eval_rouge1": 63.5224, | |
| "eval_rouge2": 59.4962, | |
| "eval_rougeL": 63.22, | |
| "eval_rougeLsum": 63.4236, | |
| "eval_runtime": 615.3823, | |
| "eval_samples_per_second": 0.453, | |
| "eval_steps_per_second": 0.228, | |
| "step": 4392 | |
| }, | |
| { | |
| "epoch": 14.02, | |
| "learning_rate": 3.141640042598509e-06, | |
| "loss": 0.0107, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 14.06, | |
| "learning_rate": 3.0351437699680513e-06, | |
| "loss": 0.0207, | |
| "step": 4410 | |
| }, | |
| { | |
| "epoch": 14.09, | |
| "learning_rate": 2.9286474973375935e-06, | |
| "loss": 0.0132, | |
| "step": 4420 | |
| }, | |
| { | |
| "epoch": 14.12, | |
| "learning_rate": 2.8221512247071357e-06, | |
| "loss": 0.0142, | |
| "step": 4430 | |
| }, | |
| { | |
| "epoch": 14.15, | |
| "learning_rate": 2.7156549520766774e-06, | |
| "loss": 0.0139, | |
| "step": 4440 | |
| }, | |
| { | |
| "epoch": 14.18, | |
| "learning_rate": 2.6091586794462196e-06, | |
| "loss": 0.0127, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 14.22, | |
| "learning_rate": 2.5026624068157617e-06, | |
| "loss": 0.0136, | |
| "step": 4460 | |
| }, | |
| { | |
| "epoch": 14.25, | |
| "learning_rate": 2.3961661341853035e-06, | |
| "loss": 0.0086, | |
| "step": 4470 | |
| }, | |
| { | |
| "epoch": 14.28, | |
| "learning_rate": 2.2896698615548457e-06, | |
| "loss": 0.0104, | |
| "step": 4480 | |
| }, | |
| { | |
| "epoch": 14.31, | |
| "learning_rate": 2.183173588924388e-06, | |
| "loss": 0.0205, | |
| "step": 4490 | |
| }, | |
| { | |
| "epoch": 14.34, | |
| "learning_rate": 2.0766773162939296e-06, | |
| "loss": 0.0158, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 14.37, | |
| "learning_rate": 1.9701810436634718e-06, | |
| "loss": 0.0114, | |
| "step": 4510 | |
| }, | |
| { | |
| "epoch": 14.41, | |
| "learning_rate": 1.8636847710330141e-06, | |
| "loss": 0.0296, | |
| "step": 4520 | |
| }, | |
| { | |
| "epoch": 14.44, | |
| "learning_rate": 1.7571884984025559e-06, | |
| "loss": 0.021, | |
| "step": 4530 | |
| }, | |
| { | |
| "epoch": 14.47, | |
| "learning_rate": 1.650692225772098e-06, | |
| "loss": 0.0123, | |
| "step": 4540 | |
| }, | |
| { | |
| "epoch": 14.5, | |
| "learning_rate": 1.54419595314164e-06, | |
| "loss": 0.0178, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 14.53, | |
| "learning_rate": 1.4376996805111822e-06, | |
| "loss": 0.0116, | |
| "step": 4560 | |
| }, | |
| { | |
| "epoch": 14.57, | |
| "learning_rate": 1.3312034078807244e-06, | |
| "loss": 0.0101, | |
| "step": 4570 | |
| }, | |
| { | |
| "epoch": 14.6, | |
| "learning_rate": 1.2247071352502663e-06, | |
| "loss": 0.0166, | |
| "step": 4580 | |
| }, | |
| { | |
| "epoch": 14.63, | |
| "learning_rate": 1.1182108626198083e-06, | |
| "loss": 0.0136, | |
| "step": 4590 | |
| }, | |
| { | |
| "epoch": 14.66, | |
| "learning_rate": 1.0117145899893504e-06, | |
| "loss": 0.0161, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 14.69, | |
| "learning_rate": 9.052183173588925e-07, | |
| "loss": 0.0082, | |
| "step": 4610 | |
| }, | |
| { | |
| "epoch": 14.73, | |
| "learning_rate": 7.987220447284345e-07, | |
| "loss": 0.0159, | |
| "step": 4620 | |
| }, | |
| { | |
| "epoch": 14.76, | |
| "learning_rate": 6.922257720979766e-07, | |
| "loss": 0.0093, | |
| "step": 4630 | |
| }, | |
| { | |
| "epoch": 14.79, | |
| "learning_rate": 5.857294994675187e-07, | |
| "loss": 0.0096, | |
| "step": 4640 | |
| }, | |
| { | |
| "epoch": 14.82, | |
| "learning_rate": 4.792332268370607e-07, | |
| "loss": 0.0135, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 14.85, | |
| "learning_rate": 3.727369542066028e-07, | |
| "loss": 0.0123, | |
| "step": 4660 | |
| }, | |
| { | |
| "epoch": 14.88, | |
| "learning_rate": 2.6624068157614484e-07, | |
| "loss": 0.007, | |
| "step": 4670 | |
| }, | |
| { | |
| "epoch": 14.92, | |
| "learning_rate": 1.597444089456869e-07, | |
| "loss": 0.0184, | |
| "step": 4680 | |
| }, | |
| { | |
| "epoch": 14.95, | |
| "learning_rate": 5.3248136315228974e-08, | |
| "loss": 0.0204, | |
| "step": 4690 | |
| }, | |
| { | |
| "epoch": 14.96, | |
| "eval_gen_len": 12.577060931899641, | |
| "eval_loss": 0.01156473346054554, | |
| "eval_rouge1": 63.2308, | |
| "eval_rouge2": 59.1657, | |
| "eval_rougeL": 62.9973, | |
| "eval_rougeLsum": 63.1556, | |
| "eval_runtime": 615.4381, | |
| "eval_samples_per_second": 0.453, | |
| "eval_steps_per_second": 0.227, | |
| "step": 4695 | |
| } | |
| ], | |
| "max_steps": 4695, | |
| "num_train_epochs": 15, | |
| "total_flos": 3.211052520088535e+17, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |