| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.0, | |
| "eval_steps": 500, | |
| "global_step": 5000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "grad_norm": 2.7266480922698975, | |
| "learning_rate": 1.8e-06, | |
| "loss": 1.238, | |
| "step": 10 | |
| }, | |
| { | |
| "grad_norm": 1.3996137380599976, | |
| "learning_rate": 3.8e-06, | |
| "loss": 1.2263, | |
| "step": 20 | |
| }, | |
| { | |
| "grad_norm": 0.6298696398735046, | |
| "learning_rate": 5.8e-06, | |
| "loss": 1.1558, | |
| "step": 30 | |
| }, | |
| { | |
| "grad_norm": 0.4681908190250397, | |
| "learning_rate": 7.8e-06, | |
| "loss": 1.1569, | |
| "step": 40 | |
| }, | |
| { | |
| "grad_norm": 0.9008304476737976, | |
| "learning_rate": 9.800000000000001e-06, | |
| "loss": 1.1582, | |
| "step": 50 | |
| }, | |
| { | |
| "grad_norm": 0.8695996999740601, | |
| "learning_rate": 1.18e-05, | |
| "loss": 1.1469, | |
| "step": 60 | |
| }, | |
| { | |
| "grad_norm": 0.754650890827179, | |
| "learning_rate": 1.3800000000000002e-05, | |
| "loss": 1.0852, | |
| "step": 70 | |
| }, | |
| { | |
| "grad_norm": 0.7990859746932983, | |
| "learning_rate": 1.58e-05, | |
| "loss": 1.0661, | |
| "step": 80 | |
| }, | |
| { | |
| "grad_norm": 0.6180333495140076, | |
| "learning_rate": 1.78e-05, | |
| "loss": 1.0627, | |
| "step": 90 | |
| }, | |
| { | |
| "grad_norm": 0.46167218685150146, | |
| "learning_rate": 1.9800000000000004e-05, | |
| "loss": 1.053, | |
| "step": 100 | |
| }, | |
| { | |
| "grad_norm": 0.5125234723091125, | |
| "learning_rate": 2.18e-05, | |
| "loss": 1.0584, | |
| "step": 110 | |
| }, | |
| { | |
| "grad_norm": 0.615847647190094, | |
| "learning_rate": 2.38e-05, | |
| "loss": 1.0547, | |
| "step": 120 | |
| }, | |
| { | |
| "grad_norm": 0.7808847427368164, | |
| "learning_rate": 2.58e-05, | |
| "loss": 1.0613, | |
| "step": 130 | |
| }, | |
| { | |
| "grad_norm": 0.6452605128288269, | |
| "learning_rate": 2.7800000000000005e-05, | |
| "loss": 1.0389, | |
| "step": 140 | |
| }, | |
| { | |
| "grad_norm": 0.7127711176872253, | |
| "learning_rate": 2.98e-05, | |
| "loss": 1.0285, | |
| "step": 150 | |
| }, | |
| { | |
| "grad_norm": 1.9395393133163452, | |
| "learning_rate": 3.18e-05, | |
| "loss": 0.9987, | |
| "step": 160 | |
| }, | |
| { | |
| "grad_norm": 2.0809590816497803, | |
| "learning_rate": 3.38e-05, | |
| "loss": 0.9503, | |
| "step": 170 | |
| }, | |
| { | |
| "grad_norm": 1.116188406944275, | |
| "learning_rate": 3.58e-05, | |
| "loss": 0.8721, | |
| "step": 180 | |
| }, | |
| { | |
| "grad_norm": 1.0570509433746338, | |
| "learning_rate": 3.7800000000000004e-05, | |
| "loss": 0.8447, | |
| "step": 190 | |
| }, | |
| { | |
| "grad_norm": 2.021897077560425, | |
| "learning_rate": 3.9800000000000005e-05, | |
| "loss": 0.7972, | |
| "step": 200 | |
| }, | |
| { | |
| "grad_norm": 1.9479414224624634, | |
| "learning_rate": 4.18e-05, | |
| "loss": 0.7084, | |
| "step": 210 | |
| }, | |
| { | |
| "grad_norm": 2.420440912246704, | |
| "learning_rate": 4.38e-05, | |
| "loss": 0.638, | |
| "step": 220 | |
| }, | |
| { | |
| "grad_norm": 1.9939135313034058, | |
| "learning_rate": 4.58e-05, | |
| "loss": 0.5441, | |
| "step": 230 | |
| }, | |
| { | |
| "grad_norm": 1.766343116760254, | |
| "learning_rate": 4.78e-05, | |
| "loss": 0.4735, | |
| "step": 240 | |
| }, | |
| { | |
| "grad_norm": 2.4787240028381348, | |
| "learning_rate": 4.9800000000000004e-05, | |
| "loss": 0.3976, | |
| "step": 250 | |
| }, | |
| { | |
| "grad_norm": 1.8052730560302734, | |
| "learning_rate": 4.9999557099392795e-05, | |
| "loss": 0.2773, | |
| "step": 260 | |
| }, | |
| { | |
| "grad_norm": 1.487770915031433, | |
| "learning_rate": 4.9998026105095405e-05, | |
| "loss": 0.2167, | |
| "step": 270 | |
| }, | |
| { | |
| "grad_norm": 1.126673936843872, | |
| "learning_rate": 4.9995401616153804e-05, | |
| "loss": 0.1938, | |
| "step": 280 | |
| }, | |
| { | |
| "grad_norm": 1.2377407550811768, | |
| "learning_rate": 4.9991683747371645e-05, | |
| "loss": 0.1878, | |
| "step": 290 | |
| }, | |
| { | |
| "grad_norm": 2.0436601638793945, | |
| "learning_rate": 4.9986872661380534e-05, | |
| "loss": 0.1571, | |
| "step": 300 | |
| }, | |
| { | |
| "grad_norm": 1.2251367568969727, | |
| "learning_rate": 4.998096856863298e-05, | |
| "loss": 0.1595, | |
| "step": 310 | |
| }, | |
| { | |
| "grad_norm": 1.2823742628097534, | |
| "learning_rate": 4.997397172739312e-05, | |
| "loss": 0.1464, | |
| "step": 320 | |
| }, | |
| { | |
| "grad_norm": 1.0287004709243774, | |
| "learning_rate": 4.996588244372545e-05, | |
| "loss": 0.1295, | |
| "step": 330 | |
| }, | |
| { | |
| "grad_norm": 1.0611475706100464, | |
| "learning_rate": 4.995670107148146e-05, | |
| "loss": 0.1309, | |
| "step": 340 | |
| }, | |
| { | |
| "grad_norm": 1.0434561967849731, | |
| "learning_rate": 4.9946428012284095e-05, | |
| "loss": 0.1344, | |
| "step": 350 | |
| }, | |
| { | |
| "grad_norm": 1.203682780265808, | |
| "learning_rate": 4.993506371551025e-05, | |
| "loss": 0.1339, | |
| "step": 360 | |
| }, | |
| { | |
| "grad_norm": 0.7937515377998352, | |
| "learning_rate": 4.992260867827109e-05, | |
| "loss": 0.1247, | |
| "step": 370 | |
| }, | |
| { | |
| "grad_norm": 0.9882832169532776, | |
| "learning_rate": 4.9909063445390285e-05, | |
| "loss": 0.1221, | |
| "step": 380 | |
| }, | |
| { | |
| "grad_norm": 1.3059258460998535, | |
| "learning_rate": 4.9894428609380206e-05, | |
| "loss": 0.1215, | |
| "step": 390 | |
| }, | |
| { | |
| "grad_norm": 1.3231853246688843, | |
| "learning_rate": 4.987870481041599e-05, | |
| "loss": 0.1257, | |
| "step": 400 | |
| }, | |
| { | |
| "grad_norm": 0.7158247828483582, | |
| "learning_rate": 4.986189273630752e-05, | |
| "loss": 0.1243, | |
| "step": 410 | |
| }, | |
| { | |
| "grad_norm": 1.0682462453842163, | |
| "learning_rate": 4.9843993122469424e-05, | |
| "loss": 0.1131, | |
| "step": 420 | |
| }, | |
| { | |
| "grad_norm": 0.914338231086731, | |
| "learning_rate": 4.9825006751888767e-05, | |
| "loss": 0.1135, | |
| "step": 430 | |
| }, | |
| { | |
| "grad_norm": 0.8019906878471375, | |
| "learning_rate": 4.9804934455090915e-05, | |
| "loss": 0.1135, | |
| "step": 440 | |
| }, | |
| { | |
| "grad_norm": 0.878158450126648, | |
| "learning_rate": 4.978377711010315e-05, | |
| "loss": 0.1176, | |
| "step": 450 | |
| }, | |
| { | |
| "grad_norm": 1.1396379470825195, | |
| "learning_rate": 4.976153564241628e-05, | |
| "loss": 0.1077, | |
| "step": 460 | |
| }, | |
| { | |
| "grad_norm": 1.298904538154602, | |
| "learning_rate": 4.973821102494417e-05, | |
| "loss": 0.1208, | |
| "step": 470 | |
| }, | |
| { | |
| "grad_norm": 0.8515462279319763, | |
| "learning_rate": 4.971380427798113e-05, | |
| "loss": 0.1024, | |
| "step": 480 | |
| }, | |
| { | |
| "grad_norm": 0.8240450620651245, | |
| "learning_rate": 4.9688316469157354e-05, | |
| "loss": 0.1018, | |
| "step": 490 | |
| }, | |
| { | |
| "grad_norm": 0.6559143662452698, | |
| "learning_rate": 4.9661748713392165e-05, | |
| "loss": 0.0972, | |
| "step": 500 | |
| }, | |
| { | |
| "grad_norm": 0.9851711392402649, | |
| "learning_rate": 4.963410217284526e-05, | |
| "loss": 0.113, | |
| "step": 510 | |
| }, | |
| { | |
| "grad_norm": 0.8339484333992004, | |
| "learning_rate": 4.960537805686589e-05, | |
| "loss": 0.0973, | |
| "step": 520 | |
| }, | |
| { | |
| "grad_norm": 0.8457871675491333, | |
| "learning_rate": 4.957557762193994e-05, | |
| "loss": 0.101, | |
| "step": 530 | |
| }, | |
| { | |
| "grad_norm": 0.9261676669120789, | |
| "learning_rate": 4.954470217163498e-05, | |
| "loss": 0.0894, | |
| "step": 540 | |
| }, | |
| { | |
| "grad_norm": 0.8262922763824463, | |
| "learning_rate": 4.9512753056543226e-05, | |
| "loss": 0.0825, | |
| "step": 550 | |
| }, | |
| { | |
| "grad_norm": 0.6502310633659363, | |
| "learning_rate": 4.947973167422247e-05, | |
| "loss": 0.0876, | |
| "step": 560 | |
| }, | |
| { | |
| "grad_norm": 0.7397481203079224, | |
| "learning_rate": 4.9445639469134944e-05, | |
| "loss": 0.0768, | |
| "step": 570 | |
| }, | |
| { | |
| "grad_norm": 0.7882953882217407, | |
| "learning_rate": 4.941047793258416e-05, | |
| "loss": 0.0969, | |
| "step": 580 | |
| }, | |
| { | |
| "grad_norm": 0.79180508852005, | |
| "learning_rate": 4.937424860264961e-05, | |
| "loss": 0.108, | |
| "step": 590 | |
| }, | |
| { | |
| "grad_norm": 0.7894780039787292, | |
| "learning_rate": 4.933695306411957e-05, | |
| "loss": 0.097, | |
| "step": 600 | |
| }, | |
| { | |
| "grad_norm": 0.8765780925750732, | |
| "learning_rate": 4.929859294842172e-05, | |
| "loss": 0.1013, | |
| "step": 610 | |
| }, | |
| { | |
| "grad_norm": 0.9128709435462952, | |
| "learning_rate": 4.9259169933551765e-05, | |
| "loss": 0.0854, | |
| "step": 620 | |
| }, | |
| { | |
| "grad_norm": 0.6108020544052124, | |
| "learning_rate": 4.921868574400011e-05, | |
| "loss": 0.0809, | |
| "step": 630 | |
| }, | |
| { | |
| "grad_norm": 0.7296820878982544, | |
| "learning_rate": 4.9177142150676355e-05, | |
| "loss": 0.0858, | |
| "step": 640 | |
| }, | |
| { | |
| "grad_norm": 0.8388088345527649, | |
| "learning_rate": 4.913454097083185e-05, | |
| "loss": 0.0974, | |
| "step": 650 | |
| }, | |
| { | |
| "grad_norm": 0.920002818107605, | |
| "learning_rate": 4.9090884067980205e-05, | |
| "loss": 0.0857, | |
| "step": 660 | |
| }, | |
| { | |
| "grad_norm": 0.9557753205299377, | |
| "learning_rate": 4.9046173351815796e-05, | |
| "loss": 0.1021, | |
| "step": 670 | |
| }, | |
| { | |
| "grad_norm": 1.0099128484725952, | |
| "learning_rate": 4.900041077813017e-05, | |
| "loss": 0.0937, | |
| "step": 680 | |
| }, | |
| { | |
| "grad_norm": 0.7957805395126343, | |
| "learning_rate": 4.895359834872656e-05, | |
| "loss": 0.0919, | |
| "step": 690 | |
| }, | |
| { | |
| "grad_norm": 0.5897656083106995, | |
| "learning_rate": 4.8905738111332275e-05, | |
| "loss": 0.0868, | |
| "step": 700 | |
| }, | |
| { | |
| "grad_norm": 0.82470703125, | |
| "learning_rate": 4.8856832159509156e-05, | |
| "loss": 0.0897, | |
| "step": 710 | |
| }, | |
| { | |
| "grad_norm": 0.5583815574645996, | |
| "learning_rate": 4.880688263256197e-05, | |
| "loss": 0.0831, | |
| "step": 720 | |
| }, | |
| { | |
| "grad_norm": 0.5602760910987854, | |
| "learning_rate": 4.8755891715444814e-05, | |
| "loss": 0.0864, | |
| "step": 730 | |
| }, | |
| { | |
| "grad_norm": 0.6645472645759583, | |
| "learning_rate": 4.870386163866561e-05, | |
| "loss": 0.0785, | |
| "step": 740 | |
| }, | |
| { | |
| "grad_norm": 0.4997585117816925, | |
| "learning_rate": 4.8650794678188485e-05, | |
| "loss": 0.0736, | |
| "step": 750 | |
| }, | |
| { | |
| "grad_norm": 0.6175997853279114, | |
| "learning_rate": 4.859669315533417e-05, | |
| "loss": 0.0802, | |
| "step": 760 | |
| }, | |
| { | |
| "grad_norm": 0.6164193749427795, | |
| "learning_rate": 4.8541559436678563e-05, | |
| "loss": 0.0741, | |
| "step": 770 | |
| }, | |
| { | |
| "grad_norm": 0.692690908908844, | |
| "learning_rate": 4.8485395933949114e-05, | |
| "loss": 0.0803, | |
| "step": 780 | |
| }, | |
| { | |
| "grad_norm": 0.7447015047073364, | |
| "learning_rate": 4.842820510391938e-05, | |
| "loss": 0.0747, | |
| "step": 790 | |
| }, | |
| { | |
| "grad_norm": 0.9005197882652283, | |
| "learning_rate": 4.836998944830155e-05, | |
| "loss": 0.0859, | |
| "step": 800 | |
| }, | |
| { | |
| "grad_norm": 0.6014283895492554, | |
| "learning_rate": 4.8310751513636975e-05, | |
| "loss": 0.0847, | |
| "step": 810 | |
| }, | |
| { | |
| "grad_norm": 0.5599387884140015, | |
| "learning_rate": 4.825049389118484e-05, | |
| "loss": 0.0709, | |
| "step": 820 | |
| }, | |
| { | |
| "grad_norm": 0.6468709111213684, | |
| "learning_rate": 4.818921921680874e-05, | |
| "loss": 0.0887, | |
| "step": 830 | |
| }, | |
| { | |
| "grad_norm": 0.6334129571914673, | |
| "learning_rate": 4.812693017086145e-05, | |
| "loss": 0.0689, | |
| "step": 840 | |
| }, | |
| { | |
| "grad_norm": 0.6896155476570129, | |
| "learning_rate": 4.806362947806763e-05, | |
| "loss": 0.0839, | |
| "step": 850 | |
| }, | |
| { | |
| "grad_norm": 0.8341414928436279, | |
| "learning_rate": 4.799931990740463e-05, | |
| "loss": 0.0785, | |
| "step": 860 | |
| }, | |
| { | |
| "grad_norm": 0.5337378978729248, | |
| "learning_rate": 4.7934004271981415e-05, | |
| "loss": 0.0769, | |
| "step": 870 | |
| }, | |
| { | |
| "grad_norm": 0.7976155877113342, | |
| "learning_rate": 4.786768542891548e-05, | |
| "loss": 0.0738, | |
| "step": 880 | |
| }, | |
| { | |
| "grad_norm": 0.3165757358074188, | |
| "learning_rate": 4.7800366279207854e-05, | |
| "loss": 0.0713, | |
| "step": 890 | |
| }, | |
| { | |
| "grad_norm": 0.7630752325057983, | |
| "learning_rate": 4.773204976761623e-05, | |
| "loss": 0.0719, | |
| "step": 900 | |
| }, | |
| { | |
| "grad_norm": 0.6145244836807251, | |
| "learning_rate": 4.7662738882526146e-05, | |
| "loss": 0.0644, | |
| "step": 910 | |
| }, | |
| { | |
| "grad_norm": 0.6122027635574341, | |
| "learning_rate": 4.759243665582024e-05, | |
| "loss": 0.0728, | |
| "step": 920 | |
| }, | |
| { | |
| "grad_norm": 0.4458109736442566, | |
| "learning_rate": 4.752114616274567e-05, | |
| "loss": 0.0678, | |
| "step": 930 | |
| }, | |
| { | |
| "grad_norm": 0.6862826347351074, | |
| "learning_rate": 4.744887052177954e-05, | |
| "loss": 0.0694, | |
| "step": 940 | |
| }, | |
| { | |
| "grad_norm": 0.5472699403762817, | |
| "learning_rate": 4.7375612894492536e-05, | |
| "loss": 0.0739, | |
| "step": 950 | |
| }, | |
| { | |
| "grad_norm": 0.6469587087631226, | |
| "learning_rate": 4.7301376485410596e-05, | |
| "loss": 0.0612, | |
| "step": 960 | |
| }, | |
| { | |
| "grad_norm": 0.5633054971694946, | |
| "learning_rate": 4.722616454187474e-05, | |
| "loss": 0.0752, | |
| "step": 970 | |
| }, | |
| { | |
| "grad_norm": 0.611199140548706, | |
| "learning_rate": 4.714998035389904e-05, | |
| "loss": 0.0682, | |
| "step": 980 | |
| }, | |
| { | |
| "grad_norm": 0.6113159656524658, | |
| "learning_rate": 4.7072827254026666e-05, | |
| "loss": 0.0734, | |
| "step": 990 | |
| }, | |
| { | |
| "grad_norm": 0.5616899728775024, | |
| "learning_rate": 4.6994708617184155e-05, | |
| "loss": 0.0647, | |
| "step": 1000 | |
| }, | |
| { | |
| "grad_norm": 0.5405145883560181, | |
| "learning_rate": 4.691562786053376e-05, | |
| "loss": 0.0652, | |
| "step": 1010 | |
| }, | |
| { | |
| "grad_norm": 0.5287688374519348, | |
| "learning_rate": 4.6835588443323956e-05, | |
| "loss": 0.071, | |
| "step": 1020 | |
| }, | |
| { | |
| "grad_norm": 0.6897308230400085, | |
| "learning_rate": 4.675459386673815e-05, | |
| "loss": 0.0711, | |
| "step": 1030 | |
| }, | |
| { | |
| "grad_norm": 0.43904393911361694, | |
| "learning_rate": 4.6672647673741485e-05, | |
| "loss": 0.0762, | |
| "step": 1040 | |
| }, | |
| { | |
| "grad_norm": 0.502697765827179, | |
| "learning_rate": 4.658975344892594e-05, | |
| "loss": 0.0892, | |
| "step": 1050 | |
| }, | |
| { | |
| "grad_norm": 0.718237042427063, | |
| "learning_rate": 4.650591481835344e-05, | |
| "loss": 0.0862, | |
| "step": 1060 | |
| }, | |
| { | |
| "grad_norm": 0.570436954498291, | |
| "learning_rate": 4.642113544939728e-05, | |
| "loss": 0.07, | |
| "step": 1070 | |
| }, | |
| { | |
| "grad_norm": 0.4335796535015106, | |
| "learning_rate": 4.63354190505817e-05, | |
| "loss": 0.0811, | |
| "step": 1080 | |
| }, | |
| { | |
| "grad_norm": 0.6726382374763489, | |
| "learning_rate": 4.6248769371419684e-05, | |
| "loss": 0.0675, | |
| "step": 1090 | |
| }, | |
| { | |
| "grad_norm": 0.5062233805656433, | |
| "learning_rate": 4.6161190202248897e-05, | |
| "loss": 0.064, | |
| "step": 1100 | |
| }, | |
| { | |
| "grad_norm": 0.3789091408252716, | |
| "learning_rate": 4.607268537406591e-05, | |
| "loss": 0.0706, | |
| "step": 1110 | |
| }, | |
| { | |
| "grad_norm": 0.5326701402664185, | |
| "learning_rate": 4.598325875835862e-05, | |
| "loss": 0.071, | |
| "step": 1120 | |
| }, | |
| { | |
| "grad_norm": 0.5220073461532593, | |
| "learning_rate": 4.589291426693692e-05, | |
| "loss": 0.0727, | |
| "step": 1130 | |
| }, | |
| { | |
| "grad_norm": 0.5300417542457581, | |
| "learning_rate": 4.580165585176152e-05, | |
| "loss": 0.0678, | |
| "step": 1140 | |
| }, | |
| { | |
| "grad_norm": 0.524254560470581, | |
| "learning_rate": 4.570948750477114e-05, | |
| "loss": 0.0652, | |
| "step": 1150 | |
| }, | |
| { | |
| "grad_norm": 0.5178232192993164, | |
| "learning_rate": 4.561641325770788e-05, | |
| "loss": 0.0636, | |
| "step": 1160 | |
| }, | |
| { | |
| "grad_norm": 0.35104331374168396, | |
| "learning_rate": 4.5522437181940804e-05, | |
| "loss": 0.0624, | |
| "step": 1170 | |
| }, | |
| { | |
| "grad_norm": 0.5118323564529419, | |
| "learning_rate": 4.542756338828791e-05, | |
| "loss": 0.0622, | |
| "step": 1180 | |
| }, | |
| { | |
| "grad_norm": 0.4421655535697937, | |
| "learning_rate": 4.533179602683629e-05, | |
| "loss": 0.0629, | |
| "step": 1190 | |
| }, | |
| { | |
| "grad_norm": 0.7161966562271118, | |
| "learning_rate": 4.523513928676056e-05, | |
| "loss": 0.0599, | |
| "step": 1200 | |
| }, | |
| { | |
| "grad_norm": 0.4908861219882965, | |
| "learning_rate": 4.5137597396139676e-05, | |
| "loss": 0.0653, | |
| "step": 1210 | |
| }, | |
| { | |
| "grad_norm": 1.035172939300537, | |
| "learning_rate": 4.503917462177192e-05, | |
| "loss": 0.0592, | |
| "step": 1220 | |
| }, | |
| { | |
| "grad_norm": 0.6242174506187439, | |
| "learning_rate": 4.4939875268988273e-05, | |
| "loss": 0.0679, | |
| "step": 1230 | |
| }, | |
| { | |
| "grad_norm": 0.34626179933547974, | |
| "learning_rate": 4.483970368146413e-05, | |
| "loss": 0.0608, | |
| "step": 1240 | |
| }, | |
| { | |
| "grad_norm": 0.41088080406188965, | |
| "learning_rate": 4.473866424102923e-05, | |
| "loss": 0.0715, | |
| "step": 1250 | |
| }, | |
| { | |
| "grad_norm": 0.4913835823535919, | |
| "learning_rate": 4.463676136747602e-05, | |
| "loss": 0.0519, | |
| "step": 1260 | |
| }, | |
| { | |
| "grad_norm": 0.39666661620140076, | |
| "learning_rate": 4.453399951836632e-05, | |
| "loss": 0.0559, | |
| "step": 1270 | |
| }, | |
| { | |
| "grad_norm": 0.5405796766281128, | |
| "learning_rate": 4.44303831888363e-05, | |
| "loss": 0.0645, | |
| "step": 1280 | |
| }, | |
| { | |
| "grad_norm": 0.6450210809707642, | |
| "learning_rate": 4.432591691139989e-05, | |
| "loss": 0.0692, | |
| "step": 1290 | |
| }, | |
| { | |
| "grad_norm": 0.6847130656242371, | |
| "learning_rate": 4.422060525575048e-05, | |
| "loss": 0.0665, | |
| "step": 1300 | |
| }, | |
| { | |
| "grad_norm": 0.5179506540298462, | |
| "learning_rate": 4.411445282856106e-05, | |
| "loss": 0.0612, | |
| "step": 1310 | |
| }, | |
| { | |
| "grad_norm": 0.76016765832901, | |
| "learning_rate": 4.400746427328268e-05, | |
| "loss": 0.0618, | |
| "step": 1320 | |
| }, | |
| { | |
| "grad_norm": 0.522050678730011, | |
| "learning_rate": 4.389964426994134e-05, | |
| "loss": 0.0632, | |
| "step": 1330 | |
| }, | |
| { | |
| "grad_norm": 0.5314463973045349, | |
| "learning_rate": 4.3790997534933276e-05, | |
| "loss": 0.0628, | |
| "step": 1340 | |
| }, | |
| { | |
| "grad_norm": 0.4170549213886261, | |
| "learning_rate": 4.368152882081865e-05, | |
| "loss": 0.0595, | |
| "step": 1350 | |
| }, | |
| { | |
| "grad_norm": 0.5747459530830383, | |
| "learning_rate": 4.357124291611363e-05, | |
| "loss": 0.0549, | |
| "step": 1360 | |
| }, | |
| { | |
| "grad_norm": 0.5837365388870239, | |
| "learning_rate": 4.346014464508098e-05, | |
| "loss": 0.0519, | |
| "step": 1370 | |
| }, | |
| { | |
| "grad_norm": 0.35218408703804016, | |
| "learning_rate": 4.3348238867518985e-05, | |
| "loss": 0.06, | |
| "step": 1380 | |
| }, | |
| { | |
| "grad_norm": 0.40569043159484863, | |
| "learning_rate": 4.323553047854887e-05, | |
| "loss": 0.0484, | |
| "step": 1390 | |
| }, | |
| { | |
| "grad_norm": 0.5539340972900391, | |
| "learning_rate": 4.3122024408400693e-05, | |
| "loss": 0.0496, | |
| "step": 1400 | |
| }, | |
| { | |
| "grad_norm": 0.33733201026916504, | |
| "learning_rate": 4.3007725622197674e-05, | |
| "loss": 0.0508, | |
| "step": 1410 | |
| }, | |
| { | |
| "grad_norm": 0.5494152903556824, | |
| "learning_rate": 4.2892639119739e-05, | |
| "loss": 0.0621, | |
| "step": 1420 | |
| }, | |
| { | |
| "grad_norm": 0.39040982723236084, | |
| "learning_rate": 4.277676993528112e-05, | |
| "loss": 0.0528, | |
| "step": 1430 | |
| }, | |
| { | |
| "grad_norm": 0.5634658932685852, | |
| "learning_rate": 4.2660123137317523e-05, | |
| "loss": 0.0586, | |
| "step": 1440 | |
| }, | |
| { | |
| "grad_norm": 0.5504993796348572, | |
| "learning_rate": 4.254270382835703e-05, | |
| "loss": 0.0617, | |
| "step": 1450 | |
| }, | |
| { | |
| "grad_norm": 0.3419093191623688, | |
| "learning_rate": 4.2424517144700603e-05, | |
| "loss": 0.0573, | |
| "step": 1460 | |
| }, | |
| { | |
| "grad_norm": 0.3770234286785126, | |
| "learning_rate": 4.230556825621667e-05, | |
| "loss": 0.0552, | |
| "step": 1470 | |
| }, | |
| { | |
| "grad_norm": 0.5088812708854675, | |
| "learning_rate": 4.2185862366114936e-05, | |
| "loss": 0.061, | |
| "step": 1480 | |
| }, | |
| { | |
| "grad_norm": 0.5062453746795654, | |
| "learning_rate": 4.2065404710718834e-05, | |
| "loss": 0.0528, | |
| "step": 1490 | |
| }, | |
| { | |
| "grad_norm": 0.7702981233596802, | |
| "learning_rate": 4.194420055923644e-05, | |
| "loss": 0.0637, | |
| "step": 1500 | |
| }, | |
| { | |
| "grad_norm": 0.6036869287490845, | |
| "learning_rate": 4.182225521352999e-05, | |
| "loss": 0.0702, | |
| "step": 1510 | |
| }, | |
| { | |
| "grad_norm": 0.5008228421211243, | |
| "learning_rate": 4.169957400788395e-05, | |
| "loss": 0.0624, | |
| "step": 1520 | |
| }, | |
| { | |
| "grad_norm": 0.41124141216278076, | |
| "learning_rate": 4.157616230877169e-05, | |
| "loss": 0.0563, | |
| "step": 1530 | |
| }, | |
| { | |
| "grad_norm": 0.7010630369186401, | |
| "learning_rate": 4.145202551462072e-05, | |
| "loss": 0.062, | |
| "step": 1540 | |
| }, | |
| { | |
| "grad_norm": 0.4738198220729828, | |
| "learning_rate": 4.132716905557658e-05, | |
| "loss": 0.0636, | |
| "step": 1550 | |
| }, | |
| { | |
| "grad_norm": 0.5558097958564758, | |
| "learning_rate": 4.1201598393265244e-05, | |
| "loss": 0.0627, | |
| "step": 1560 | |
| }, | |
| { | |
| "grad_norm": 0.4446842074394226, | |
| "learning_rate": 4.1075319020554286e-05, | |
| "loss": 0.0577, | |
| "step": 1570 | |
| }, | |
| { | |
| "grad_norm": 0.476675808429718, | |
| "learning_rate": 4.094833646131256e-05, | |
| "loss": 0.0544, | |
| "step": 1580 | |
| }, | |
| { | |
| "grad_norm": 0.4960286617279053, | |
| "learning_rate": 4.082065627016858e-05, | |
| "loss": 0.0644, | |
| "step": 1590 | |
| }, | |
| { | |
| "grad_norm": 0.46870744228363037, | |
| "learning_rate": 4.0692284032267516e-05, | |
| "loss": 0.054, | |
| "step": 1600 | |
| }, | |
| { | |
| "grad_norm": 0.7339263558387756, | |
| "learning_rate": 4.056322536302693e-05, | |
| "loss": 0.0552, | |
| "step": 1610 | |
| }, | |
| { | |
| "grad_norm": 0.3929803669452667, | |
| "learning_rate": 4.043348590789111e-05, | |
| "loss": 0.0554, | |
| "step": 1620 | |
| }, | |
| { | |
| "grad_norm": 0.4788963198661804, | |
| "learning_rate": 4.030307134208411e-05, | |
| "loss": 0.052, | |
| "step": 1630 | |
| }, | |
| { | |
| "grad_norm": 0.4371449649333954, | |
| "learning_rate": 4.017198737036154e-05, | |
| "loss": 0.0531, | |
| "step": 1640 | |
| }, | |
| { | |
| "grad_norm": 0.4106438159942627, | |
| "learning_rate": 4.004023972676097e-05, | |
| "loss": 0.0569, | |
| "step": 1650 | |
| }, | |
| { | |
| "grad_norm": 0.3421051800251007, | |
| "learning_rate": 3.9907834174351125e-05, | |
| "loss": 0.0509, | |
| "step": 1660 | |
| }, | |
| { | |
| "grad_norm": 0.45882418751716614, | |
| "learning_rate": 3.9774776504979805e-05, | |
| "loss": 0.058, | |
| "step": 1670 | |
| }, | |
| { | |
| "grad_norm": 0.5227372646331787, | |
| "learning_rate": 3.964107253902052e-05, | |
| "loss": 0.0552, | |
| "step": 1680 | |
| }, | |
| { | |
| "grad_norm": 0.567558228969574, | |
| "learning_rate": 3.950672812511788e-05, | |
| "loss": 0.0617, | |
| "step": 1690 | |
| }, | |
| { | |
| "grad_norm": 0.6489816904067993, | |
| "learning_rate": 3.937174913993177e-05, | |
| "loss": 0.0563, | |
| "step": 1700 | |
| }, | |
| { | |
| "grad_norm": 0.4380536377429962, | |
| "learning_rate": 3.9236141487880265e-05, | |
| "loss": 0.0508, | |
| "step": 1710 | |
| }, | |
| { | |
| "grad_norm": 0.5422831177711487, | |
| "learning_rate": 3.909991110088138e-05, | |
| "loss": 0.0541, | |
| "step": 1720 | |
| }, | |
| { | |
| "grad_norm": 0.7260904908180237, | |
| "learning_rate": 3.896306393809357e-05, | |
| "loss": 0.0563, | |
| "step": 1730 | |
| }, | |
| { | |
| "grad_norm": 0.4155614972114563, | |
| "learning_rate": 3.882560598565504e-05, | |
| "loss": 0.0552, | |
| "step": 1740 | |
| }, | |
| { | |
| "grad_norm": 0.35693034529685974, | |
| "learning_rate": 3.8687543256421955e-05, | |
| "loss": 0.0521, | |
| "step": 1750 | |
| }, | |
| { | |
| "grad_norm": 0.591141402721405, | |
| "learning_rate": 3.8548881789705345e-05, | |
| "loss": 0.0571, | |
| "step": 1760 | |
| }, | |
| { | |
| "grad_norm": 0.5429609417915344, | |
| "learning_rate": 3.840962765100696e-05, | |
| "loss": 0.0575, | |
| "step": 1770 | |
| }, | |
| { | |
| "grad_norm": 0.4284355938434601, | |
| "learning_rate": 3.826978693175395e-05, | |
| "loss": 0.0547, | |
| "step": 1780 | |
| }, | |
| { | |
| "grad_norm": 0.720148503780365, | |
| "learning_rate": 3.81293657490324e-05, | |
| "loss": 0.0575, | |
| "step": 1790 | |
| }, | |
| { | |
| "grad_norm": 0.49093666672706604, | |
| "learning_rate": 3.798837024531974e-05, | |
| "loss": 0.0513, | |
| "step": 1800 | |
| }, | |
| { | |
| "grad_norm": 0.5715184211730957, | |
| "learning_rate": 3.7846806588216055e-05, | |
| "loss": 0.0571, | |
| "step": 1810 | |
| }, | |
| { | |
| "grad_norm": 0.45693865418434143, | |
| "learning_rate": 3.770468097017432e-05, | |
| "loss": 0.0481, | |
| "step": 1820 | |
| }, | |
| { | |
| "grad_norm": 0.49179983139038086, | |
| "learning_rate": 3.7561999608229505e-05, | |
| "loss": 0.0528, | |
| "step": 1830 | |
| }, | |
| { | |
| "grad_norm": 0.4709620177745819, | |
| "learning_rate": 3.741876874372659e-05, | |
| "loss": 0.0501, | |
| "step": 1840 | |
| }, | |
| { | |
| "grad_norm": 0.4550098776817322, | |
| "learning_rate": 3.727499464204758e-05, | |
| "loss": 0.0493, | |
| "step": 1850 | |
| }, | |
| { | |
| "grad_norm": 0.4054962396621704, | |
| "learning_rate": 3.7130683592337465e-05, | |
| "loss": 0.0476, | |
| "step": 1860 | |
| }, | |
| { | |
| "grad_norm": 0.541397750377655, | |
| "learning_rate": 3.698584190722906e-05, | |
| "loss": 0.0534, | |
| "step": 1870 | |
| }, | |
| { | |
| "grad_norm": 0.36921900510787964, | |
| "learning_rate": 3.6840475922566886e-05, | |
| "loss": 0.0478, | |
| "step": 1880 | |
| }, | |
| { | |
| "grad_norm": 0.5922922492027283, | |
| "learning_rate": 3.6694591997130026e-05, | |
| "loss": 0.0501, | |
| "step": 1890 | |
| }, | |
| { | |
| "grad_norm": 0.43542882800102234, | |
| "learning_rate": 3.654819651235401e-05, | |
| "loss": 0.0541, | |
| "step": 1900 | |
| }, | |
| { | |
| "grad_norm": 0.48189231753349304, | |
| "learning_rate": 3.640129587205156e-05, | |
| "loss": 0.0521, | |
| "step": 1910 | |
| }, | |
| { | |
| "grad_norm": 0.4395929276943207, | |
| "learning_rate": 3.625389650213259e-05, | |
| "loss": 0.0429, | |
| "step": 1920 | |
| }, | |
| { | |
| "grad_norm": 0.5689699649810791, | |
| "learning_rate": 3.610600485032305e-05, | |
| "loss": 0.0586, | |
| "step": 1930 | |
| }, | |
| { | |
| "grad_norm": 0.5108771920204163, | |
| "learning_rate": 3.5957627385882884e-05, | |
| "loss": 0.0523, | |
| "step": 1940 | |
| }, | |
| { | |
| "grad_norm": 0.5463844537734985, | |
| "learning_rate": 3.580877059932308e-05, | |
| "loss": 0.0449, | |
| "step": 1950 | |
| }, | |
| { | |
| "grad_norm": 0.3828030526638031, | |
| "learning_rate": 3.5659441002121694e-05, | |
| "loss": 0.0509, | |
| "step": 1960 | |
| }, | |
| { | |
| "grad_norm": 0.3314559757709503, | |
| "learning_rate": 3.550964512643908e-05, | |
| "loss": 0.0435, | |
| "step": 1970 | |
| }, | |
| { | |
| "grad_norm": 0.8163153529167175, | |
| "learning_rate": 3.535938952483211e-05, | |
| "loss": 0.0558, | |
| "step": 1980 | |
| }, | |
| { | |
| "grad_norm": 0.7200665473937988, | |
| "learning_rate": 3.520868076996755e-05, | |
| "loss": 0.0432, | |
| "step": 1990 | |
| }, | |
| { | |
| "grad_norm": 0.47182565927505493, | |
| "learning_rate": 3.5057525454334564e-05, | |
| "loss": 0.0455, | |
| "step": 2000 | |
| }, | |
| { | |
| "grad_norm": 0.5063982009887695, | |
| "learning_rate": 3.490593018995636e-05, | |
| "loss": 0.0446, | |
| "step": 2010 | |
| }, | |
| { | |
| "grad_norm": 0.44255322217941284, | |
| "learning_rate": 3.475390160810087e-05, | |
| "loss": 0.0434, | |
| "step": 2020 | |
| }, | |
| { | |
| "grad_norm": 0.4013705849647522, | |
| "learning_rate": 3.460144635899079e-05, | |
| "loss": 0.0501, | |
| "step": 2030 | |
| }, | |
| { | |
| "grad_norm": 0.5221493244171143, | |
| "learning_rate": 3.444857111151259e-05, | |
| "loss": 0.0547, | |
| "step": 2040 | |
| }, | |
| { | |
| "grad_norm": 0.4845060110092163, | |
| "learning_rate": 3.429528255292485e-05, | |
| "loss": 0.0497, | |
| "step": 2050 | |
| }, | |
| { | |
| "grad_norm": 0.4617994725704193, | |
| "learning_rate": 3.41415873885657e-05, | |
| "loss": 0.0489, | |
| "step": 2060 | |
| }, | |
| { | |
| "grad_norm": 0.4810671806335449, | |
| "learning_rate": 3.3987492341559535e-05, | |
| "loss": 0.0478, | |
| "step": 2070 | |
| }, | |
| { | |
| "grad_norm": 0.5026255249977112, | |
| "learning_rate": 3.383300415252292e-05, | |
| "loss": 0.0446, | |
| "step": 2080 | |
| }, | |
| { | |
| "grad_norm": 0.477425754070282, | |
| "learning_rate": 3.367812957926971e-05, | |
| "loss": 0.0517, | |
| "step": 2090 | |
| }, | |
| { | |
| "grad_norm": 0.3495548367500305, | |
| "learning_rate": 3.352287539651545e-05, | |
| "loss": 0.0456, | |
| "step": 2100 | |
| }, | |
| { | |
| "grad_norm": 0.45665979385375977, | |
| "learning_rate": 3.3367248395581076e-05, | |
| "loss": 0.046, | |
| "step": 2110 | |
| }, | |
| { | |
| "grad_norm": 0.41686445474624634, | |
| "learning_rate": 3.321125538409574e-05, | |
| "loss": 0.0486, | |
| "step": 2120 | |
| }, | |
| { | |
| "grad_norm": 0.3617282211780548, | |
| "learning_rate": 3.3054903185699135e-05, | |
| "loss": 0.0476, | |
| "step": 2130 | |
| }, | |
| { | |
| "grad_norm": 0.47553274035453796, | |
| "learning_rate": 3.2898198639742915e-05, | |
| "loss": 0.0432, | |
| "step": 2140 | |
| }, | |
| { | |
| "grad_norm": 0.3067774176597595, | |
| "learning_rate": 3.2741148600991576e-05, | |
| "loss": 0.0522, | |
| "step": 2150 | |
| }, | |
| { | |
| "grad_norm": 0.5735144019126892, | |
| "learning_rate": 3.258375993932259e-05, | |
| "loss": 0.0476, | |
| "step": 2160 | |
| }, | |
| { | |
| "grad_norm": 0.2728676497936249, | |
| "learning_rate": 3.2426039539425876e-05, | |
| "loss": 0.0414, | |
| "step": 2170 | |
| }, | |
| { | |
| "grad_norm": 0.6051283478736877, | |
| "learning_rate": 3.226799430050268e-05, | |
| "loss": 0.0499, | |
| "step": 2180 | |
| }, | |
| { | |
| "grad_norm": 0.5630646347999573, | |
| "learning_rate": 3.210963113596374e-05, | |
| "loss": 0.0418, | |
| "step": 2190 | |
| }, | |
| { | |
| "grad_norm": 0.4311167001724243, | |
| "learning_rate": 3.1950956973126906e-05, | |
| "loss": 0.0499, | |
| "step": 2200 | |
| }, | |
| { | |
| "grad_norm": 0.4380161166191101, | |
| "learning_rate": 3.1791978752914085e-05, | |
| "loss": 0.0401, | |
| "step": 2210 | |
| }, | |
| { | |
| "grad_norm": 0.43494531512260437, | |
| "learning_rate": 3.163270342954766e-05, | |
| "loss": 0.0469, | |
| "step": 2220 | |
| }, | |
| { | |
| "grad_norm": 0.5626083612442017, | |
| "learning_rate": 3.1473137970246246e-05, | |
| "loss": 0.049, | |
| "step": 2230 | |
| }, | |
| { | |
| "grad_norm": 0.4352441132068634, | |
| "learning_rate": 3.1313289354919945e-05, | |
| "loss": 0.0459, | |
| "step": 2240 | |
| }, | |
| { | |
| "grad_norm": 0.48047247529029846, | |
| "learning_rate": 3.1153164575865046e-05, | |
| "loss": 0.0459, | |
| "step": 2250 | |
| }, | |
| { | |
| "grad_norm": 0.6736582517623901, | |
| "learning_rate": 3.099277063745811e-05, | |
| "loss": 0.0539, | |
| "step": 2260 | |
| }, | |
| { | |
| "grad_norm": 0.40036192536354065, | |
| "learning_rate": 3.0832114555849614e-05, | |
| "loss": 0.0438, | |
| "step": 2270 | |
| }, | |
| { | |
| "grad_norm": 0.42229387164115906, | |
| "learning_rate": 3.0671203358657e-05, | |
| "loss": 0.0413, | |
| "step": 2280 | |
| }, | |
| { | |
| "grad_norm": 0.3298429250717163, | |
| "learning_rate": 3.051004408465733e-05, | |
| "loss": 0.04, | |
| "step": 2290 | |
| }, | |
| { | |
| "grad_norm": 0.43710318207740784, | |
| "learning_rate": 3.034864378347933e-05, | |
| "loss": 0.0454, | |
| "step": 2300 | |
| }, | |
| { | |
| "grad_norm": 0.4662761390209198, | |
| "learning_rate": 3.018700951529504e-05, | |
| "loss": 0.0521, | |
| "step": 2310 | |
| }, | |
| { | |
| "grad_norm": 0.5804249048233032, | |
| "learning_rate": 3.0025148350510974e-05, | |
| "loss": 0.0475, | |
| "step": 2320 | |
| }, | |
| { | |
| "grad_norm": 0.4282180070877075, | |
| "learning_rate": 2.986306736945883e-05, | |
| "loss": 0.0466, | |
| "step": 2330 | |
| }, | |
| { | |
| "grad_norm": 0.45659002661705017, | |
| "learning_rate": 2.970077366208579e-05, | |
| "loss": 0.0387, | |
| "step": 2340 | |
| }, | |
| { | |
| "grad_norm": 0.5479471683502197, | |
| "learning_rate": 2.953827432764438e-05, | |
| "loss": 0.0461, | |
| "step": 2350 | |
| }, | |
| { | |
| "grad_norm": 0.31041985750198364, | |
| "learning_rate": 2.9375576474381905e-05, | |
| "loss": 0.042, | |
| "step": 2360 | |
| }, | |
| { | |
| "grad_norm": 0.42949941754341125, | |
| "learning_rate": 2.921268721922954e-05, | |
| "loss": 0.0435, | |
| "step": 2370 | |
| }, | |
| { | |
| "grad_norm": 0.4704861044883728, | |
| "learning_rate": 2.904961368749099e-05, | |
| "loss": 0.0447, | |
| "step": 2380 | |
| }, | |
| { | |
| "grad_norm": 0.4041045606136322, | |
| "learning_rate": 2.8886363012530826e-05, | |
| "loss": 0.0429, | |
| "step": 2390 | |
| }, | |
| { | |
| "grad_norm": 0.3385079503059387, | |
| "learning_rate": 2.8722942335462415e-05, | |
| "loss": 0.0454, | |
| "step": 2400 | |
| }, | |
| { | |
| "grad_norm": 0.5993208289146423, | |
| "learning_rate": 2.8559358804835597e-05, | |
| "loss": 0.0426, | |
| "step": 2410 | |
| }, | |
| { | |
| "grad_norm": 0.36963415145874023, | |
| "learning_rate": 2.839561957632393e-05, | |
| "loss": 0.0515, | |
| "step": 2420 | |
| }, | |
| { | |
| "grad_norm": 0.5569792985916138, | |
| "learning_rate": 2.823173181241171e-05, | |
| "loss": 0.0484, | |
| "step": 2430 | |
| }, | |
| { | |
| "grad_norm": 0.4478709101676941, | |
| "learning_rate": 2.806770268208066e-05, | |
| "loss": 0.0442, | |
| "step": 2440 | |
| }, | |
| { | |
| "grad_norm": 0.39773526787757874, | |
| "learning_rate": 2.7903539360496323e-05, | |
| "loss": 0.0362, | |
| "step": 2450 | |
| }, | |
| { | |
| "grad_norm": 0.4185701906681061, | |
| "learning_rate": 2.773924902869418e-05, | |
| "loss": 0.046, | |
| "step": 2460 | |
| }, | |
| { | |
| "grad_norm": 0.43949025869369507, | |
| "learning_rate": 2.757483887326559e-05, | |
| "loss": 0.0433, | |
| "step": 2470 | |
| }, | |
| { | |
| "grad_norm": 0.4299829602241516, | |
| "learning_rate": 2.741031608604337e-05, | |
| "loss": 0.0394, | |
| "step": 2480 | |
| }, | |
| { | |
| "grad_norm": 0.27871155738830566, | |
| "learning_rate": 2.7245687863787195e-05, | |
| "loss": 0.0367, | |
| "step": 2490 | |
| }, | |
| { | |
| "grad_norm": 0.4810972511768341, | |
| "learning_rate": 2.7080961407868848e-05, | |
| "loss": 0.0428, | |
| "step": 2500 | |
| }, | |
| { | |
| "grad_norm": 0.3548348546028137, | |
| "learning_rate": 2.6916143923957138e-05, | |
| "loss": 0.0435, | |
| "step": 2510 | |
| }, | |
| { | |
| "grad_norm": 0.3431377410888672, | |
| "learning_rate": 2.675124262170275e-05, | |
| "loss": 0.0336, | |
| "step": 2520 | |
| }, | |
| { | |
| "grad_norm": 0.4408491551876068, | |
| "learning_rate": 2.6586264714422836e-05, | |
| "loss": 0.0404, | |
| "step": 2530 | |
| }, | |
| { | |
| "grad_norm": 1.0779521465301514, | |
| "learning_rate": 2.6421217418785544e-05, | |
| "loss": 0.0424, | |
| "step": 2540 | |
| }, | |
| { | |
| "grad_norm": 0.40511152148246765, | |
| "learning_rate": 2.6256107954494242e-05, | |
| "loss": 0.0476, | |
| "step": 2550 | |
| }, | |
| { | |
| "grad_norm": 0.36247146129608154, | |
| "learning_rate": 2.6090943543971784e-05, | |
| "loss": 0.0436, | |
| "step": 2560 | |
| }, | |
| { | |
| "grad_norm": 0.3747636079788208, | |
| "learning_rate": 2.592573141204455e-05, | |
| "loss": 0.0404, | |
| "step": 2570 | |
| }, | |
| { | |
| "grad_norm": 0.5357540249824524, | |
| "learning_rate": 2.5760478785626397e-05, | |
| "loss": 0.0444, | |
| "step": 2580 | |
| }, | |
| { | |
| "grad_norm": 0.3767753839492798, | |
| "learning_rate": 2.5595192893402553e-05, | |
| "loss": 0.0397, | |
| "step": 2590 | |
| }, | |
| { | |
| "grad_norm": 0.3285132944583893, | |
| "learning_rate": 2.5429880965513386e-05, | |
| "loss": 0.0381, | |
| "step": 2600 | |
| }, | |
| { | |
| "grad_norm": 0.4897662401199341, | |
| "learning_rate": 2.526455023323817e-05, | |
| "loss": 0.039, | |
| "step": 2610 | |
| }, | |
| { | |
| "grad_norm": 0.4508710205554962, | |
| "learning_rate": 2.5099207928678732e-05, | |
| "loss": 0.0395, | |
| "step": 2620 | |
| }, | |
| { | |
| "grad_norm": 0.3298162519931793, | |
| "learning_rate": 2.4933861284443112e-05, | |
| "loss": 0.0413, | |
| "step": 2630 | |
| }, | |
| { | |
| "grad_norm": 0.42930713295936584, | |
| "learning_rate": 2.4768517533329157e-05, | |
| "loss": 0.0461, | |
| "step": 2640 | |
| }, | |
| { | |
| "grad_norm": 0.43155762553215027, | |
| "learning_rate": 2.460318390800819e-05, | |
| "loss": 0.0453, | |
| "step": 2650 | |
| }, | |
| { | |
| "grad_norm": 0.3605857789516449, | |
| "learning_rate": 2.44378676407086e-05, | |
| "loss": 0.0445, | |
| "step": 2660 | |
| }, | |
| { | |
| "grad_norm": 0.2702181041240692, | |
| "learning_rate": 2.4272575962899462e-05, | |
| "loss": 0.0421, | |
| "step": 2670 | |
| }, | |
| { | |
| "grad_norm": 0.3844555616378784, | |
| "learning_rate": 2.410731610497424e-05, | |
| "loss": 0.0418, | |
| "step": 2680 | |
| }, | |
| { | |
| "grad_norm": 0.4364534914493561, | |
| "learning_rate": 2.3942095295934476e-05, | |
| "loss": 0.0379, | |
| "step": 2690 | |
| }, | |
| { | |
| "grad_norm": 0.33137261867523193, | |
| "learning_rate": 2.3776920763073602e-05, | |
| "loss": 0.0422, | |
| "step": 2700 | |
| }, | |
| { | |
| "grad_norm": 0.2672342360019684, | |
| "learning_rate": 2.361179973166078e-05, | |
| "loss": 0.0424, | |
| "step": 2710 | |
| }, | |
| { | |
| "grad_norm": 0.34315600991249084, | |
| "learning_rate": 2.3446739424624827e-05, | |
| "loss": 0.0411, | |
| "step": 2720 | |
| }, | |
| { | |
| "grad_norm": 0.4923914670944214, | |
| "learning_rate": 2.328174706223832e-05, | |
| "loss": 0.0439, | |
| "step": 2730 | |
| }, | |
| { | |
| "grad_norm": 0.26800429821014404, | |
| "learning_rate": 2.3116829861801686e-05, | |
| "loss": 0.0391, | |
| "step": 2740 | |
| }, | |
| { | |
| "grad_norm": 0.29833143949508667, | |
| "learning_rate": 2.2951995037327515e-05, | |
| "loss": 0.0397, | |
| "step": 2750 | |
| }, | |
| { | |
| "grad_norm": 0.3327856957912445, | |
| "learning_rate": 2.2787249799225024e-05, | |
| "loss": 0.0381, | |
| "step": 2760 | |
| }, | |
| { | |
| "grad_norm": 0.46502891182899475, | |
| "learning_rate": 2.2622601353984635e-05, | |
| "loss": 0.0411, | |
| "step": 2770 | |
| }, | |
| { | |
| "grad_norm": 0.40785571932792664, | |
| "learning_rate": 2.2458056903862726e-05, | |
| "loss": 0.0379, | |
| "step": 2780 | |
| }, | |
| { | |
| "grad_norm": 0.3968541622161865, | |
| "learning_rate": 2.229362364656659e-05, | |
| "loss": 0.0429, | |
| "step": 2790 | |
| }, | |
| { | |
| "grad_norm": 0.40930506587028503, | |
| "learning_rate": 2.21293087749396e-05, | |
| "loss": 0.0374, | |
| "step": 2800 | |
| }, | |
| { | |
| "grad_norm": 0.4295065701007843, | |
| "learning_rate": 2.1965119476646547e-05, | |
| "loss": 0.0366, | |
| "step": 2810 | |
| }, | |
| { | |
| "grad_norm": 0.4438641667366028, | |
| "learning_rate": 2.1801062933859235e-05, | |
| "loss": 0.0378, | |
| "step": 2820 | |
| }, | |
| { | |
| "grad_norm": 0.3975362181663513, | |
| "learning_rate": 2.1637146322942315e-05, | |
| "loss": 0.036, | |
| "step": 2830 | |
| }, | |
| { | |
| "grad_norm": 0.44547799229621887, | |
| "learning_rate": 2.1473376814139362e-05, | |
| "loss": 0.0395, | |
| "step": 2840 | |
| }, | |
| { | |
| "grad_norm": 0.4006638526916504, | |
| "learning_rate": 2.1309761571259237e-05, | |
| "loss": 0.0367, | |
| "step": 2850 | |
| }, | |
| { | |
| "grad_norm": 0.27414795756340027, | |
| "learning_rate": 2.1146307751362694e-05, | |
| "loss": 0.0312, | |
| "step": 2860 | |
| }, | |
| { | |
| "grad_norm": 0.4664889872074127, | |
| "learning_rate": 2.098302250444934e-05, | |
| "loss": 0.0413, | |
| "step": 2870 | |
| }, | |
| { | |
| "grad_norm": 0.19164593517780304, | |
| "learning_rate": 2.0819912973144843e-05, | |
| "loss": 0.0345, | |
| "step": 2880 | |
| }, | |
| { | |
| "grad_norm": 0.33589157462120056, | |
| "learning_rate": 2.065698629238851e-05, | |
| "loss": 0.0355, | |
| "step": 2890 | |
| }, | |
| { | |
| "grad_norm": 0.33917805552482605, | |
| "learning_rate": 2.0494249589121157e-05, | |
| "loss": 0.0393, | |
| "step": 2900 | |
| }, | |
| { | |
| "grad_norm": 0.32117098569869995, | |
| "learning_rate": 2.033170998197339e-05, | |
| "loss": 0.0448, | |
| "step": 2910 | |
| }, | |
| { | |
| "grad_norm": 0.6140998005867004, | |
| "learning_rate": 2.0169374580954165e-05, | |
| "loss": 0.0378, | |
| "step": 2920 | |
| }, | |
| { | |
| "grad_norm": 0.5331240296363831, | |
| "learning_rate": 2.000725048713983e-05, | |
| "loss": 0.0365, | |
| "step": 2930 | |
| }, | |
| { | |
| "grad_norm": 0.3481861352920532, | |
| "learning_rate": 1.9845344792363447e-05, | |
| "loss": 0.0356, | |
| "step": 2940 | |
| }, | |
| { | |
| "grad_norm": 0.5170403718948364, | |
| "learning_rate": 1.9683664578904614e-05, | |
| "loss": 0.0403, | |
| "step": 2950 | |
| }, | |
| { | |
| "grad_norm": 0.31584399938583374, | |
| "learning_rate": 1.9522216919179644e-05, | |
| "loss": 0.0404, | |
| "step": 2960 | |
| }, | |
| { | |
| "grad_norm": 0.3115321099758148, | |
| "learning_rate": 1.9361008875432185e-05, | |
| "loss": 0.0358, | |
| "step": 2970 | |
| }, | |
| { | |
| "grad_norm": 0.338223397731781, | |
| "learning_rate": 1.9200047499424308e-05, | |
| "loss": 0.0331, | |
| "step": 2980 | |
| }, | |
| { | |
| "grad_norm": 0.48482340574264526, | |
| "learning_rate": 1.9039339832128057e-05, | |
| "loss": 0.0347, | |
| "step": 2990 | |
| }, | |
| { | |
| "grad_norm": 0.4181440770626068, | |
| "learning_rate": 1.8878892903417404e-05, | |
| "loss": 0.0389, | |
| "step": 3000 | |
| }, | |
| { | |
| "grad_norm": 0.42534828186035156, | |
| "learning_rate": 1.871871373176078e-05, | |
| "loss": 0.0392, | |
| "step": 3010 | |
| }, | |
| { | |
| "grad_norm": 0.5645197629928589, | |
| "learning_rate": 1.8558809323914083e-05, | |
| "loss": 0.0461, | |
| "step": 3020 | |
| }, | |
| { | |
| "grad_norm": 0.36499112844467163, | |
| "learning_rate": 1.8399186674614126e-05, | |
| "loss": 0.0343, | |
| "step": 3030 | |
| }, | |
| { | |
| "grad_norm": 0.2742806673049927, | |
| "learning_rate": 1.823985276627269e-05, | |
| "loss": 0.0283, | |
| "step": 3040 | |
| }, | |
| { | |
| "grad_norm": 0.39008432626724243, | |
| "learning_rate": 1.80808145686711e-05, | |
| "loss": 0.0374, | |
| "step": 3050 | |
| }, | |
| { | |
| "grad_norm": 0.36456918716430664, | |
| "learning_rate": 1.7922079038655326e-05, | |
| "loss": 0.037, | |
| "step": 3060 | |
| }, | |
| { | |
| "grad_norm": 0.2564813792705536, | |
| "learning_rate": 1.7763653119831684e-05, | |
| "loss": 0.0374, | |
| "step": 3070 | |
| }, | |
| { | |
| "grad_norm": 0.4312773048877716, | |
| "learning_rate": 1.7605543742263087e-05, | |
| "loss": 0.0353, | |
| "step": 3080 | |
| }, | |
| { | |
| "grad_norm": 0.30969807505607605, | |
| "learning_rate": 1.744775782216593e-05, | |
| "loss": 0.0339, | |
| "step": 3090 | |
| }, | |
| { | |
| "grad_norm": 0.3753385543823242, | |
| "learning_rate": 1.7290302261607503e-05, | |
| "loss": 0.0323, | |
| "step": 3100 | |
| }, | |
| { | |
| "grad_norm": 0.3404577672481537, | |
| "learning_rate": 1.7133183948204108e-05, | |
| "loss": 0.0346, | |
| "step": 3110 | |
| }, | |
| { | |
| "grad_norm": 0.33649688959121704, | |
| "learning_rate": 1.6976409754819767e-05, | |
| "loss": 0.0332, | |
| "step": 3120 | |
| }, | |
| { | |
| "grad_norm": 0.7442240715026855, | |
| "learning_rate": 1.6819986539265582e-05, | |
| "loss": 0.0392, | |
| "step": 3130 | |
| }, | |
| { | |
| "grad_norm": 0.3342534899711609, | |
| "learning_rate": 1.6663921143999733e-05, | |
| "loss": 0.0368, | |
| "step": 3140 | |
| }, | |
| { | |
| "grad_norm": 0.32587477564811707, | |
| "learning_rate": 1.650822039582819e-05, | |
| "loss": 0.0318, | |
| "step": 3150 | |
| }, | |
| { | |
| "grad_norm": 0.5487452149391174, | |
| "learning_rate": 1.63528911056061e-05, | |
| "loss": 0.0376, | |
| "step": 3160 | |
| }, | |
| { | |
| "grad_norm": 0.5345733165740967, | |
| "learning_rate": 1.619794006793979e-05, | |
| "loss": 0.0386, | |
| "step": 3170 | |
| }, | |
| { | |
| "grad_norm": 0.3318815231323242, | |
| "learning_rate": 1.604337406088963e-05, | |
| "loss": 0.0308, | |
| "step": 3180 | |
| }, | |
| { | |
| "grad_norm": 0.41096043586730957, | |
| "learning_rate": 1.588919984567349e-05, | |
| "loss": 0.036, | |
| "step": 3190 | |
| }, | |
| { | |
| "grad_norm": 0.49638691544532776, | |
| "learning_rate": 1.5735424166371e-05, | |
| "loss": 0.0404, | |
| "step": 3200 | |
| }, | |
| { | |
| "grad_norm": 0.38568755984306335, | |
| "learning_rate": 1.558205374962854e-05, | |
| "loss": 0.033, | |
| "step": 3210 | |
| }, | |
| { | |
| "grad_norm": 0.34658241271972656, | |
| "learning_rate": 1.5429095304364976e-05, | |
| "loss": 0.0327, | |
| "step": 3220 | |
| }, | |
| { | |
| "grad_norm": 0.33020010590553284, | |
| "learning_rate": 1.527655552147824e-05, | |
| "loss": 0.0373, | |
| "step": 3230 | |
| }, | |
| { | |
| "grad_norm": 0.5657230615615845, | |
| "learning_rate": 1.5124441073552585e-05, | |
| "loss": 0.0354, | |
| "step": 3240 | |
| }, | |
| { | |
| "grad_norm": 0.3239688277244568, | |
| "learning_rate": 1.4972758614566746e-05, | |
| "loss": 0.0295, | |
| "step": 3250 | |
| }, | |
| { | |
| "grad_norm": 0.387555330991745, | |
| "learning_rate": 1.4821514779602864e-05, | |
| "loss": 0.0436, | |
| "step": 3260 | |
| }, | |
| { | |
| "grad_norm": 0.3232133388519287, | |
| "learning_rate": 1.467071618455624e-05, | |
| "loss": 0.0322, | |
| "step": 3270 | |
| }, | |
| { | |
| "grad_norm": 0.2886604964733124, | |
| "learning_rate": 1.452036942584595e-05, | |
| "loss": 0.0304, | |
| "step": 3280 | |
| }, | |
| { | |
| "grad_norm": 0.2882493734359741, | |
| "learning_rate": 1.4370481080126248e-05, | |
| "loss": 0.0314, | |
| "step": 3290 | |
| }, | |
| { | |
| "grad_norm": 0.4715385138988495, | |
| "learning_rate": 1.4221057703998985e-05, | |
| "loss": 0.0334, | |
| "step": 3300 | |
| }, | |
| { | |
| "grad_norm": 0.3316683769226074, | |
| "learning_rate": 1.4072105833726684e-05, | |
| "loss": 0.0351, | |
| "step": 3310 | |
| }, | |
| { | |
| "grad_norm": 0.3624657988548279, | |
| "learning_rate": 1.3923631984946672e-05, | |
| "loss": 0.035, | |
| "step": 3320 | |
| }, | |
| { | |
| "grad_norm": 0.36212462186813354, | |
| "learning_rate": 1.3775642652386103e-05, | |
| "loss": 0.0301, | |
| "step": 3330 | |
| }, | |
| { | |
| "grad_norm": 0.36848610639572144, | |
| "learning_rate": 1.3628144309577784e-05, | |
| "loss": 0.0324, | |
| "step": 3340 | |
| }, | |
| { | |
| "grad_norm": 0.4272776246070862, | |
| "learning_rate": 1.3481143408577079e-05, | |
| "loss": 0.0373, | |
| "step": 3350 | |
| }, | |
| { | |
| "grad_norm": 0.37170878052711487, | |
| "learning_rate": 1.3334646379679583e-05, | |
| "loss": 0.033, | |
| "step": 3360 | |
| }, | |
| { | |
| "grad_norm": 0.48944494128227234, | |
| "learning_rate": 1.3188659631139966e-05, | |
| "loss": 0.0414, | |
| "step": 3370 | |
| }, | |
| { | |
| "grad_norm": 0.3089650869369507, | |
| "learning_rate": 1.3043189548891516e-05, | |
| "loss": 0.0338, | |
| "step": 3380 | |
| }, | |
| { | |
| "grad_norm": 0.3926386535167694, | |
| "learning_rate": 1.2898242496266884e-05, | |
| "loss": 0.0351, | |
| "step": 3390 | |
| }, | |
| { | |
| "grad_norm": 0.40642455220222473, | |
| "learning_rate": 1.2753824813719733e-05, | |
| "loss": 0.0412, | |
| "step": 3400 | |
| }, | |
| { | |
| "grad_norm": 0.35056406259536743, | |
| "learning_rate": 1.2609942818547327e-05, | |
| "loss": 0.035, | |
| "step": 3410 | |
| }, | |
| { | |
| "grad_norm": 0.2923920750617981, | |
| "learning_rate": 1.2466602804614266e-05, | |
| "loss": 0.035, | |
| "step": 3420 | |
| }, | |
| { | |
| "grad_norm": 0.9390926957130432, | |
| "learning_rate": 1.2323811042077096e-05, | |
| "loss": 0.0298, | |
| "step": 3430 | |
| }, | |
| { | |
| "grad_norm": 0.5153149366378784, | |
| "learning_rate": 1.2181573777110106e-05, | |
| "loss": 0.0376, | |
| "step": 3440 | |
| }, | |
| { | |
| "grad_norm": 0.2597624659538269, | |
| "learning_rate": 1.2039897231632055e-05, | |
| "loss": 0.0333, | |
| "step": 3450 | |
| }, | |
| { | |
| "grad_norm": 0.36722615361213684, | |
| "learning_rate": 1.1898787603033995e-05, | |
| "loss": 0.038, | |
| "step": 3460 | |
| }, | |
| { | |
| "grad_norm": 0.23120303452014923, | |
| "learning_rate": 1.1758251063908227e-05, | |
| "loss": 0.0305, | |
| "step": 3470 | |
| }, | |
| { | |
| "grad_norm": 0.4411832094192505, | |
| "learning_rate": 1.1618293761778235e-05, | |
| "loss": 0.0305, | |
| "step": 3480 | |
| }, | |
| { | |
| "grad_norm": 0.36204567551612854, | |
| "learning_rate": 1.1478921818829827e-05, | |
| "loss": 0.0325, | |
| "step": 3490 | |
| }, | |
| { | |
| "grad_norm": 0.38650524616241455, | |
| "learning_rate": 1.1340141331643276e-05, | |
| "loss": 0.037, | |
| "step": 3500 | |
| }, | |
| { | |
| "grad_norm": 0.29249370098114014, | |
| "learning_rate": 1.1201958370926682e-05, | |
| "loss": 0.0304, | |
| "step": 3510 | |
| }, | |
| { | |
| "grad_norm": 0.3050512969493866, | |
| "learning_rate": 1.1064378981250409e-05, | |
| "loss": 0.0322, | |
| "step": 3520 | |
| }, | |
| { | |
| "grad_norm": 0.36122196912765503, | |
| "learning_rate": 1.0927409180782638e-05, | |
| "loss": 0.0325, | |
| "step": 3530 | |
| }, | |
| { | |
| "grad_norm": 0.3820047974586487, | |
| "learning_rate": 1.0791054961026182e-05, | |
| "loss": 0.0361, | |
| "step": 3540 | |
| }, | |
| { | |
| "grad_norm": 0.3520725965499878, | |
| "learning_rate": 1.0655322286556318e-05, | |
| "loss": 0.0333, | |
| "step": 3550 | |
| }, | |
| { | |
| "grad_norm": 0.3393506705760956, | |
| "learning_rate": 1.0520217094759962e-05, | |
| "loss": 0.0352, | |
| "step": 3560 | |
| }, | |
| { | |
| "grad_norm": 0.3986015021800995, | |
| "learning_rate": 1.0385745295575866e-05, | |
| "loss": 0.0331, | |
| "step": 3570 | |
| }, | |
| { | |
| "grad_norm": 0.3583716154098511, | |
| "learning_rate": 1.0251912771236159e-05, | |
| "loss": 0.0319, | |
| "step": 3580 | |
| }, | |
| { | |
| "grad_norm": 0.27882295846939087, | |
| "learning_rate": 1.0118725376009025e-05, | |
| "loss": 0.0341, | |
| "step": 3590 | |
| }, | |
| { | |
| "grad_norm": 0.32304292917251587, | |
| "learning_rate": 9.986188935942579e-06, | |
| "loss": 0.0291, | |
| "step": 3600 | |
| }, | |
| { | |
| "grad_norm": 0.5138395428657532, | |
| "learning_rate": 9.854309248610085e-06, | |
| "loss": 0.0313, | |
| "step": 3610 | |
| }, | |
| { | |
| "grad_norm": 0.31992360949516296, | |
| "learning_rate": 9.723092082856294e-06, | |
| "loss": 0.0305, | |
| "step": 3620 | |
| }, | |
| { | |
| "grad_norm": 0.2782381772994995, | |
| "learning_rate": 9.592543178545107e-06, | |
| "loss": 0.029, | |
| "step": 3630 | |
| }, | |
| { | |
| "grad_norm": 0.27642911672592163, | |
| "learning_rate": 9.462668246308529e-06, | |
| "loss": 0.0329, | |
| "step": 3640 | |
| }, | |
| { | |
| "grad_norm": 0.33659425377845764, | |
| "learning_rate": 9.333472967296833e-06, | |
| "loss": 0.0328, | |
| "step": 3650 | |
| }, | |
| { | |
| "grad_norm": 0.40272244811058044, | |
| "learning_rate": 9.204962992930063e-06, | |
| "loss": 0.0284, | |
| "step": 3660 | |
| }, | |
| { | |
| "grad_norm": 0.4781797528266907, | |
| "learning_rate": 9.077143944650802e-06, | |
| "loss": 0.0323, | |
| "step": 3670 | |
| }, | |
| { | |
| "grad_norm": 0.34181609749794006, | |
| "learning_rate": 8.950021413678306e-06, | |
| "loss": 0.0305, | |
| "step": 3680 | |
| }, | |
| { | |
| "grad_norm": 0.3515138328075409, | |
| "learning_rate": 8.8236009607639e-06, | |
| "loss": 0.0311, | |
| "step": 3690 | |
| }, | |
| { | |
| "grad_norm": 0.42265385389328003, | |
| "learning_rate": 8.697888115947725e-06, | |
| "loss": 0.0294, | |
| "step": 3700 | |
| }, | |
| { | |
| "grad_norm": 0.24089132249355316, | |
| "learning_rate": 8.572888378316884e-06, | |
| "loss": 0.0346, | |
| "step": 3710 | |
| }, | |
| { | |
| "grad_norm": 0.3806060254573822, | |
| "learning_rate": 8.448607215764823e-06, | |
| "loss": 0.0274, | |
| "step": 3720 | |
| }, | |
| { | |
| "grad_norm": 0.40154969692230225, | |
| "learning_rate": 8.325050064752238e-06, | |
| "loss": 0.0321, | |
| "step": 3730 | |
| }, | |
| { | |
| "grad_norm": 0.6564791798591614, | |
| "learning_rate": 8.202222330069167e-06, | |
| "loss": 0.0324, | |
| "step": 3740 | |
| }, | |
| { | |
| "grad_norm": 0.4108399748802185, | |
| "learning_rate": 8.080129384598625e-06, | |
| "loss": 0.0301, | |
| "step": 3750 | |
| }, | |
| { | |
| "grad_norm": 0.19246874749660492, | |
| "learning_rate": 7.958776569081586e-06, | |
| "loss": 0.0288, | |
| "step": 3760 | |
| }, | |
| { | |
| "grad_norm": 0.4110735058784485, | |
| "learning_rate": 7.838169191883316e-06, | |
| "loss": 0.0382, | |
| "step": 3770 | |
| }, | |
| { | |
| "grad_norm": 0.25348952412605286, | |
| "learning_rate": 7.718312528761223e-06, | |
| "loss": 0.0311, | |
| "step": 3780 | |
| }, | |
| { | |
| "grad_norm": 0.3308151662349701, | |
| "learning_rate": 7.59921182263402e-06, | |
| "loss": 0.0274, | |
| "step": 3790 | |
| }, | |
| { | |
| "grad_norm": 0.28520438075065613, | |
| "learning_rate": 7.480872283352428e-06, | |
| "loss": 0.0309, | |
| "step": 3800 | |
| }, | |
| { | |
| "grad_norm": 0.337143212556839, | |
| "learning_rate": 7.363299087471276e-06, | |
| "loss": 0.0259, | |
| "step": 3810 | |
| }, | |
| { | |
| "grad_norm": 0.22851786017417908, | |
| "learning_rate": 7.246497378023018e-06, | |
| "loss": 0.0323, | |
| "step": 3820 | |
| }, | |
| { | |
| "grad_norm": 0.5074667930603027, | |
| "learning_rate": 7.130472264292823e-06, | |
| "loss": 0.0292, | |
| "step": 3830 | |
| }, | |
| { | |
| "grad_norm": 0.2871147096157074, | |
| "learning_rate": 7.015228821595024e-06, | |
| "loss": 0.0296, | |
| "step": 3840 | |
| }, | |
| { | |
| "grad_norm": 0.30016905069351196, | |
| "learning_rate": 6.9007720910511554e-06, | |
| "loss": 0.0231, | |
| "step": 3850 | |
| }, | |
| { | |
| "grad_norm": 0.28288936614990234, | |
| "learning_rate": 6.7871070793693815e-06, | |
| "loss": 0.033, | |
| "step": 3860 | |
| }, | |
| { | |
| "grad_norm": 0.44114962220191956, | |
| "learning_rate": 6.6742387586255505e-06, | |
| "loss": 0.0343, | |
| "step": 3870 | |
| }, | |
| { | |
| "grad_norm": 0.22615277767181396, | |
| "learning_rate": 6.562172066045655e-06, | |
| "loss": 0.0267, | |
| "step": 3880 | |
| }, | |
| { | |
| "grad_norm": 0.24401821196079254, | |
| "learning_rate": 6.450911903789863e-06, | |
| "loss": 0.0322, | |
| "step": 3890 | |
| }, | |
| { | |
| "grad_norm": 0.4524429738521576, | |
| "learning_rate": 6.340463138738123e-06, | |
| "loss": 0.0292, | |
| "step": 3900 | |
| }, | |
| { | |
| "grad_norm": 0.3167012929916382, | |
| "learning_rate": 6.230830602277199e-06, | |
| "loss": 0.0325, | |
| "step": 3910 | |
| }, | |
| { | |
| "grad_norm": 0.2682662010192871, | |
| "learning_rate": 6.1220190900894175e-06, | |
| "loss": 0.0273, | |
| "step": 3920 | |
| }, | |
| { | |
| "grad_norm": 0.29721036553382874, | |
| "learning_rate": 6.014033361942806e-06, | |
| "loss": 0.0266, | |
| "step": 3930 | |
| }, | |
| { | |
| "grad_norm": 0.42319223284721375, | |
| "learning_rate": 5.906878141482944e-06, | |
| "loss": 0.0322, | |
| "step": 3940 | |
| }, | |
| { | |
| "grad_norm": 0.38211673498153687, | |
| "learning_rate": 5.800558116026319e-06, | |
| "loss": 0.0312, | |
| "step": 3950 | |
| }, | |
| { | |
| "grad_norm": 0.27969926595687866, | |
| "learning_rate": 5.695077936355259e-06, | |
| "loss": 0.0263, | |
| "step": 3960 | |
| }, | |
| { | |
| "grad_norm": 0.22936464846134186, | |
| "learning_rate": 5.590442216514543e-06, | |
| "loss": 0.0309, | |
| "step": 3970 | |
| }, | |
| { | |
| "grad_norm": 0.3358422517776489, | |
| "learning_rate": 5.48665553360952e-06, | |
| "loss": 0.0243, | |
| "step": 3980 | |
| }, | |
| { | |
| "grad_norm": 0.2455916702747345, | |
| "learning_rate": 5.383722427605931e-06, | |
| "loss": 0.0293, | |
| "step": 3990 | |
| }, | |
| { | |
| "grad_norm": 0.24449455738067627, | |
| "learning_rate": 5.281647401131279e-06, | |
| "loss": 0.0283, | |
| "step": 4000 | |
| }, | |
| { | |
| "grad_norm": 0.40746790170669556, | |
| "learning_rate": 5.1804349192779044e-06, | |
| "loss": 0.0324, | |
| "step": 4010 | |
| }, | |
| { | |
| "grad_norm": 0.3240891695022583, | |
| "learning_rate": 5.0800894094076565e-06, | |
| "loss": 0.024, | |
| "step": 4020 | |
| }, | |
| { | |
| "grad_norm": 0.25076866149902344, | |
| "learning_rate": 4.980615260958193e-06, | |
| "loss": 0.0328, | |
| "step": 4030 | |
| }, | |
| { | |
| "grad_norm": 0.1966366469860077, | |
| "learning_rate": 4.882016825251037e-06, | |
| "loss": 0.0303, | |
| "step": 4040 | |
| }, | |
| { | |
| "grad_norm": 0.2640022337436676, | |
| "learning_rate": 4.784298415301172e-06, | |
| "loss": 0.0254, | |
| "step": 4050 | |
| }, | |
| { | |
| "grad_norm": 0.33577871322631836, | |
| "learning_rate": 4.687464305628405e-06, | |
| "loss": 0.0296, | |
| "step": 4060 | |
| }, | |
| { | |
| "grad_norm": 0.40677428245544434, | |
| "learning_rate": 4.591518732070402e-06, | |
| "loss": 0.0278, | |
| "step": 4070 | |
| }, | |
| { | |
| "grad_norm": 0.46990591287612915, | |
| "learning_rate": 4.4964658915973676e-06, | |
| "loss": 0.0279, | |
| "step": 4080 | |
| }, | |
| { | |
| "grad_norm": 0.22323285043239594, | |
| "learning_rate": 4.40230994212848e-06, | |
| "loss": 0.0293, | |
| "step": 4090 | |
| }, | |
| { | |
| "grad_norm": 0.2634666860103607, | |
| "learning_rate": 4.309055002349987e-06, | |
| "loss": 0.0303, | |
| "step": 4100 | |
| }, | |
| { | |
| "grad_norm": 0.45921197533607483, | |
| "learning_rate": 4.21670515153505e-06, | |
| "loss": 0.0297, | |
| "step": 4110 | |
| }, | |
| { | |
| "grad_norm": 0.640613853931427, | |
| "learning_rate": 4.12526442936533e-06, | |
| "loss": 0.0241, | |
| "step": 4120 | |
| }, | |
| { | |
| "grad_norm": 0.3607765734195709, | |
| "learning_rate": 4.03473683575423e-06, | |
| "loss": 0.0263, | |
| "step": 4130 | |
| }, | |
| { | |
| "grad_norm": 0.3656626045703888, | |
| "learning_rate": 3.945126330671969e-06, | |
| "loss": 0.028, | |
| "step": 4140 | |
| }, | |
| { | |
| "grad_norm": 0.8340692520141602, | |
| "learning_rate": 3.856436833972341e-06, | |
| "loss": 0.0298, | |
| "step": 4150 | |
| }, | |
| { | |
| "grad_norm": 0.5998902320861816, | |
| "learning_rate": 3.7686722252212344e-06, | |
| "loss": 0.0285, | |
| "step": 4160 | |
| }, | |
| { | |
| "grad_norm": 0.36758774518966675, | |
| "learning_rate": 3.681836343526965e-06, | |
| "loss": 0.0329, | |
| "step": 4170 | |
| }, | |
| { | |
| "grad_norm": 0.31286585330963135, | |
| "learning_rate": 3.5959329873723e-06, | |
| "loss": 0.0296, | |
| "step": 4180 | |
| }, | |
| { | |
| "grad_norm": 0.38393154740333557, | |
| "learning_rate": 3.510965914448333e-06, | |
| "loss": 0.0329, | |
| "step": 4190 | |
| }, | |
| { | |
| "grad_norm": 0.396363228559494, | |
| "learning_rate": 3.4269388414900792e-06, | |
| "loss": 0.0317, | |
| "step": 4200 | |
| }, | |
| { | |
| "grad_norm": 0.222749263048172, | |
| "learning_rate": 3.3438554441139247e-06, | |
| "loss": 0.0223, | |
| "step": 4210 | |
| }, | |
| { | |
| "grad_norm": 0.24695293605327606, | |
| "learning_rate": 3.2617193566568283e-06, | |
| "loss": 0.0241, | |
| "step": 4220 | |
| }, | |
| { | |
| "grad_norm": 0.4028778374195099, | |
| "learning_rate": 3.1805341720173326e-06, | |
| "loss": 0.0226, | |
| "step": 4230 | |
| }, | |
| { | |
| "grad_norm": 0.2558240592479706, | |
| "learning_rate": 3.100303441498423e-06, | |
| "loss": 0.0315, | |
| "step": 4240 | |
| }, | |
| { | |
| "grad_norm": 0.5111129283905029, | |
| "learning_rate": 3.021030674652156e-06, | |
| "loss": 0.0291, | |
| "step": 4250 | |
| }, | |
| { | |
| "grad_norm": 0.28568732738494873, | |
| "learning_rate": 2.942719339126171e-06, | |
| "loss": 0.0288, | |
| "step": 4260 | |
| }, | |
| { | |
| "grad_norm": 0.9272612929344177, | |
| "learning_rate": 2.8653728605119695e-06, | |
| "loss": 0.0265, | |
| "step": 4270 | |
| }, | |
| { | |
| "grad_norm": 0.2502414584159851, | |
| "learning_rate": 2.788994622195096e-06, | |
| "loss": 0.0257, | |
| "step": 4280 | |
| }, | |
| { | |
| "grad_norm": 0.3649044930934906, | |
| "learning_rate": 2.7135879652071318e-06, | |
| "loss": 0.0257, | |
| "step": 4290 | |
| }, | |
| { | |
| "grad_norm": 0.26225656270980835, | |
| "learning_rate": 2.6391561880795253e-06, | |
| "loss": 0.0256, | |
| "step": 4300 | |
| }, | |
| { | |
| "grad_norm": 0.38593482971191406, | |
| "learning_rate": 2.5657025466993472e-06, | |
| "loss": 0.0238, | |
| "step": 4310 | |
| }, | |
| { | |
| "grad_norm": 0.24045054614543915, | |
| "learning_rate": 2.493230254166817e-06, | |
| "loss": 0.0275, | |
| "step": 4320 | |
| }, | |
| { | |
| "grad_norm": 0.33231034874916077, | |
| "learning_rate": 2.4217424806547984e-06, | |
| "loss": 0.0251, | |
| "step": 4330 | |
| }, | |
| { | |
| "grad_norm": 0.31093278527259827, | |
| "learning_rate": 2.3512423532700805e-06, | |
| "loss": 0.0244, | |
| "step": 4340 | |
| }, | |
| { | |
| "grad_norm": 0.3532539904117584, | |
| "learning_rate": 2.2817329559166293e-06, | |
| "loss": 0.0252, | |
| "step": 4350 | |
| }, | |
| { | |
| "grad_norm": 0.5126403570175171, | |
| "learning_rate": 2.213217329160672e-06, | |
| "loss": 0.0263, | |
| "step": 4360 | |
| }, | |
| { | |
| "grad_norm": 0.3334461450576782, | |
| "learning_rate": 2.1456984700976733e-06, | |
| "loss": 0.0315, | |
| "step": 4370 | |
| }, | |
| { | |
| "grad_norm": 0.3070288598537445, | |
| "learning_rate": 2.079179332221265e-06, | |
| "loss": 0.0283, | |
| "step": 4380 | |
| }, | |
| { | |
| "grad_norm": 0.4461181163787842, | |
| "learning_rate": 2.0136628252940217e-06, | |
| "loss": 0.0274, | |
| "step": 4390 | |
| }, | |
| { | |
| "grad_norm": 0.6602442264556885, | |
| "learning_rate": 1.9491518152202096e-06, | |
| "loss": 0.0317, | |
| "step": 4400 | |
| }, | |
| { | |
| "grad_norm": 0.2886788547039032, | |
| "learning_rate": 1.8856491239203938e-06, | |
| "loss": 0.0253, | |
| "step": 4410 | |
| }, | |
| { | |
| "grad_norm": 0.3347417116165161, | |
| "learning_rate": 1.8231575292080027e-06, | |
| "loss": 0.0316, | |
| "step": 4420 | |
| }, | |
| { | |
| "grad_norm": 0.3918112814426422, | |
| "learning_rate": 1.7616797646678479e-06, | |
| "loss": 0.0289, | |
| "step": 4430 | |
| }, | |
| { | |
| "grad_norm": 0.41051027178764343, | |
| "learning_rate": 1.7012185195365016e-06, | |
| "loss": 0.0304, | |
| "step": 4440 | |
| }, | |
| { | |
| "grad_norm": 0.2741791903972626, | |
| "learning_rate": 1.6417764385846996e-06, | |
| "loss": 0.0241, | |
| "step": 4450 | |
| }, | |
| { | |
| "grad_norm": 0.337773859500885, | |
| "learning_rate": 1.5833561220016253e-06, | |
| "loss": 0.0276, | |
| "step": 4460 | |
| }, | |
| { | |
| "grad_norm": 0.394645631313324, | |
| "learning_rate": 1.5259601252811755e-06, | |
| "loss": 0.0229, | |
| "step": 4470 | |
| }, | |
| { | |
| "grad_norm": 0.26460105180740356, | |
| "learning_rate": 1.4695909591101924e-06, | |
| "loss": 0.0275, | |
| "step": 4480 | |
| }, | |
| { | |
| "grad_norm": 0.30859678983688354, | |
| "learning_rate": 1.4142510892586113e-06, | |
| "loss": 0.0255, | |
| "step": 4490 | |
| }, | |
| { | |
| "grad_norm": 0.3854379951953888, | |
| "learning_rate": 1.3599429364716144e-06, | |
| "loss": 0.0276, | |
| "step": 4500 | |
| }, | |
| { | |
| "grad_norm": 0.2807616591453552, | |
| "learning_rate": 1.3066688763637453e-06, | |
| "loss": 0.028, | |
| "step": 4510 | |
| }, | |
| { | |
| "grad_norm": 0.364891916513443, | |
| "learning_rate": 1.2544312393149683e-06, | |
| "loss": 0.0258, | |
| "step": 4520 | |
| }, | |
| { | |
| "grad_norm": 0.8333351016044617, | |
| "learning_rate": 1.2032323103687654e-06, | |
| "loss": 0.0302, | |
| "step": 4530 | |
| }, | |
| { | |
| "grad_norm": 0.31938087940216064, | |
| "learning_rate": 1.1530743291321367e-06, | |
| "loss": 0.0255, | |
| "step": 4540 | |
| }, | |
| { | |
| "grad_norm": 0.2908027172088623, | |
| "learning_rate": 1.1039594896776834e-06, | |
| "loss": 0.0227, | |
| "step": 4550 | |
| }, | |
| { | |
| "grad_norm": 0.2906731963157654, | |
| "learning_rate": 1.0558899404475824e-06, | |
| "loss": 0.0254, | |
| "step": 4560 | |
| }, | |
| { | |
| "grad_norm": 0.4353867471218109, | |
| "learning_rate": 1.008867784159645e-06, | |
| "loss": 0.0267, | |
| "step": 4570 | |
| }, | |
| { | |
| "grad_norm": 0.30952566862106323, | |
| "learning_rate": 9.628950777153256e-07, | |
| "loss": 0.0262, | |
| "step": 4580 | |
| }, | |
| { | |
| "grad_norm": 0.7123703956604004, | |
| "learning_rate": 9.179738321097225e-07, | |
| "loss": 0.0262, | |
| "step": 4590 | |
| }, | |
| { | |
| "grad_norm": 0.4011175334453583, | |
| "learning_rate": 8.741060123436535e-07, | |
| "loss": 0.0232, | |
| "step": 4600 | |
| }, | |
| { | |
| "grad_norm": 0.305926114320755, | |
| "learning_rate": 8.312935373376573e-07, | |
| "loss": 0.0259, | |
| "step": 4610 | |
| }, | |
| { | |
| "grad_norm": 0.33336731791496277, | |
| "learning_rate": 7.895382798480927e-07, | |
| "loss": 0.0283, | |
| "step": 4620 | |
| }, | |
| { | |
| "grad_norm": 0.3118588328361511, | |
| "learning_rate": 7.488420663851859e-07, | |
| "loss": 0.0224, | |
| "step": 4630 | |
| }, | |
| { | |
| "grad_norm": 0.21298760175704956, | |
| "learning_rate": 7.092066771331507e-07, | |
| "loss": 0.0282, | |
| "step": 4640 | |
| }, | |
| { | |
| "grad_norm": 0.4557483196258545, | |
| "learning_rate": 6.706338458723205e-07, | |
| "loss": 0.0277, | |
| "step": 4650 | |
| }, | |
| { | |
| "grad_norm": 0.33916375041007996, | |
| "learning_rate": 6.331252599032833e-07, | |
| "loss": 0.0322, | |
| "step": 4660 | |
| }, | |
| { | |
| "grad_norm": 0.36580178141593933, | |
| "learning_rate": 5.96682559973108e-07, | |
| "loss": 0.0234, | |
| "step": 4670 | |
| }, | |
| { | |
| "grad_norm": 0.23574711382389069, | |
| "learning_rate": 5.613073402035429e-07, | |
| "loss": 0.0229, | |
| "step": 4680 | |
| }, | |
| { | |
| "grad_norm": 0.34600311517715454, | |
| "learning_rate": 5.270011480213056e-07, | |
| "loss": 0.0332, | |
| "step": 4690 | |
| }, | |
| { | |
| "grad_norm": 0.28487834334373474, | |
| "learning_rate": 4.937654840903722e-07, | |
| "loss": 0.0355, | |
| "step": 4700 | |
| }, | |
| { | |
| "grad_norm": 0.36917445063591003, | |
| "learning_rate": 4.6160180224635305e-07, | |
| "loss": 0.0245, | |
| "step": 4710 | |
| }, | |
| { | |
| "grad_norm": 0.2571835517883301, | |
| "learning_rate": 4.3051150943289596e-07, | |
| "loss": 0.0265, | |
| "step": 4720 | |
| }, | |
| { | |
| "grad_norm": 0.3375624716281891, | |
| "learning_rate": 4.004959656401186e-07, | |
| "loss": 0.0274, | |
| "step": 4730 | |
| }, | |
| { | |
| "grad_norm": 0.35926106572151184, | |
| "learning_rate": 3.715564838451452e-07, | |
| "loss": 0.0266, | |
| "step": 4740 | |
| }, | |
| { | |
| "grad_norm": 0.4017363488674164, | |
| "learning_rate": 3.436943299546608e-07, | |
| "loss": 0.0278, | |
| "step": 4750 | |
| }, | |
| { | |
| "grad_norm": 0.2312772423028946, | |
| "learning_rate": 3.169107227495388e-07, | |
| "loss": 0.0256, | |
| "step": 4760 | |
| }, | |
| { | |
| "grad_norm": 0.2088507115840912, | |
| "learning_rate": 2.912068338315255e-07, | |
| "loss": 0.0277, | |
| "step": 4770 | |
| }, | |
| { | |
| "grad_norm": 0.168931245803833, | |
| "learning_rate": 2.6658378757198623e-07, | |
| "loss": 0.0234, | |
| "step": 4780 | |
| }, | |
| { | |
| "grad_norm": 0.43259093165397644, | |
| "learning_rate": 2.430426610627395e-07, | |
| "loss": 0.0248, | |
| "step": 4790 | |
| }, | |
| { | |
| "grad_norm": 0.45266538858413696, | |
| "learning_rate": 2.205844840689142e-07, | |
| "loss": 0.0298, | |
| "step": 4800 | |
| }, | |
| { | |
| "grad_norm": 0.41012969613075256, | |
| "learning_rate": 1.9921023898393233e-07, | |
| "loss": 0.0258, | |
| "step": 4810 | |
| }, | |
| { | |
| "grad_norm": 0.4440428614616394, | |
| "learning_rate": 1.7892086078651616e-07, | |
| "loss": 0.021, | |
| "step": 4820 | |
| }, | |
| { | |
| "grad_norm": 0.3820282518863678, | |
| "learning_rate": 1.5971723699979013e-07, | |
| "loss": 0.0274, | |
| "step": 4830 | |
| }, | |
| { | |
| "grad_norm": 0.166753888130188, | |
| "learning_rate": 1.4160020765247862e-07, | |
| "loss": 0.0338, | |
| "step": 4840 | |
| }, | |
| { | |
| "grad_norm": 0.2630172669887543, | |
| "learning_rate": 1.2457056524212696e-07, | |
| "loss": 0.0303, | |
| "step": 4850 | |
| }, | |
| { | |
| "grad_norm": 0.28150999546051025, | |
| "learning_rate": 1.0862905470047091e-07, | |
| "loss": 0.0248, | |
| "step": 4860 | |
| }, | |
| { | |
| "grad_norm": 0.19534726440906525, | |
| "learning_rate": 9.377637336082101e-08, | |
| "loss": 0.0218, | |
| "step": 4870 | |
| }, | |
| { | |
| "grad_norm": 0.21588121354579926, | |
| "learning_rate": 8.001317092757864e-08, | |
| "loss": 0.0292, | |
| "step": 4880 | |
| }, | |
| { | |
| "grad_norm": 0.3922027349472046, | |
| "learning_rate": 6.73400494477977e-08, | |
| "loss": 0.0273, | |
| "step": 4890 | |
| }, | |
| { | |
| "grad_norm": 0.3174291253089905, | |
| "learning_rate": 5.5757563284875026e-08, | |
| "loss": 0.0276, | |
| "step": 4900 | |
| }, | |
| { | |
| "grad_norm": 0.24576455354690552, | |
| "learning_rate": 4.526621909426987e-08, | |
| "loss": 0.0259, | |
| "step": 4910 | |
| }, | |
| { | |
| "grad_norm": 0.22395384311676025, | |
| "learning_rate": 3.586647580136881e-08, | |
| "loss": 0.023, | |
| "step": 4920 | |
| }, | |
| { | |
| "grad_norm": 0.8619683980941772, | |
| "learning_rate": 2.755874458139629e-08, | |
| "loss": 0.0288, | |
| "step": 4930 | |
| }, | |
| { | |
| "grad_norm": 0.31981348991394043, | |
| "learning_rate": 2.034338884142617e-08, | |
| "loss": 0.0246, | |
| "step": 4940 | |
| }, | |
| { | |
| "grad_norm": 0.27290672063827515, | |
| "learning_rate": 1.422072420449172e-08, | |
| "loss": 0.0277, | |
| "step": 4950 | |
| }, | |
| { | |
| "grad_norm": 0.2460501492023468, | |
| "learning_rate": 9.191018495779968e-09, | |
| "loss": 0.0247, | |
| "step": 4960 | |
| }, | |
| { | |
| "grad_norm": 0.27399739623069763, | |
| "learning_rate": 5.254491730916078e-09, | |
| "loss": 0.0333, | |
| "step": 4970 | |
| }, | |
| { | |
| "grad_norm": 0.3019241988658905, | |
| "learning_rate": 2.411316106326611e-09, | |
| "loss": 0.0295, | |
| "step": 4980 | |
| }, | |
| { | |
| "grad_norm": 0.2698662281036377, | |
| "learning_rate": 6.616159917316456e-10, | |
| "loss": 0.0259, | |
| "step": 4990 | |
| }, | |
| { | |
| "grad_norm": 0.282757043838501, | |
| "learning_rate": 5.4679246769273606e-12, | |
| "loss": 0.0249, | |
| "step": 5000 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 5000, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 9223372036854775807, | |
| "save_steps": 5000, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 0.0, | |
| "train_batch_size": 32, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |