| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.0, | |
| "eval_steps": 500, | |
| "global_step": 5000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.002, | |
| "grad_norm": 0.24343955516815186, | |
| "learning_rate": 2e-08, | |
| "loss": 0.4501, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.004, | |
| "grad_norm": 0.3954734802246094, | |
| "learning_rate": 4e-08, | |
| "loss": 0.5889, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.006, | |
| "grad_norm": 0.24582786858081818, | |
| "learning_rate": 6e-08, | |
| "loss": 0.7849, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.008, | |
| "grad_norm": 0.4748157560825348, | |
| "learning_rate": 8e-08, | |
| "loss": 0.3773, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 0.22575026750564575, | |
| "learning_rate": 1e-07, | |
| "loss": 0.5379, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.012, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.2e-07, | |
| "loss": 0.5542, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.014, | |
| "grad_norm": 0.5761760473251343, | |
| "learning_rate": 1.4e-07, | |
| "loss": 0.6973, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.016, | |
| "grad_norm": 0.5927178263664246, | |
| "learning_rate": 1.6e-07, | |
| "loss": 0.674, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.018, | |
| "grad_norm": 0.3376465439796448, | |
| "learning_rate": 1.8e-07, | |
| "loss": 0.4789, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 0.3363400995731354, | |
| "learning_rate": 2e-07, | |
| "loss": 0.518, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.022, | |
| "grad_norm": 1.267148494720459, | |
| "learning_rate": 2.1999999999999998e-07, | |
| "loss": 0.7268, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.024, | |
| "grad_norm": 0.4405953288078308, | |
| "learning_rate": 2.4e-07, | |
| "loss": 0.5595, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.026, | |
| "grad_norm": 0.6260291934013367, | |
| "learning_rate": 2.6e-07, | |
| "loss": 0.6663, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.028, | |
| "grad_norm": 0.16322936117649078, | |
| "learning_rate": 2.8e-07, | |
| "loss": 0.5629, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 0.5333911776542664, | |
| "learning_rate": 3e-07, | |
| "loss": 0.7167, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.032, | |
| "grad_norm": 0.38673460483551025, | |
| "learning_rate": 3.2e-07, | |
| "loss": 0.4412, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.034, | |
| "grad_norm": 0.5723925828933716, | |
| "learning_rate": 3.4000000000000003e-07, | |
| "loss": 0.5445, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.036, | |
| "grad_norm": 0.3431761860847473, | |
| "learning_rate": 3.6e-07, | |
| "loss": 0.5964, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.038, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.7999999999999996e-07, | |
| "loss": 0.5541, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4e-07, | |
| "loss": 0.6355, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.042, | |
| "grad_norm": 0.5891884565353394, | |
| "learning_rate": 4.1999999999999995e-07, | |
| "loss": 0.5136, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.044, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.3999999999999997e-07, | |
| "loss": 0.4751, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.046, | |
| "grad_norm": 0.28812310099601746, | |
| "learning_rate": 4.6e-07, | |
| "loss": 0.524, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.048, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.8e-07, | |
| "loss": 0.2113, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 0.5220464468002319, | |
| "learning_rate": 5e-07, | |
| "loss": 0.6769, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.052, | |
| "grad_norm": 0.21842537820339203, | |
| "learning_rate": 5.2e-07, | |
| "loss": 0.6747, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.054, | |
| "grad_norm": 0.8564176559448242, | |
| "learning_rate": 5.4e-07, | |
| "loss": 0.5687, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.056, | |
| "grad_norm": 0.3472450375556946, | |
| "learning_rate": 5.6e-07, | |
| "loss": 0.4823, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.058, | |
| "grad_norm": 0.3166986107826233, | |
| "learning_rate": 5.8e-07, | |
| "loss": 0.5185, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 0.4498099088668823, | |
| "learning_rate": 6e-07, | |
| "loss": 0.5141, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.062, | |
| "grad_norm": 0.3302588164806366, | |
| "learning_rate": 6.2e-07, | |
| "loss": 0.6248, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.064, | |
| "grad_norm": 0.32614701986312866, | |
| "learning_rate": 6.4e-07, | |
| "loss": 0.4848, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.066, | |
| "grad_norm": 0.3647592067718506, | |
| "learning_rate": 6.6e-07, | |
| "loss": 0.5498, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.068, | |
| "grad_norm": 0.5517160892486572, | |
| "learning_rate": 6.800000000000001e-07, | |
| "loss": 0.711, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 0.43151068687438965, | |
| "learning_rate": 7e-07, | |
| "loss": 0.527, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.072, | |
| "grad_norm": 0.46171778440475464, | |
| "learning_rate": 7.2e-07, | |
| "loss": 0.465, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.074, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.4e-07, | |
| "loss": 0.6132, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.076, | |
| "grad_norm": 0.34477466344833374, | |
| "learning_rate": 7.599999999999999e-07, | |
| "loss": 0.5284, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.078, | |
| "grad_norm": 1.2470405101776123, | |
| "learning_rate": 7.799999999999999e-07, | |
| "loss": 0.6293, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 0.3687979578971863, | |
| "learning_rate": 8e-07, | |
| "loss": 0.5045, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.082, | |
| "grad_norm": 0.4226088225841522, | |
| "learning_rate": 8.199999999999999e-07, | |
| "loss": 0.6637, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.084, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.399999999999999e-07, | |
| "loss": 0.7251, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.086, | |
| "grad_norm": 0.9228200316429138, | |
| "learning_rate": 8.599999999999999e-07, | |
| "loss": 0.6431, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.088, | |
| "grad_norm": 0.2559778690338135, | |
| "learning_rate": 8.799999999999999e-07, | |
| "loss": 0.4074, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 0.7684836983680725, | |
| "learning_rate": 9e-07, | |
| "loss": 0.5902, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.092, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.2e-07, | |
| "loss": 0.4689, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.094, | |
| "grad_norm": 0.3533630669116974, | |
| "learning_rate": 9.399999999999999e-07, | |
| "loss": 0.7271, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.096, | |
| "grad_norm": 0.6006487011909485, | |
| "learning_rate": 9.6e-07, | |
| "loss": 0.5598, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.098, | |
| "grad_norm": 0.9538562893867493, | |
| "learning_rate": 9.8e-07, | |
| "loss": 0.8427, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 0.6611864566802979, | |
| "learning_rate": 1e-06, | |
| "loss": 0.5613, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.102, | |
| "grad_norm": 0.44144096970558167, | |
| "learning_rate": 9.999878153526972e-07, | |
| "loss": 0.2706, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.104, | |
| "grad_norm": 0.437011182308197, | |
| "learning_rate": 9.99951262004652e-07, | |
| "loss": 0.7309, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.106, | |
| "grad_norm": 0.49905768036842346, | |
| "learning_rate": 9.998903417374226e-07, | |
| "loss": 0.5009, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.108, | |
| "grad_norm": 0.3730575442314148, | |
| "learning_rate": 9.99805057520177e-07, | |
| "loss": 0.5414, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.996954135095478e-07, | |
| "loss": 0.3806, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.112, | |
| "grad_norm": 0.717401385307312, | |
| "learning_rate": 9.99561415049429e-07, | |
| "loss": 0.574, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.114, | |
| "grad_norm": 0.4745548963546753, | |
| "learning_rate": 9.99403068670717e-07, | |
| "loss": 0.422, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.116, | |
| "grad_norm": 0.5111886858940125, | |
| "learning_rate": 9.992203820909905e-07, | |
| "loss": 0.4709, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.118, | |
| "grad_norm": 0.4847474694252014, | |
| "learning_rate": 9.990133642141357e-07, | |
| "loss": 0.5755, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 0.5600196123123169, | |
| "learning_rate": 9.98782025129912e-07, | |
| "loss": 0.6245, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.122, | |
| "grad_norm": 0.4002205729484558, | |
| "learning_rate": 9.9852637611346e-07, | |
| "loss": 0.5492, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.124, | |
| "grad_norm": 0.25081363320350647, | |
| "learning_rate": 9.982464296247522e-07, | |
| "loss": 0.471, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.126, | |
| "grad_norm": 0.371477872133255, | |
| "learning_rate": 9.97942199307985e-07, | |
| "loss": 0.6561, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.128, | |
| "grad_norm": 0.5661497116088867, | |
| "learning_rate": 9.976136999909155e-07, | |
| "loss": 0.5399, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 0.41831666231155396, | |
| "learning_rate": 9.972609476841365e-07, | |
| "loss": 0.6655, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.132, | |
| "grad_norm": 0.27902722358703613, | |
| "learning_rate": 9.968839595802981e-07, | |
| "loss": 0.4324, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.134, | |
| "grad_norm": 0.3698316216468811, | |
| "learning_rate": 9.964827540532684e-07, | |
| "loss": 0.3613, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.136, | |
| "grad_norm": 0.2788097560405731, | |
| "learning_rate": 9.960573506572389e-07, | |
| "loss": 0.4812, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.138, | |
| "grad_norm": 0.5648425221443176, | |
| "learning_rate": 9.956077701257707e-07, | |
| "loss": 0.5351, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 0.48793095350265503, | |
| "learning_rate": 9.95134034370785e-07, | |
| "loss": 0.6143, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.142, | |
| "grad_norm": 0.4436402916908264, | |
| "learning_rate": 9.946361664814943e-07, | |
| "loss": 0.6084, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.144, | |
| "grad_norm": 0.6165416836738586, | |
| "learning_rate": 9.941141907232763e-07, | |
| "loss": 0.5375, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.146, | |
| "grad_norm": 0.40755054354667664, | |
| "learning_rate": 9.93568132536494e-07, | |
| "loss": 0.422, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.148, | |
| "grad_norm": 0.3307097852230072, | |
| "learning_rate": 9.929980185352525e-07, | |
| "loss": 0.6054, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 1.1209286451339722, | |
| "learning_rate": 9.92403876506104e-07, | |
| "loss": 0.6197, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.152, | |
| "grad_norm": 0.28426626324653625, | |
| "learning_rate": 9.91785735406693e-07, | |
| "loss": 0.6463, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.154, | |
| "grad_norm": 0.7093852758407593, | |
| "learning_rate": 9.911436253643443e-07, | |
| "loss": 0.5661, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.156, | |
| "grad_norm": 0.6072311401367188, | |
| "learning_rate": 9.904775776745956e-07, | |
| "loss": 0.5019, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.158, | |
| "grad_norm": 1.8654353618621826, | |
| "learning_rate": 9.89787624799672e-07, | |
| "loss": 0.6612, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 0.45068666338920593, | |
| "learning_rate": 9.890738003669027e-07, | |
| "loss": 0.4475, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.162, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.883361391670839e-07, | |
| "loss": 0.3523, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.164, | |
| "grad_norm": 0.37118417024612427, | |
| "learning_rate": 9.875746771527815e-07, | |
| "loss": 0.4543, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.166, | |
| "grad_norm": 0.3821951746940613, | |
| "learning_rate": 9.8678945143658e-07, | |
| "loss": 0.5156, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.168, | |
| "grad_norm": 0.5337674021720886, | |
| "learning_rate": 9.859805002892731e-07, | |
| "loss": 0.4811, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 0.5889901518821716, | |
| "learning_rate": 9.851478631379982e-07, | |
| "loss": 0.6555, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.172, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.842915805643156e-07, | |
| "loss": 0.3551, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.174, | |
| "grad_norm": 0.24697642028331757, | |
| "learning_rate": 9.834116943022297e-07, | |
| "loss": 0.4373, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.176, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.825082472361556e-07, | |
| "loss": 0.4122, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.178, | |
| "grad_norm": 0.4018269181251526, | |
| "learning_rate": 9.81581283398829e-07, | |
| "loss": 0.3821, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 0.283811092376709, | |
| "learning_rate": 9.806308479691594e-07, | |
| "loss": 0.4605, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.182, | |
| "grad_norm": 0.5428382158279419, | |
| "learning_rate": 9.796569872700287e-07, | |
| "loss": 0.5417, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.184, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.786597487660335e-07, | |
| "loss": 0.3459, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.186, | |
| "grad_norm": 0.4676339626312256, | |
| "learning_rate": 9.776391810611718e-07, | |
| "loss": 0.4893, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.188, | |
| "grad_norm": 0.5452865958213806, | |
| "learning_rate": 9.765953338964734e-07, | |
| "loss": 0.3598, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 0.3055381178855896, | |
| "learning_rate": 9.755282581475767e-07, | |
| "loss": 0.3582, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.192, | |
| "grad_norm": 0.6679813861846924, | |
| "learning_rate": 9.744380058222482e-07, | |
| "loss": 0.571, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.194, | |
| "grad_norm": 0.534697949886322, | |
| "learning_rate": 9.733246300578482e-07, | |
| "loss": 0.4952, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.196, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.721881851187405e-07, | |
| "loss": 0.4123, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.198, | |
| "grad_norm": 0.6762744784355164, | |
| "learning_rate": 9.710287263936483e-07, | |
| "loss": 0.4716, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 0.6695389151573181, | |
| "learning_rate": 9.698463103929541e-07, | |
| "loss": 0.4759, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.202, | |
| "grad_norm": 0.2647770941257477, | |
| "learning_rate": 9.686409947459457e-07, | |
| "loss": 0.5234, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.204, | |
| "grad_norm": 0.710737943649292, | |
| "learning_rate": 9.674128381980071e-07, | |
| "loss": 0.4103, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.206, | |
| "grad_norm": 0.7678632140159607, | |
| "learning_rate": 9.661619006077561e-07, | |
| "loss": 0.4646, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.208, | |
| "grad_norm": 0.41716259717941284, | |
| "learning_rate": 9.648882429441256e-07, | |
| "loss": 0.5082, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.635919272833937e-07, | |
| "loss": 0.3218, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.212, | |
| "grad_norm": 0.5014677047729492, | |
| "learning_rate": 9.622730168061567e-07, | |
| "loss": 0.5026, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.214, | |
| "grad_norm": 0.5589115619659424, | |
| "learning_rate": 9.609315757942502e-07, | |
| "loss": 0.5518, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.216, | |
| "grad_norm": 0.32412189245224, | |
| "learning_rate": 9.595676696276171e-07, | |
| "loss": 0.6244, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.218, | |
| "grad_norm": 0.48868948221206665, | |
| "learning_rate": 9.581813647811197e-07, | |
| "loss": 0.4662, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 1.1036161184310913, | |
| "learning_rate": 9.567727288213004e-07, | |
| "loss": 0.5146, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.222, | |
| "grad_norm": 0.6447727084159851, | |
| "learning_rate": 9.553418304030885e-07, | |
| "loss": 0.4169, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.224, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.538887392664543e-07, | |
| "loss": 0.3333, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.226, | |
| "grad_norm": 0.2741583585739136, | |
| "learning_rate": 9.524135262330098e-07, | |
| "loss": 0.5284, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.228, | |
| "grad_norm": 0.35744038224220276, | |
| "learning_rate": 9.509162632025569e-07, | |
| "loss": 0.5026, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 0.5547361969947815, | |
| "learning_rate": 9.493970231495834e-07, | |
| "loss": 0.5528, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.232, | |
| "grad_norm": 0.7139315009117126, | |
| "learning_rate": 9.478558801197064e-07, | |
| "loss": 0.5158, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.234, | |
| "grad_norm": 0.47351908683776855, | |
| "learning_rate": 9.462929092260628e-07, | |
| "loss": 0.4882, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.236, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.447081866456487e-07, | |
| "loss": 0.4078, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.238, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.431017896156073e-07, | |
| "loss": 0.3839, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 1.2698532342910767, | |
| "learning_rate": 9.414737964294634e-07, | |
| "loss": 0.4445, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.242, | |
| "grad_norm": 0.3917461037635803, | |
| "learning_rate": 9.398242864333083e-07, | |
| "loss": 0.4356, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 0.244, | |
| "grad_norm": 0.7485201954841614, | |
| "learning_rate": 9.381533400219317e-07, | |
| "loss": 0.5457, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.246, | |
| "grad_norm": 0.6434317827224731, | |
| "learning_rate": 9.364610386349047e-07, | |
| "loss": 0.4488, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.248, | |
| "grad_norm": 0.7366912364959717, | |
| "learning_rate": 9.347474647526095e-07, | |
| "loss": 0.5612, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 0.5144039392471313, | |
| "learning_rate": 9.330127018922193e-07, | |
| "loss": 0.3663, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.252, | |
| "grad_norm": 1.0590194463729858, | |
| "learning_rate": 9.312568346036287e-07, | |
| "loss": 0.5273, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.254, | |
| "grad_norm": 0.517156720161438, | |
| "learning_rate": 9.294799484653322e-07, | |
| "loss": 0.5462, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 0.256, | |
| "grad_norm": 0.7070313096046448, | |
| "learning_rate": 9.276821300802533e-07, | |
| "loss": 0.3943, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.258, | |
| "grad_norm": 0.39627718925476074, | |
| "learning_rate": 9.258634670715237e-07, | |
| "loss": 0.3459, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 0.28623461723327637, | |
| "learning_rate": 9.240240480782129e-07, | |
| "loss": 0.4004, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.262, | |
| "grad_norm": 0.2795998156070709, | |
| "learning_rate": 9.221639627510075e-07, | |
| "loss": 0.3504, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 0.264, | |
| "grad_norm": 0.5013036727905273, | |
| "learning_rate": 9.202833017478421e-07, | |
| "loss": 0.4138, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.266, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.183821567294808e-07, | |
| "loss": 0.317, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 0.268, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.164606203550497e-07, | |
| "loss": 0.3859, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 0.2559995651245117, | |
| "learning_rate": 9.145187862775208e-07, | |
| "loss": 0.4547, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.272, | |
| "grad_norm": 1.344838261604309, | |
| "learning_rate": 9.125567491391475e-07, | |
| "loss": 0.439, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.274, | |
| "grad_norm": 0.5493496060371399, | |
| "learning_rate": 9.10574604566852e-07, | |
| "loss": 0.3391, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 0.276, | |
| "grad_norm": 2.435943841934204, | |
| "learning_rate": 9.085724491675642e-07, | |
| "loss": 0.4042, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.278, | |
| "grad_norm": 0.49110209941864014, | |
| "learning_rate": 9.065503805235137e-07, | |
| "loss": 0.4792, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 0.3752264976501465, | |
| "learning_rate": 9.045084971874737e-07, | |
| "loss": 0.3786, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.282, | |
| "grad_norm": 1.7920351028442383, | |
| "learning_rate": 9.02446898677957e-07, | |
| "loss": 0.4598, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 0.284, | |
| "grad_norm": 0.4336905777454376, | |
| "learning_rate": 9.003656854743666e-07, | |
| "loss": 0.3466, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.286, | |
| "grad_norm": 0.4503847658634186, | |
| "learning_rate": 8.982649590120981e-07, | |
| "loss": 0.4454, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 0.288, | |
| "grad_norm": 1.1344330310821533, | |
| "learning_rate": 8.961448216775953e-07, | |
| "loss": 0.4648, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 0.71676105260849, | |
| "learning_rate": 8.940053768033608e-07, | |
| "loss": 0.3969, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.292, | |
| "grad_norm": 0.734315037727356, | |
| "learning_rate": 8.918467286629198e-07, | |
| "loss": 0.3435, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.294, | |
| "grad_norm": 0.36458420753479004, | |
| "learning_rate": 8.896689824657371e-07, | |
| "loss": 0.3545, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 0.296, | |
| "grad_norm": 0.4997411072254181, | |
| "learning_rate": 8.874722443520898e-07, | |
| "loss": 0.2999, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.298, | |
| "grad_norm": 0.3361949026584625, | |
| "learning_rate": 8.852566213878946e-07, | |
| "loss": 0.2811, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 0.30942943692207336, | |
| "learning_rate": 8.83022221559489e-07, | |
| "loss": 0.6193, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.302, | |
| "grad_norm": 0.40158674120903015, | |
| "learning_rate": 8.807691537683684e-07, | |
| "loss": 0.5567, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 0.304, | |
| "grad_norm": 0.40954452753067017, | |
| "learning_rate": 8.784975278258782e-07, | |
| "loss": 0.3944, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.306, | |
| "grad_norm": 0.8099032044410706, | |
| "learning_rate": 8.762074544478621e-07, | |
| "loss": 0.4308, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 0.308, | |
| "grad_norm": 0.6486494541168213, | |
| "learning_rate": 8.73899045249266e-07, | |
| "loss": 0.3566, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 0.6595824956893921, | |
| "learning_rate": 8.71572412738697e-07, | |
| "loss": 0.3783, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.312, | |
| "grad_norm": 1.0239789485931396, | |
| "learning_rate": 8.69227670312942e-07, | |
| "loss": 0.4704, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.314, | |
| "grad_norm": 0.4185124933719635, | |
| "learning_rate": 8.668649322514381e-07, | |
| "loss": 0.2514, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 0.316, | |
| "grad_norm": 0.7972500324249268, | |
| "learning_rate": 8.644843137107057e-07, | |
| "loss": 0.3748, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.318, | |
| "grad_norm": 0.8556084632873535, | |
| "learning_rate": 8.620859307187338e-07, | |
| "loss": 0.3196, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 0.5393433570861816, | |
| "learning_rate": 8.596699001693255e-07, | |
| "loss": 0.261, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.322, | |
| "grad_norm": 0.6654929518699646, | |
| "learning_rate": 8.572363398164016e-07, | |
| "loss": 0.318, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 0.324, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.547853682682604e-07, | |
| "loss": 0.3533, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.326, | |
| "grad_norm": 0.3704928755760193, | |
| "learning_rate": 8.523171049817973e-07, | |
| "loss": 0.3844, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 0.328, | |
| "grad_norm": 0.6798892021179199, | |
| "learning_rate": 8.498316702566826e-07, | |
| "loss": 0.3807, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 0.721932053565979, | |
| "learning_rate": 8.473291852294986e-07, | |
| "loss": 0.3674, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.332, | |
| "grad_norm": 0.6556898951530457, | |
| "learning_rate": 8.448097718678348e-07, | |
| "loss": 0.3965, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.334, | |
| "grad_norm": 0.7529492974281311, | |
| "learning_rate": 8.422735529643443e-07, | |
| "loss": 0.2847, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 0.336, | |
| "grad_norm": 0.2707422077655792, | |
| "learning_rate": 8.397206521307583e-07, | |
| "loss": 0.177, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.338, | |
| "grad_norm": 0.6421071290969849, | |
| "learning_rate": 8.371511937918617e-07, | |
| "loss": 0.3504, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 0.43531012535095215, | |
| "learning_rate": 8.34565303179429e-07, | |
| "loss": 0.3154, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.342, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.319631063261207e-07, | |
| "loss": 0.2684, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 0.344, | |
| "grad_norm": 0.4423937201499939, | |
| "learning_rate": 8.293447300593402e-07, | |
| "loss": 0.3543, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 0.346, | |
| "grad_norm": 1.126133680343628, | |
| "learning_rate": 8.267103019950528e-07, | |
| "loss": 0.2968, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 0.348, | |
| "grad_norm": 1.135288119316101, | |
| "learning_rate": 8.240599505315654e-07, | |
| "loss": 0.3884, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 0.3778904676437378, | |
| "learning_rate": 8.213938048432696e-07, | |
| "loss": 0.3483, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.352, | |
| "grad_norm": 1.2009609937667847, | |
| "learning_rate": 8.187119948743449e-07, | |
| "loss": 0.308, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 0.354, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.160146513324254e-07, | |
| "loss": 0.3573, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 0.356, | |
| "grad_norm": 0.7662010192871094, | |
| "learning_rate": 8.133019056822302e-07, | |
| "loss": 0.3884, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 0.358, | |
| "grad_norm": 0.6717849373817444, | |
| "learning_rate": 8.105738901391551e-07, | |
| "loss": 0.2845, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.07830737662829e-07, | |
| "loss": 0.3274, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.362, | |
| "grad_norm": 0.6610533595085144, | |
| "learning_rate": 8.050725819506339e-07, | |
| "loss": 0.2492, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 0.364, | |
| "grad_norm": 0.6068025827407837, | |
| "learning_rate": 8.022995574311875e-07, | |
| "loss": 0.3101, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 0.366, | |
| "grad_norm": 1.280325174331665, | |
| "learning_rate": 7.995117992577928e-07, | |
| "loss": 0.4106, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 0.368, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.967094433018508e-07, | |
| "loss": 0.2845, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 0.5306461453437805, | |
| "learning_rate": 7.938926261462365e-07, | |
| "loss": 0.3267, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 0.372, | |
| "grad_norm": 0.4115411341190338, | |
| "learning_rate": 7.910614850786447e-07, | |
| "loss": 0.4351, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 0.374, | |
| "grad_norm": 0.5598284006118774, | |
| "learning_rate": 7.882161580848966e-07, | |
| "loss": 0.3982, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 0.376, | |
| "grad_norm": 0.4399189054965973, | |
| "learning_rate": 7.853567838422159e-07, | |
| "loss": 0.3284, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 0.378, | |
| "grad_norm": 0.4055747091770172, | |
| "learning_rate": 7.82483501712469e-07, | |
| "loss": 0.2678, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 0.5287490487098694, | |
| "learning_rate": 7.795964517353733e-07, | |
| "loss": 0.223, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.382, | |
| "grad_norm": 0.6047921180725098, | |
| "learning_rate": 7.76695774621672e-07, | |
| "loss": 0.3105, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 0.384, | |
| "grad_norm": 0.7903981804847717, | |
| "learning_rate": 7.737816117462751e-07, | |
| "loss": 0.3185, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 0.386, | |
| "grad_norm": 0.673017680644989, | |
| "learning_rate": 7.7085410514137e-07, | |
| "loss": 0.2129, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 0.388, | |
| "grad_norm": 0.37061551213264465, | |
| "learning_rate": 7.679133974894982e-07, | |
| "loss": 0.3608, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 0.571678638458252, | |
| "learning_rate": 7.649596321166024e-07, | |
| "loss": 0.1734, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 0.392, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.619929529850396e-07, | |
| "loss": 0.2466, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 0.394, | |
| "grad_norm": 0.8853337168693542, | |
| "learning_rate": 7.590135046865651e-07, | |
| "loss": 0.248, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 0.396, | |
| "grad_norm": 0.5480286478996277, | |
| "learning_rate": 7.560214324352858e-07, | |
| "loss": 0.2524, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 0.398, | |
| "grad_norm": 0.48180562257766724, | |
| "learning_rate": 7.530168820605818e-07, | |
| "loss": 0.3421, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 1.1237231492996216, | |
| "learning_rate": 7.5e-07, | |
| "loss": 0.2916, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.402, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.469709332921154e-07, | |
| "loss": 0.2692, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 0.404, | |
| "grad_norm": 1.1030820608139038, | |
| "learning_rate": 7.439298295693663e-07, | |
| "loss": 0.2692, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 0.406, | |
| "grad_norm": 0.7971144914627075, | |
| "learning_rate": 7.408768370508576e-07, | |
| "loss": 0.293, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 0.408, | |
| "grad_norm": 0.9243513941764832, | |
| "learning_rate": 7.378121045351377e-07, | |
| "loss": 0.2873, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 0.2981295585632324, | |
| "learning_rate": 7.347357813929454e-07, | |
| "loss": 0.3581, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 0.412, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.316480175599308e-07, | |
| "loss": 0.2246, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 0.414, | |
| "grad_norm": 0.5026867985725403, | |
| "learning_rate": 7.285489635293471e-07, | |
| "loss": 0.2117, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 0.416, | |
| "grad_norm": 0.5733957886695862, | |
| "learning_rate": 7.254387703447153e-07, | |
| "loss": 0.2838, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 0.418, | |
| "grad_norm": 0.4224608838558197, | |
| "learning_rate": 7.223175895924637e-07, | |
| "loss": 0.2593, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 0.3539625406265259, | |
| "learning_rate": 7.191855733945386e-07, | |
| "loss": 0.2341, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.422, | |
| "grad_norm": 2.312608003616333, | |
| "learning_rate": 7.160428744009912e-07, | |
| "loss": 0.3711, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 0.424, | |
| "grad_norm": 0.5831490755081177, | |
| "learning_rate": 7.128896457825363e-07, | |
| "loss": 0.345, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 0.426, | |
| "grad_norm": 0.268442302942276, | |
| "learning_rate": 7.097260412230885e-07, | |
| "loss": 0.2471, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 0.428, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.065522149122709e-07, | |
| "loss": 0.2666, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 0.3480328619480133, | |
| "learning_rate": 7.033683215379002e-07, | |
| "loss": 0.2101, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 0.432, | |
| "grad_norm": 1.392856240272522, | |
| "learning_rate": 7.001745162784475e-07, | |
| "loss": 0.3794, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 0.434, | |
| "grad_norm": 0.5403885245323181, | |
| "learning_rate": 6.969709547954755e-07, | |
| "loss": 0.4291, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 0.436, | |
| "grad_norm": 0.6326515078544617, | |
| "learning_rate": 6.937577932260514e-07, | |
| "loss": 0.2617, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 0.438, | |
| "grad_norm": 0.4600224792957306, | |
| "learning_rate": 6.905351881751371e-07, | |
| "loss": 0.2905, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 0.3074401617050171, | |
| "learning_rate": 6.87303296707956e-07, | |
| "loss": 0.1947, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.442, | |
| "grad_norm": 0.6525133848190308, | |
| "learning_rate": 6.840622763423391e-07, | |
| "loss": 0.311, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 0.444, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.80812285041046e-07, | |
| "loss": 0.2794, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 0.446, | |
| "grad_norm": 0.31862539052963257, | |
| "learning_rate": 6.775534812040686e-07, | |
| "loss": 0.2453, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 0.448, | |
| "grad_norm": 0.36761975288391113, | |
| "learning_rate": 6.742860236609076e-07, | |
| "loss": 0.3143, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 0.4649720788002014, | |
| "learning_rate": 6.710100716628344e-07, | |
| "loss": 0.2906, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 0.452, | |
| "grad_norm": 0.9335395693778992, | |
| "learning_rate": 6.677257848751276e-07, | |
| "loss": 0.2104, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 0.454, | |
| "grad_norm": 0.48736828565597534, | |
| "learning_rate": 6.644333233692916e-07, | |
| "loss": 0.2243, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 0.456, | |
| "grad_norm": 0.3455299437046051, | |
| "learning_rate": 6.611328476152556e-07, | |
| "loss": 0.2485, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 0.458, | |
| "grad_norm": 0.5446789264678955, | |
| "learning_rate": 6.578245184735512e-07, | |
| "loss": 0.2834, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 0.5425819754600525, | |
| "learning_rate": 6.545084971874736e-07, | |
| "loss": 0.2917, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.462, | |
| "grad_norm": 0.4187966585159302, | |
| "learning_rate": 6.511849453752223e-07, | |
| "loss": 0.287, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 0.464, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.478540250220233e-07, | |
| "loss": 0.2607, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 0.466, | |
| "grad_norm": 0.7009822726249695, | |
| "learning_rate": 6.445158984722358e-07, | |
| "loss": 0.241, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 0.468, | |
| "grad_norm": 0.3654937744140625, | |
| "learning_rate": 6.411707284214383e-07, | |
| "loss": 0.2763, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.378186779084995e-07, | |
| "loss": 0.3021, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 0.472, | |
| "grad_norm": 1.470301866531372, | |
| "learning_rate": 6.344599103076328e-07, | |
| "loss": 0.2751, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 0.474, | |
| "grad_norm": 0.9141136407852173, | |
| "learning_rate": 6.310945893204324e-07, | |
| "loss": 0.3617, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 0.476, | |
| "grad_norm": 0.40913861989974976, | |
| "learning_rate": 6.277228789678953e-07, | |
| "loss": 0.2258, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 0.478, | |
| "grad_norm": 0.4026692807674408, | |
| "learning_rate": 6.243449435824276e-07, | |
| "loss": 0.3555, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 0.5284442901611328, | |
| "learning_rate": 6.209609477998338e-07, | |
| "loss": 0.2222, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.482, | |
| "grad_norm": 1.552199125289917, | |
| "learning_rate": 6.17571056551295e-07, | |
| "loss": 0.2444, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 0.484, | |
| "grad_norm": 0.45044296979904175, | |
| "learning_rate": 6.141754350553279e-07, | |
| "loss": 0.3037, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 0.486, | |
| "grad_norm": 0.36123940348625183, | |
| "learning_rate": 6.107742488097338e-07, | |
| "loss": 0.2412, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 0.488, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.073676635835316e-07, | |
| "loss": 0.2901, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 0.0, | |
| "learning_rate": 6.039558454088795e-07, | |
| "loss": 0.2803, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 0.492, | |
| "grad_norm": 0.535737156867981, | |
| "learning_rate": 6.005389605729824e-07, | |
| "loss": 0.2778, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 0.494, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.97117175609986e-07, | |
| "loss": 0.198, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 0.496, | |
| "grad_norm": 0.4263180196285248, | |
| "learning_rate": 5.936906572928624e-07, | |
| "loss": 0.2764, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 0.498, | |
| "grad_norm": 0.4536767303943634, | |
| "learning_rate": 5.9025957262528e-07, | |
| "loss": 0.2533, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 0.6187867522239685, | |
| "learning_rate": 5.868240888334652e-07, | |
| "loss": 0.3028, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.502, | |
| "grad_norm": 0.526470959186554, | |
| "learning_rate": 5.833843733580512e-07, | |
| "loss": 0.2654, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 0.504, | |
| "grad_norm": 0.7789123058319092, | |
| "learning_rate": 5.799405938459174e-07, | |
| "loss": 0.1738, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 0.506, | |
| "grad_norm": 0.8183090090751648, | |
| "learning_rate": 5.764929181420191e-07, | |
| "loss": 0.2594, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 0.508, | |
| "grad_norm": 0.32851341366767883, | |
| "learning_rate": 5.730415142812058e-07, | |
| "loss": 0.2654, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 1.3538209199905396, | |
| "learning_rate": 5.695865504800327e-07, | |
| "loss": 0.3137, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 0.512, | |
| "grad_norm": 0.5997196435928345, | |
| "learning_rate": 5.661281951285612e-07, | |
| "loss": 0.2907, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 0.514, | |
| "grad_norm": 0.7326313257217407, | |
| "learning_rate": 5.626666167821521e-07, | |
| "loss": 0.2786, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 0.516, | |
| "grad_norm": 0.32429754734039307, | |
| "learning_rate": 5.592019841532506e-07, | |
| "loss": 0.2376, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 0.518, | |
| "grad_norm": 0.9618094563484192, | |
| "learning_rate": 5.557344661031627e-07, | |
| "loss": 0.2353, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 0.7093096375465393, | |
| "learning_rate": 5.522642316338268e-07, | |
| "loss": 0.2908, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.522, | |
| "grad_norm": 0.6049932241439819, | |
| "learning_rate": 5.487914498795747e-07, | |
| "loss": 0.1842, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 0.524, | |
| "grad_norm": 0.7540631294250488, | |
| "learning_rate": 5.453162900988901e-07, | |
| "loss": 0.2837, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 0.526, | |
| "grad_norm": 0.31856876611709595, | |
| "learning_rate": 5.418389216661578e-07, | |
| "loss": 0.2362, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 0.528, | |
| "grad_norm": 0.3942410349845886, | |
| "learning_rate": 5.383595140634093e-07, | |
| "loss": 0.2118, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 0.71694016456604, | |
| "learning_rate": 5.348782368720625e-07, | |
| "loss": 0.3198, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 0.532, | |
| "grad_norm": 0.6586489081382751, | |
| "learning_rate": 5.313952597646567e-07, | |
| "loss": 0.3122, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 0.534, | |
| "grad_norm": 0.8461518883705139, | |
| "learning_rate": 5.27910752496582e-07, | |
| "loss": 0.2936, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 0.536, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.244248848978067e-07, | |
| "loss": 0.2448, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 0.538, | |
| "grad_norm": 0.5395662188529968, | |
| "learning_rate": 5.209378268645997e-07, | |
| "loss": 0.2252, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 0.9108349680900574, | |
| "learning_rate": 5.174497483512505e-07, | |
| "loss": 0.3459, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.542, | |
| "grad_norm": 0.34403327107429504, | |
| "learning_rate": 5.139608193617844e-07, | |
| "loss": 0.2264, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 0.544, | |
| "grad_norm": 0.4753933846950531, | |
| "learning_rate": 5.104712099416785e-07, | |
| "loss": 0.2055, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 0.546, | |
| "grad_norm": 0.39978426694869995, | |
| "learning_rate": 5.069810901695727e-07, | |
| "loss": 0.2603, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 0.548, | |
| "grad_norm": 0.48148438334465027, | |
| "learning_rate": 5.034906301489807e-07, | |
| "loss": 0.1868, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 0.4292818307876587, | |
| "learning_rate": 5e-07, | |
| "loss": 0.2567, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 0.552, | |
| "grad_norm": 0.8320960402488708, | |
| "learning_rate": 4.965093698510192e-07, | |
| "loss": 0.275, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 0.554, | |
| "grad_norm": 0.4082556366920471, | |
| "learning_rate": 4.930189098304274e-07, | |
| "loss": 0.2878, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 0.556, | |
| "grad_norm": 0.7608616352081299, | |
| "learning_rate": 4.895287900583216e-07, | |
| "loss": 0.2047, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 0.558, | |
| "grad_norm": 0.3502591550350189, | |
| "learning_rate": 4.860391806382156e-07, | |
| "loss": 0.2268, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 0.800480306148529, | |
| "learning_rate": 4.825502516487496e-07, | |
| "loss": 0.2261, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.562, | |
| "grad_norm": 0.35348179936408997, | |
| "learning_rate": 4.790621731354002e-07, | |
| "loss": 0.2703, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 0.564, | |
| "grad_norm": 0.42046764492988586, | |
| "learning_rate": 4.7557511510219335e-07, | |
| "loss": 0.1906, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 0.566, | |
| "grad_norm": 0.6101659536361694, | |
| "learning_rate": 4.7208924750341805e-07, | |
| "loss": 0.2274, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 0.568, | |
| "grad_norm": 0.9455362558364868, | |
| "learning_rate": 4.686047402353433e-07, | |
| "loss": 0.2739, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.6512176312793735e-07, | |
| "loss": 0.1789, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 0.572, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.6164048593659065e-07, | |
| "loss": 0.201, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 0.574, | |
| "grad_norm": 0.5131359100341797, | |
| "learning_rate": 4.5816107833384233e-07, | |
| "loss": 0.2712, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 0.576, | |
| "grad_norm": 0.32932206988334656, | |
| "learning_rate": 4.5468370990110997e-07, | |
| "loss": 0.1617, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 0.578, | |
| "grad_norm": 1.0272895097732544, | |
| "learning_rate": 4.512085501204253e-07, | |
| "loss": 0.274, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 0.4852657616138458, | |
| "learning_rate": 4.477357683661733e-07, | |
| "loss": 0.2418, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.582, | |
| "grad_norm": 0.5325042009353638, | |
| "learning_rate": 4.442655338968373e-07, | |
| "loss": 0.2262, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 0.584, | |
| "grad_norm": 0.2604765295982361, | |
| "learning_rate": 4.407980158467495e-07, | |
| "loss": 0.2924, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 0.586, | |
| "grad_norm": 0.34971269965171814, | |
| "learning_rate": 4.3733338321784777e-07, | |
| "loss": 0.2656, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 0.588, | |
| "grad_norm": 0.7735595107078552, | |
| "learning_rate": 4.338718048714387e-07, | |
| "loss": 0.3221, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 1.6270298957824707, | |
| "learning_rate": 4.304134495199674e-07, | |
| "loss": 0.2028, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 0.592, | |
| "grad_norm": 0.5100713968276978, | |
| "learning_rate": 4.2695848571879424e-07, | |
| "loss": 0.258, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 0.594, | |
| "grad_norm": 1.0808939933776855, | |
| "learning_rate": 4.23507081857981e-07, | |
| "loss": 0.2214, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 0.596, | |
| "grad_norm": 0.4352293312549591, | |
| "learning_rate": 4.200594061540826e-07, | |
| "loss": 0.1716, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 0.598, | |
| "grad_norm": 0.4087446928024292, | |
| "learning_rate": 4.166156266419489e-07, | |
| "loss": 0.1444, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 0.5501864552497864, | |
| "learning_rate": 4.131759111665348e-07, | |
| "loss": 0.2196, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.602, | |
| "grad_norm": 0.3372376263141632, | |
| "learning_rate": 4.0974042737472005e-07, | |
| "loss": 0.251, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 0.604, | |
| "grad_norm": 0.7682265043258667, | |
| "learning_rate": 4.0630934270713755e-07, | |
| "loss": 0.2154, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 0.606, | |
| "grad_norm": 0.2871810495853424, | |
| "learning_rate": 4.028828243900141e-07, | |
| "loss": 0.227, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 0.608, | |
| "grad_norm": 0.7498311996459961, | |
| "learning_rate": 3.9946103942701775e-07, | |
| "loss": 0.2345, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 0.7101619243621826, | |
| "learning_rate": 3.960441545911204e-07, | |
| "loss": 0.2822, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 0.612, | |
| "grad_norm": 1.4612855911254883, | |
| "learning_rate": 3.9263233641646836e-07, | |
| "loss": 0.3137, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 0.614, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.8922575119026635e-07, | |
| "loss": 0.2221, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 0.616, | |
| "grad_norm": 0.29206129908561707, | |
| "learning_rate": 3.8582456494467206e-07, | |
| "loss": 0.2724, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 0.618, | |
| "grad_norm": 0.8707839846611023, | |
| "learning_rate": 3.8242894344870495e-07, | |
| "loss": 0.2708, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 0.42069849371910095, | |
| "learning_rate": 3.790390522001662e-07, | |
| "loss": 0.261, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.622, | |
| "grad_norm": 0.4054507911205292, | |
| "learning_rate": 3.7565505641757266e-07, | |
| "loss": 0.2331, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 0.624, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.722771210321048e-07, | |
| "loss": 0.1907, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 0.626, | |
| "grad_norm": 0.4163876473903656, | |
| "learning_rate": 3.689054106795677e-07, | |
| "loss": 0.2347, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 0.628, | |
| "grad_norm": 0.4715331792831421, | |
| "learning_rate": 3.6554008969236715e-07, | |
| "loss": 0.1964, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 0.5196170806884766, | |
| "learning_rate": 3.621813220915004e-07, | |
| "loss": 0.2348, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 0.632, | |
| "grad_norm": 0.4927294850349426, | |
| "learning_rate": 3.5882927157856167e-07, | |
| "loss": 0.1554, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 0.634, | |
| "grad_norm": 0.4638221561908722, | |
| "learning_rate": 3.554841015277641e-07, | |
| "loss": 0.216, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 0.636, | |
| "grad_norm": 0.38084176182746887, | |
| "learning_rate": 3.521459749779768e-07, | |
| "loss": 0.2756, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 0.638, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.488150546247778e-07, | |
| "loss": 0.302, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 0.5371063351631165, | |
| "learning_rate": 3.454915028125263e-07, | |
| "loss": 0.2763, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.642, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.421754815264488e-07, | |
| "loss": 0.2124, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 0.644, | |
| "grad_norm": 0.645055890083313, | |
| "learning_rate": 3.388671523847445e-07, | |
| "loss": 0.2644, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 0.646, | |
| "grad_norm": 0.4749380350112915, | |
| "learning_rate": 3.3556667663070835e-07, | |
| "loss": 0.2956, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 0.648, | |
| "grad_norm": 0.5023802518844604, | |
| "learning_rate": 3.3227421512487255e-07, | |
| "loss": 0.2258, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 0.48916229605674744, | |
| "learning_rate": 3.2898992833716563e-07, | |
| "loss": 0.2342, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 0.652, | |
| "grad_norm": 0.33849379420280457, | |
| "learning_rate": 3.257139763390925e-07, | |
| "loss": 0.2239, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 0.654, | |
| "grad_norm": 0.5677793622016907, | |
| "learning_rate": 3.2244651879593156e-07, | |
| "loss": 0.261, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 0.656, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.191877149589539e-07, | |
| "loss": 0.2248, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 0.658, | |
| "grad_norm": 2.0566842555999756, | |
| "learning_rate": 3.15937723657661e-07, | |
| "loss": 0.2277, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 0.5822269916534424, | |
| "learning_rate": 3.1269670329204393e-07, | |
| "loss": 0.234, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.662, | |
| "grad_norm": 0.4788433015346527, | |
| "learning_rate": 3.0946481182486297e-07, | |
| "loss": 0.1916, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 0.664, | |
| "grad_norm": 0.45295077562332153, | |
| "learning_rate": 3.0624220677394854e-07, | |
| "loss": 0.2814, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 0.666, | |
| "grad_norm": 0.9427077770233154, | |
| "learning_rate": 3.0302904520452443e-07, | |
| "loss": 0.3316, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 0.668, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.9982548372155256e-07, | |
| "loss": 0.2791, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.9663167846209996e-07, | |
| "loss": 0.1789, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 0.672, | |
| "grad_norm": 0.8427194952964783, | |
| "learning_rate": 2.9344778508772914e-07, | |
| "loss": 0.2714, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 0.674, | |
| "grad_norm": 0.5976389050483704, | |
| "learning_rate": 2.902739587769114e-07, | |
| "loss": 0.2024, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 0.676, | |
| "grad_norm": 1.0579211711883545, | |
| "learning_rate": 2.8711035421746363e-07, | |
| "loss": 0.1822, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 0.678, | |
| "grad_norm": 0.3103402256965637, | |
| "learning_rate": 2.8395712559900874e-07, | |
| "loss": 0.3106, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 0.5267159342765808, | |
| "learning_rate": 2.808144266054612e-07, | |
| "loss": 0.3567, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.682, | |
| "grad_norm": 0.8383456468582153, | |
| "learning_rate": 2.776824104075364e-07, | |
| "loss": 0.2286, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 0.684, | |
| "grad_norm": 0.3062600791454315, | |
| "learning_rate": 2.745612296552847e-07, | |
| "loss": 0.2748, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 0.686, | |
| "grad_norm": 0.7528430223464966, | |
| "learning_rate": 2.71451036470653e-07, | |
| "loss": 0.2441, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 0.688, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.683519824400692e-07, | |
| "loss": 0.2142, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 0.3675922453403473, | |
| "learning_rate": 2.6526421860705473e-07, | |
| "loss": 0.2415, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 0.692, | |
| "grad_norm": 0.5668989419937134, | |
| "learning_rate": 2.621878954648623e-07, | |
| "loss": 0.2005, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 0.694, | |
| "grad_norm": 0.5000125169754028, | |
| "learning_rate": 2.591231629491423e-07, | |
| "loss": 0.1944, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 0.696, | |
| "grad_norm": 0.4605919420719147, | |
| "learning_rate": 2.5607017043063353e-07, | |
| "loss": 0.1608, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 0.698, | |
| "grad_norm": 0.5186160802841187, | |
| "learning_rate": 2.530290667078846e-07, | |
| "loss": 0.245, | |
| "step": 3490 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 0.32804951071739197, | |
| "learning_rate": 2.500000000000001e-07, | |
| "loss": 0.2557, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.702, | |
| "grad_norm": 1.0500743389129639, | |
| "learning_rate": 2.469831179394182e-07, | |
| "loss": 0.24, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 0.704, | |
| "grad_norm": 0.3622412383556366, | |
| "learning_rate": 2.439785675647143e-07, | |
| "loss": 0.1951, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 0.706, | |
| "grad_norm": 0.5787561535835266, | |
| "learning_rate": 2.4098649531343494e-07, | |
| "loss": 0.2027, | |
| "step": 3530 | |
| }, | |
| { | |
| "epoch": 0.708, | |
| "grad_norm": 0.6269544363021851, | |
| "learning_rate": 2.380070470149605e-07, | |
| "loss": 0.2072, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 0.5072006583213806, | |
| "learning_rate": 2.350403678833976e-07, | |
| "loss": 0.2077, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 0.712, | |
| "grad_norm": 0.3611772656440735, | |
| "learning_rate": 2.3208660251050156e-07, | |
| "loss": 0.1896, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 0.714, | |
| "grad_norm": 0.3685358762741089, | |
| "learning_rate": 2.2914589485863012e-07, | |
| "loss": 0.1974, | |
| "step": 3570 | |
| }, | |
| { | |
| "epoch": 0.716, | |
| "grad_norm": 0.4205205738544464, | |
| "learning_rate": 2.262183882537249e-07, | |
| "loss": 0.2447, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 0.718, | |
| "grad_norm": 0.6153053641319275, | |
| "learning_rate": 2.23304225378328e-07, | |
| "loss": 0.2697, | |
| "step": 3590 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 0.5538428425788879, | |
| "learning_rate": 2.2040354826462664e-07, | |
| "loss": 0.2222, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.722, | |
| "grad_norm": 0.7773278951644897, | |
| "learning_rate": 2.1751649828753106e-07, | |
| "loss": 0.2851, | |
| "step": 3610 | |
| }, | |
| { | |
| "epoch": 0.724, | |
| "grad_norm": 0.8577904105186462, | |
| "learning_rate": 2.146432161577842e-07, | |
| "loss": 0.3621, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 0.726, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.117838419151034e-07, | |
| "loss": 0.1773, | |
| "step": 3630 | |
| }, | |
| { | |
| "epoch": 0.728, | |
| "grad_norm": 0.3094559609889984, | |
| "learning_rate": 2.0893851492135532e-07, | |
| "loss": 0.2208, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 0.5256013870239258, | |
| "learning_rate": 2.0610737385376348e-07, | |
| "loss": 0.2264, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 0.732, | |
| "grad_norm": 0.7884941101074219, | |
| "learning_rate": 2.0329055669814933e-07, | |
| "loss": 0.2455, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 0.734, | |
| "grad_norm": 1.0568552017211914, | |
| "learning_rate": 2.0048820074220711e-07, | |
| "loss": 0.2818, | |
| "step": 3670 | |
| }, | |
| { | |
| "epoch": 0.736, | |
| "grad_norm": 0.7856800556182861, | |
| "learning_rate": 1.9770044256881258e-07, | |
| "loss": 0.2015, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 0.738, | |
| "grad_norm": 0.3122967779636383, | |
| "learning_rate": 1.9492741804936618e-07, | |
| "loss": 0.2609, | |
| "step": 3690 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 0.430240273475647, | |
| "learning_rate": 1.9216926233717084e-07, | |
| "loss": 0.2465, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.742, | |
| "grad_norm": 0.9452837109565735, | |
| "learning_rate": 1.8942610986084484e-07, | |
| "loss": 0.1857, | |
| "step": 3710 | |
| }, | |
| { | |
| "epoch": 0.744, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.8669809431776988e-07, | |
| "loss": 0.1873, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 0.746, | |
| "grad_norm": 0.3083716928958893, | |
| "learning_rate": 1.8398534866757455e-07, | |
| "loss": 0.267, | |
| "step": 3730 | |
| }, | |
| { | |
| "epoch": 0.748, | |
| "grad_norm": 0.46697720885276794, | |
| "learning_rate": 1.812880051256551e-07, | |
| "loss": 0.2621, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 0.6552309393882751, | |
| "learning_rate": 1.7860619515673032e-07, | |
| "loss": 0.2336, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 0.752, | |
| "grad_norm": 0.7122039198875427, | |
| "learning_rate": 1.7594004946843454e-07, | |
| "loss": 0.2676, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 0.754, | |
| "grad_norm": 1.1886111497879028, | |
| "learning_rate": 1.7328969800494726e-07, | |
| "loss": 0.291, | |
| "step": 3770 | |
| }, | |
| { | |
| "epoch": 0.756, | |
| "grad_norm": 0.5067508220672607, | |
| "learning_rate": 1.7065526994065972e-07, | |
| "loss": 0.2135, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 0.758, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.6803689367387918e-07, | |
| "loss": 0.2224, | |
| "step": 3790 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 0.9650610685348511, | |
| "learning_rate": 1.6543469682057104e-07, | |
| "loss": 0.1859, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.762, | |
| "grad_norm": 1.2125353813171387, | |
| "learning_rate": 1.6284880620813846e-07, | |
| "loss": 0.3694, | |
| "step": 3810 | |
| }, | |
| { | |
| "epoch": 0.764, | |
| "grad_norm": 0.6890068054199219, | |
| "learning_rate": 1.6027934786924185e-07, | |
| "loss": 0.2287, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 0.766, | |
| "grad_norm": 0.41877079010009766, | |
| "learning_rate": 1.5772644703565564e-07, | |
| "loss": 0.2163, | |
| "step": 3830 | |
| }, | |
| { | |
| "epoch": 0.768, | |
| "grad_norm": 0.3495817482471466, | |
| "learning_rate": 1.551902281321651e-07, | |
| "loss": 0.2781, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 0.5780860185623169, | |
| "learning_rate": 1.5267081477050131e-07, | |
| "loss": 0.283, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 0.772, | |
| "grad_norm": 0.40200501680374146, | |
| "learning_rate": 1.5016832974331723e-07, | |
| "loss": 0.2548, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 0.774, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.4768289501820263e-07, | |
| "loss": 0.2185, | |
| "step": 3870 | |
| }, | |
| { | |
| "epoch": 0.776, | |
| "grad_norm": 0.7352402806282043, | |
| "learning_rate": 1.4521463173173965e-07, | |
| "loss": 0.2304, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 0.778, | |
| "grad_norm": 0.47302520275115967, | |
| "learning_rate": 1.4276366018359842e-07, | |
| "loss": 0.2641, | |
| "step": 3890 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.4033009983067452e-07, | |
| "loss": 0.1994, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.782, | |
| "grad_norm": 0.2861184775829315, | |
| "learning_rate": 1.3791406928126635e-07, | |
| "loss": 0.1426, | |
| "step": 3910 | |
| }, | |
| { | |
| "epoch": 0.784, | |
| "grad_norm": 1.4924343824386597, | |
| "learning_rate": 1.3551568628929432e-07, | |
| "loss": 0.3393, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 0.786, | |
| "grad_norm": 0.5602060556411743, | |
| "learning_rate": 1.3313506774856175e-07, | |
| "loss": 0.2017, | |
| "step": 3930 | |
| }, | |
| { | |
| "epoch": 0.788, | |
| "grad_norm": 0.778820812702179, | |
| "learning_rate": 1.3077232968705805e-07, | |
| "loss": 0.2557, | |
| "step": 3940 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 0.45512667298316956, | |
| "learning_rate": 1.284275872613028e-07, | |
| "loss": 0.1412, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 0.792, | |
| "grad_norm": 0.8825421333312988, | |
| "learning_rate": 1.2610095475073413e-07, | |
| "loss": 0.2161, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 0.794, | |
| "grad_norm": 0.6215066313743591, | |
| "learning_rate": 1.2379254555213786e-07, | |
| "loss": 0.2547, | |
| "step": 3970 | |
| }, | |
| { | |
| "epoch": 0.796, | |
| "grad_norm": 1.2641725540161133, | |
| "learning_rate": 1.2150247217412185e-07, | |
| "loss": 0.2618, | |
| "step": 3980 | |
| }, | |
| { | |
| "epoch": 0.798, | |
| "grad_norm": 0.782548189163208, | |
| "learning_rate": 1.192308462316317e-07, | |
| "loss": 0.2514, | |
| "step": 3990 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 0.5113526582717896, | |
| "learning_rate": 1.1697777844051104e-07, | |
| "loss": 0.1731, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.802, | |
| "grad_norm": 0.2878662943840027, | |
| "learning_rate": 1.1474337861210543e-07, | |
| "loss": 0.174, | |
| "step": 4010 | |
| }, | |
| { | |
| "epoch": 0.804, | |
| "grad_norm": 0.7684484720230103, | |
| "learning_rate": 1.1252775564791023e-07, | |
| "loss": 0.2255, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 0.806, | |
| "grad_norm": 1.5482900142669678, | |
| "learning_rate": 1.1033101753426282e-07, | |
| "loss": 0.222, | |
| "step": 4030 | |
| }, | |
| { | |
| "epoch": 0.808, | |
| "grad_norm": 0.5780858397483826, | |
| "learning_rate": 1.0815327133708013e-07, | |
| "loss": 0.192, | |
| "step": 4040 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 0.4939838945865631, | |
| "learning_rate": 1.0599462319663904e-07, | |
| "loss": 0.2864, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 0.812, | |
| "grad_norm": 0.4838394224643707, | |
| "learning_rate": 1.038551783224047e-07, | |
| "loss": 0.1951, | |
| "step": 4060 | |
| }, | |
| { | |
| "epoch": 0.814, | |
| "grad_norm": 0.7597946524620056, | |
| "learning_rate": 1.0173504098790186e-07, | |
| "loss": 0.2488, | |
| "step": 4070 | |
| }, | |
| { | |
| "epoch": 0.816, | |
| "grad_norm": 0.692353367805481, | |
| "learning_rate": 9.963431452563331e-08, | |
| "loss": 0.2394, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 0.818, | |
| "grad_norm": 0.8279557228088379, | |
| "learning_rate": 9.755310132204297e-08, | |
| "loss": 0.2723, | |
| "step": 4090 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 0.5341720581054688, | |
| "learning_rate": 9.549150281252632e-08, | |
| "loss": 0.2534, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 0.822, | |
| "grad_norm": 0.47120317816734314, | |
| "learning_rate": 9.344961947648622e-08, | |
| "loss": 0.2336, | |
| "step": 4110 | |
| }, | |
| { | |
| "epoch": 0.824, | |
| "grad_norm": 0.8090870976448059, | |
| "learning_rate": 9.142755083243575e-08, | |
| "loss": 0.2316, | |
| "step": 4120 | |
| }, | |
| { | |
| "epoch": 0.826, | |
| "grad_norm": 0.49878591299057007, | |
| "learning_rate": 8.942539543314798e-08, | |
| "loss": 0.1542, | |
| "step": 4130 | |
| }, | |
| { | |
| "epoch": 0.828, | |
| "grad_norm": 0.7938223481178284, | |
| "learning_rate": 8.744325086085247e-08, | |
| "loss": 0.2709, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 0.0, | |
| "learning_rate": 8.548121372247919e-08, | |
| "loss": 0.1859, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 0.832, | |
| "grad_norm": 0.9708574414253235, | |
| "learning_rate": 8.353937964495028e-08, | |
| "loss": 0.2504, | |
| "step": 4160 | |
| }, | |
| { | |
| "epoch": 0.834, | |
| "grad_norm": 0.487113356590271, | |
| "learning_rate": 8.161784327051919e-08, | |
| "loss": 0.192, | |
| "step": 4170 | |
| }, | |
| { | |
| "epoch": 0.836, | |
| "grad_norm": 0.5531389713287354, | |
| "learning_rate": 7.971669825215787e-08, | |
| "loss": 0.2226, | |
| "step": 4180 | |
| }, | |
| { | |
| "epoch": 0.838, | |
| "grad_norm": 0.3792192339897156, | |
| "learning_rate": 7.783603724899257e-08, | |
| "loss": 0.1823, | |
| "step": 4190 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 0.9227120280265808, | |
| "learning_rate": 7.597595192178702e-08, | |
| "loss": 0.2134, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.842, | |
| "grad_norm": 0.33157506585121155, | |
| "learning_rate": 7.413653292847616e-08, | |
| "loss": 0.2049, | |
| "step": 4210 | |
| }, | |
| { | |
| "epoch": 0.844, | |
| "grad_norm": 0.3281475603580475, | |
| "learning_rate": 7.23178699197467e-08, | |
| "loss": 0.2148, | |
| "step": 4220 | |
| }, | |
| { | |
| "epoch": 0.846, | |
| "grad_norm": 0.9669944643974304, | |
| "learning_rate": 7.052005153466778e-08, | |
| "loss": 0.2296, | |
| "step": 4230 | |
| }, | |
| { | |
| "epoch": 0.848, | |
| "grad_norm": 0.39704257249832153, | |
| "learning_rate": 6.874316539637126e-08, | |
| "loss": 0.2082, | |
| "step": 4240 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 0.2964787781238556, | |
| "learning_rate": 6.698729810778064e-08, | |
| "loss": 0.2279, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 0.852, | |
| "grad_norm": 0.6239051222801208, | |
| "learning_rate": 6.52525352473905e-08, | |
| "loss": 0.2191, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 0.854, | |
| "grad_norm": 0.5734108686447144, | |
| "learning_rate": 6.353896136509524e-08, | |
| "loss": 0.2577, | |
| "step": 4270 | |
| }, | |
| { | |
| "epoch": 0.856, | |
| "grad_norm": 0.6253735423088074, | |
| "learning_rate": 6.184665997806831e-08, | |
| "loss": 0.2616, | |
| "step": 4280 | |
| }, | |
| { | |
| "epoch": 0.858, | |
| "grad_norm": 0.7374691963195801, | |
| "learning_rate": 6.017571356669182e-08, | |
| "loss": 0.2377, | |
| "step": 4290 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 0.3983461856842041, | |
| "learning_rate": 5.8526203570536504e-08, | |
| "loss": 0.1969, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 0.862, | |
| "grad_norm": 0.33935466408729553, | |
| "learning_rate": 5.689821038439263e-08, | |
| "loss": 0.1751, | |
| "step": 4310 | |
| }, | |
| { | |
| "epoch": 0.864, | |
| "grad_norm": 0.5033247470855713, | |
| "learning_rate": 5.529181335435124e-08, | |
| "loss": 0.2077, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 0.866, | |
| "grad_norm": 0.9917274117469788, | |
| "learning_rate": 5.37070907739372e-08, | |
| "loss": 0.2, | |
| "step": 4330 | |
| }, | |
| { | |
| "epoch": 0.868, | |
| "grad_norm": 0.5331639647483826, | |
| "learning_rate": 5.2144119880293544e-08, | |
| "loss": 0.1532, | |
| "step": 4340 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 0.0, | |
| "learning_rate": 5.060297685041659e-08, | |
| "loss": 0.189, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 0.872, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.908373679744315e-08, | |
| "loss": 0.2108, | |
| "step": 4360 | |
| }, | |
| { | |
| "epoch": 0.874, | |
| "grad_norm": 0.40476343035697937, | |
| "learning_rate": 4.758647376699032e-08, | |
| "loss": 0.2197, | |
| "step": 4370 | |
| }, | |
| { | |
| "epoch": 0.876, | |
| "grad_norm": 1.3688344955444336, | |
| "learning_rate": 4.611126073354571e-08, | |
| "loss": 0.3262, | |
| "step": 4380 | |
| }, | |
| { | |
| "epoch": 0.878, | |
| "grad_norm": 0.636691153049469, | |
| "learning_rate": 4.465816959691149e-08, | |
| "loss": 0.2669, | |
| "step": 4390 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 0.0, | |
| "learning_rate": 4.322727117869951e-08, | |
| "loss": 0.1597, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 0.882, | |
| "grad_norm": 0.5956746339797974, | |
| "learning_rate": 4.181863521888018e-08, | |
| "loss": 0.2288, | |
| "step": 4410 | |
| }, | |
| { | |
| "epoch": 0.884, | |
| "grad_norm": 1.2436405420303345, | |
| "learning_rate": 4.043233037238281e-08, | |
| "loss": 0.1909, | |
| "step": 4420 | |
| }, | |
| { | |
| "epoch": 0.886, | |
| "grad_norm": 0.6453717947006226, | |
| "learning_rate": 3.9068424205749794e-08, | |
| "loss": 0.2444, | |
| "step": 4430 | |
| }, | |
| { | |
| "epoch": 0.888, | |
| "grad_norm": 1.696014642715454, | |
| "learning_rate": 3.7726983193843485e-08, | |
| "loss": 0.3259, | |
| "step": 4440 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 2.008049726486206, | |
| "learning_rate": 3.6408072716606345e-08, | |
| "loss": 0.2821, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 0.892, | |
| "grad_norm": 1.3175231218338013, | |
| "learning_rate": 3.5111757055874326e-08, | |
| "loss": 0.2012, | |
| "step": 4460 | |
| }, | |
| { | |
| "epoch": 0.894, | |
| "grad_norm": 0.41426166892051697, | |
| "learning_rate": 3.3838099392243915e-08, | |
| "loss": 0.2701, | |
| "step": 4470 | |
| }, | |
| { | |
| "epoch": 0.896, | |
| "grad_norm": 0.49628499150276184, | |
| "learning_rate": 3.258716180199278e-08, | |
| "loss": 0.1985, | |
| "step": 4480 | |
| }, | |
| { | |
| "epoch": 0.898, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.135900525405427e-08, | |
| "loss": 0.1697, | |
| "step": 4490 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 0.7981248497962952, | |
| "learning_rate": 3.015368960704584e-08, | |
| "loss": 0.2987, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.902, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.8971273606351655e-08, | |
| "loss": 0.1614, | |
| "step": 4510 | |
| }, | |
| { | |
| "epoch": 0.904, | |
| "grad_norm": 0.32520195841789246, | |
| "learning_rate": 2.78118148812595e-08, | |
| "loss": 0.2718, | |
| "step": 4520 | |
| }, | |
| { | |
| "epoch": 0.906, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.667536994215186e-08, | |
| "loss": 0.1763, | |
| "step": 4530 | |
| }, | |
| { | |
| "epoch": 0.908, | |
| "grad_norm": 0.9632695913314819, | |
| "learning_rate": 2.5561994177751732e-08, | |
| "loss": 0.2242, | |
| "step": 4540 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 0.3187273144721985, | |
| "learning_rate": 2.4471741852423233e-08, | |
| "loss": 0.1798, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 0.912, | |
| "grad_norm": 0.0, | |
| "learning_rate": 2.3404666103526537e-08, | |
| "loss": 0.2245, | |
| "step": 4560 | |
| }, | |
| { | |
| "epoch": 0.914, | |
| "grad_norm": 0.640092670917511, | |
| "learning_rate": 2.2360818938828187e-08, | |
| "loss": 0.1911, | |
| "step": 4570 | |
| }, | |
| { | |
| "epoch": 0.916, | |
| "grad_norm": 0.43212875723838806, | |
| "learning_rate": 2.1340251233966377e-08, | |
| "loss": 0.2488, | |
| "step": 4580 | |
| }, | |
| { | |
| "epoch": 0.918, | |
| "grad_norm": 1.5328131914138794, | |
| "learning_rate": 2.0343012729971243e-08, | |
| "loss": 0.1862, | |
| "step": 4590 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 0.5170596241950989, | |
| "learning_rate": 1.936915203084055e-08, | |
| "loss": 0.2004, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 0.922, | |
| "grad_norm": 0.3647305965423584, | |
| "learning_rate": 1.8418716601170947e-08, | |
| "loss": 0.2377, | |
| "step": 4610 | |
| }, | |
| { | |
| "epoch": 0.924, | |
| "grad_norm": 0.9026084542274475, | |
| "learning_rate": 1.7491752763844292e-08, | |
| "loss": 0.2804, | |
| "step": 4620 | |
| }, | |
| { | |
| "epoch": 0.926, | |
| "grad_norm": 0.5593412518501282, | |
| "learning_rate": 1.658830569777031e-08, | |
| "loss": 0.1757, | |
| "step": 4630 | |
| }, | |
| { | |
| "epoch": 0.928, | |
| "grad_norm": 0.413272887468338, | |
| "learning_rate": 1.570841943568446e-08, | |
| "loss": 0.2462, | |
| "step": 4640 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 0.28161871433258057, | |
| "learning_rate": 1.4852136862001763e-08, | |
| "loss": 0.2418, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 0.932, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.4019499710726911e-08, | |
| "loss": 0.1699, | |
| "step": 4660 | |
| }, | |
| { | |
| "epoch": 0.934, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.3210548563419855e-08, | |
| "loss": 0.1736, | |
| "step": 4670 | |
| }, | |
| { | |
| "epoch": 0.936, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.2425322847218367e-08, | |
| "loss": 0.2341, | |
| "step": 4680 | |
| }, | |
| { | |
| "epoch": 0.938, | |
| "grad_norm": 0.4318271279335022, | |
| "learning_rate": 1.166386083291604e-08, | |
| "loss": 0.2545, | |
| "step": 4690 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 0.7377297282218933, | |
| "learning_rate": 1.0926199633097154e-08, | |
| "loss": 0.2266, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 0.942, | |
| "grad_norm": 0.419969379901886, | |
| "learning_rate": 1.0212375200327972e-08, | |
| "loss": 0.2356, | |
| "step": 4710 | |
| }, | |
| { | |
| "epoch": 0.944, | |
| "grad_norm": 0.0, | |
| "learning_rate": 9.522422325404233e-09, | |
| "loss": 0.1757, | |
| "step": 4720 | |
| }, | |
| { | |
| "epoch": 0.946, | |
| "grad_norm": 0.39264801144599915, | |
| "learning_rate": 8.856374635655695e-09, | |
| "loss": 0.303, | |
| "step": 4730 | |
| }, | |
| { | |
| "epoch": 0.948, | |
| "grad_norm": 0.24799562990665436, | |
| "learning_rate": 8.214264593307096e-09, | |
| "loss": 0.2823, | |
| "step": 4740 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 0.28534314036369324, | |
| "learning_rate": 7.59612349389599e-09, | |
| "loss": 0.2156, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 0.952, | |
| "grad_norm": 0.0, | |
| "learning_rate": 7.0019814647475636e-09, | |
| "loss": 0.0769, | |
| "step": 4760 | |
| }, | |
| { | |
| "epoch": 0.954, | |
| "grad_norm": 1.0923534631729126, | |
| "learning_rate": 6.431867463506046e-09, | |
| "loss": 0.2032, | |
| "step": 4770 | |
| }, | |
| { | |
| "epoch": 0.956, | |
| "grad_norm": 0.4145523011684418, | |
| "learning_rate": 5.8858092767236076e-09, | |
| "loss": 0.2456, | |
| "step": 4780 | |
| }, | |
| { | |
| "epoch": 0.958, | |
| "grad_norm": 0.2941437065601349, | |
| "learning_rate": 5.3638335185058335e-09, | |
| "loss": 0.2009, | |
| "step": 4790 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 0.9238514304161072, | |
| "learning_rate": 4.865965629214819e-09, | |
| "loss": 0.2775, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 0.962, | |
| "grad_norm": 0.8849767446517944, | |
| "learning_rate": 4.3922298742291585e-09, | |
| "loss": 0.1857, | |
| "step": 4810 | |
| }, | |
| { | |
| "epoch": 0.964, | |
| "grad_norm": 0.5588952302932739, | |
| "learning_rate": 3.9426493427611175e-09, | |
| "loss": 0.2474, | |
| "step": 4820 | |
| }, | |
| { | |
| "epoch": 0.966, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.5172459467315286e-09, | |
| "loss": 0.1684, | |
| "step": 4830 | |
| }, | |
| { | |
| "epoch": 0.968, | |
| "grad_norm": 0.6499897241592407, | |
| "learning_rate": 3.116040419701815e-09, | |
| "loss": 0.2, | |
| "step": 4840 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 0.49148255586624146, | |
| "learning_rate": 2.739052315863355e-09, | |
| "loss": 0.24, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 0.972, | |
| "grad_norm": 0.3919503092765808, | |
| "learning_rate": 2.3863000090844076e-09, | |
| "loss": 0.2908, | |
| "step": 4860 | |
| }, | |
| { | |
| "epoch": 0.974, | |
| "grad_norm": 0.6548805236816406, | |
| "learning_rate": 2.057800692014833e-09, | |
| "loss": 0.2264, | |
| "step": 4870 | |
| }, | |
| { | |
| "epoch": 0.976, | |
| "grad_norm": 0.7637394070625305, | |
| "learning_rate": 1.7535703752478147e-09, | |
| "loss": 0.3091, | |
| "step": 4880 | |
| }, | |
| { | |
| "epoch": 0.978, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.4736238865398765e-09, | |
| "loss": 0.2276, | |
| "step": 4890 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 0.43148624897003174, | |
| "learning_rate": 1.217974870087901e-09, | |
| "loss": 0.2555, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 0.982, | |
| "grad_norm": 0.3615666627883911, | |
| "learning_rate": 9.866357858642205e-10, | |
| "loss": 0.294, | |
| "step": 4910 | |
| }, | |
| { | |
| "epoch": 0.984, | |
| "grad_norm": 0.2953715920448303, | |
| "learning_rate": 7.79617909009489e-10, | |
| "loss": 0.226, | |
| "step": 4920 | |
| }, | |
| { | |
| "epoch": 0.986, | |
| "grad_norm": 0.36070477962493896, | |
| "learning_rate": 5.969313292830125e-10, | |
| "loss": 0.2444, | |
| "step": 4930 | |
| }, | |
| { | |
| "epoch": 0.988, | |
| "grad_norm": 0.5537273287773132, | |
| "learning_rate": 4.3858495057080836e-10, | |
| "loss": 0.2805, | |
| "step": 4940 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 0.0, | |
| "learning_rate": 3.0458649045211894e-10, | |
| "loss": 0.2459, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 0.992, | |
| "grad_norm": 0.0, | |
| "learning_rate": 1.9494247982282387e-10, | |
| "loss": 0.111, | |
| "step": 4960 | |
| }, | |
| { | |
| "epoch": 0.994, | |
| "grad_norm": 0.5425102114677429, | |
| "learning_rate": 1.0965826257725019e-10, | |
| "loss": 0.1702, | |
| "step": 4970 | |
| }, | |
| { | |
| "epoch": 0.996, | |
| "grad_norm": 0.514282763004303, | |
| "learning_rate": 4.873799534788059e-11, | |
| "loss": 0.1889, | |
| "step": 4980 | |
| }, | |
| { | |
| "epoch": 0.998, | |
| "grad_norm": 0.9735618829727173, | |
| "learning_rate": 1.2184647302626582e-11, | |
| "loss": 0.249, | |
| "step": 4990 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 0.6470169425010681, | |
| "learning_rate": 0.0, | |
| "loss": 0.2384, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "step": 5000, | |
| "total_flos": 2.239561404560179e+16, | |
| "train_loss": 0.3289372156143188, | |
| "train_runtime": 837.4572, | |
| "train_samples_per_second": 5.97, | |
| "train_steps_per_second": 5.97 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 5000, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 4000, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 2.239561404560179e+16, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |