| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.9958246346555324, | |
| "eval_steps": 500, | |
| "global_step": 638, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.003131524008350731, | |
| "grad_norm": 13.917898178100586, | |
| "learning_rate": 5.0000000000000004e-08, | |
| "loss": 4.1051, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.006263048016701462, | |
| "grad_norm": 17.327869415283203, | |
| "learning_rate": 1.0000000000000001e-07, | |
| "loss": 4.1048, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.009394572025052192, | |
| "grad_norm": 14.063946723937988, | |
| "learning_rate": 1.5000000000000002e-07, | |
| "loss": 4.0741, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.012526096033402923, | |
| "grad_norm": 16.817699432373047, | |
| "learning_rate": 2.0000000000000002e-07, | |
| "loss": 4.2002, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.015657620041753653, | |
| "grad_norm": 14.47036361694336, | |
| "learning_rate": 2.5000000000000004e-07, | |
| "loss": 4.2652, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.018789144050104383, | |
| "grad_norm": 14.474193572998047, | |
| "learning_rate": 3.0000000000000004e-07, | |
| "loss": 4.0888, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.021920668058455117, | |
| "grad_norm": 14.865458488464355, | |
| "learning_rate": 3.5000000000000004e-07, | |
| "loss": 4.0014, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.025052192066805846, | |
| "grad_norm": 15.338888168334961, | |
| "learning_rate": 4.0000000000000003e-07, | |
| "loss": 4.13, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.028183716075156576, | |
| "grad_norm": 15.154336929321289, | |
| "learning_rate": 4.5000000000000003e-07, | |
| "loss": 4.2493, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.031315240083507306, | |
| "grad_norm": 15.919597625732422, | |
| "learning_rate": 5.000000000000001e-07, | |
| "loss": 4.0535, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.03444676409185804, | |
| "grad_norm": 14.981926918029785, | |
| "learning_rate": 5.5e-07, | |
| "loss": 3.9064, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.037578288100208766, | |
| "grad_norm": 13.36101245880127, | |
| "learning_rate": 6.000000000000001e-07, | |
| "loss": 4.1939, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.0407098121085595, | |
| "grad_norm": 15.58773422241211, | |
| "learning_rate": 6.5e-07, | |
| "loss": 4.18, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.04384133611691023, | |
| "grad_norm": 13.560139656066895, | |
| "learning_rate": 7.000000000000001e-07, | |
| "loss": 3.9414, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.04697286012526096, | |
| "grad_norm": 12.307971954345703, | |
| "learning_rate": 7.5e-07, | |
| "loss": 3.8836, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.05010438413361169, | |
| "grad_norm": 14.533182144165039, | |
| "learning_rate": 8.000000000000001e-07, | |
| "loss": 4.1551, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.05323590814196242, | |
| "grad_norm": 13.453729629516602, | |
| "learning_rate": 8.500000000000001e-07, | |
| "loss": 4.0048, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.05636743215031315, | |
| "grad_norm": 13.45992374420166, | |
| "learning_rate": 9.000000000000001e-07, | |
| "loss": 4.0745, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.059498956158663886, | |
| "grad_norm": 11.857145309448242, | |
| "learning_rate": 9.500000000000001e-07, | |
| "loss": 3.9871, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.06263048016701461, | |
| "grad_norm": 11.872294425964355, | |
| "learning_rate": 1.0000000000000002e-06, | |
| "loss": 3.8959, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.06576200417536535, | |
| "grad_norm": 12.969825744628906, | |
| "learning_rate": 1.0500000000000001e-06, | |
| "loss": 4.0308, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.06889352818371608, | |
| "grad_norm": 12.33769416809082, | |
| "learning_rate": 1.1e-06, | |
| "loss": 3.9341, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.0720250521920668, | |
| "grad_norm": 12.669405937194824, | |
| "learning_rate": 1.1500000000000002e-06, | |
| "loss": 3.8511, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.07515657620041753, | |
| "grad_norm": 10.677213668823242, | |
| "learning_rate": 1.2000000000000002e-06, | |
| "loss": 3.7764, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.07828810020876827, | |
| "grad_norm": 10.366402626037598, | |
| "learning_rate": 1.25e-06, | |
| "loss": 3.5291, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.081419624217119, | |
| "grad_norm": 11.211421012878418, | |
| "learning_rate": 1.3e-06, | |
| "loss": 3.5765, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.08455114822546973, | |
| "grad_norm": 11.313716888427734, | |
| "learning_rate": 1.3500000000000002e-06, | |
| "loss": 3.4849, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.08768267223382047, | |
| "grad_norm": 10.41294002532959, | |
| "learning_rate": 1.4000000000000001e-06, | |
| "loss": 3.2653, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.09081419624217119, | |
| "grad_norm": 10.40064525604248, | |
| "learning_rate": 1.45e-06, | |
| "loss": 3.3384, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.09394572025052192, | |
| "grad_norm": 10.05427074432373, | |
| "learning_rate": 1.5e-06, | |
| "loss": 3.2257, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.09707724425887265, | |
| "grad_norm": 9.583163261413574, | |
| "learning_rate": 1.5500000000000002e-06, | |
| "loss": 3.1371, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.10020876826722339, | |
| "grad_norm": 10.09977912902832, | |
| "learning_rate": 1.6000000000000001e-06, | |
| "loss": 3.0658, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.10334029227557412, | |
| "grad_norm": 9.271486282348633, | |
| "learning_rate": 1.6500000000000003e-06, | |
| "loss": 2.9693, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.10647181628392484, | |
| "grad_norm": 10.687992095947266, | |
| "learning_rate": 1.7000000000000002e-06, | |
| "loss": 2.95, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.10960334029227557, | |
| "grad_norm": 8.762290000915527, | |
| "learning_rate": 1.75e-06, | |
| "loss": 2.8286, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.1127348643006263, | |
| "grad_norm": 10.13785171508789, | |
| "learning_rate": 1.8000000000000001e-06, | |
| "loss": 2.3664, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.11586638830897704, | |
| "grad_norm": 18.301353454589844, | |
| "learning_rate": 1.85e-06, | |
| "loss": 2.5533, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.11899791231732777, | |
| "grad_norm": 11.490377426147461, | |
| "learning_rate": 1.9000000000000002e-06, | |
| "loss": 2.6133, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.12212943632567849, | |
| "grad_norm": 15.614163398742676, | |
| "learning_rate": 1.9500000000000004e-06, | |
| "loss": 2.3596, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.12526096033402923, | |
| "grad_norm": 17.757442474365234, | |
| "learning_rate": 2.0000000000000003e-06, | |
| "loss": 2.3491, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.12839248434237996, | |
| "grad_norm": 17.18431854248047, | |
| "learning_rate": 2.05e-06, | |
| "loss": 2.2361, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.1315240083507307, | |
| "grad_norm": 16.149789810180664, | |
| "learning_rate": 2.1000000000000002e-06, | |
| "loss": 2.1457, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.13465553235908143, | |
| "grad_norm": 15.256914138793945, | |
| "learning_rate": 2.15e-06, | |
| "loss": 2.12, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.13778705636743216, | |
| "grad_norm": 15.537406921386719, | |
| "learning_rate": 2.2e-06, | |
| "loss": 2.1877, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.1409185803757829, | |
| "grad_norm": 7.947713851928711, | |
| "learning_rate": 2.25e-06, | |
| "loss": 2.1648, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.1440501043841336, | |
| "grad_norm": 8.818676948547363, | |
| "learning_rate": 2.3000000000000004e-06, | |
| "loss": 2.134, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.14718162839248433, | |
| "grad_norm": 5.175768852233887, | |
| "learning_rate": 2.35e-06, | |
| "loss": 2.0796, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.15031315240083507, | |
| "grad_norm": 6.750611305236816, | |
| "learning_rate": 2.4000000000000003e-06, | |
| "loss": 1.9174, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.1534446764091858, | |
| "grad_norm": 6.2147979736328125, | |
| "learning_rate": 2.4500000000000003e-06, | |
| "loss": 1.8065, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.15657620041753653, | |
| "grad_norm": 13.291611671447754, | |
| "learning_rate": 2.5e-06, | |
| "loss": 1.7061, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.15970772442588727, | |
| "grad_norm": 7.251201629638672, | |
| "learning_rate": 2.55e-06, | |
| "loss": 1.7924, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.162839248434238, | |
| "grad_norm": 5.2126054763793945, | |
| "learning_rate": 2.6e-06, | |
| "loss": 1.6735, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.16597077244258873, | |
| "grad_norm": 5.435528755187988, | |
| "learning_rate": 2.6500000000000005e-06, | |
| "loss": 1.6265, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.16910229645093947, | |
| "grad_norm": 4.505807399749756, | |
| "learning_rate": 2.7000000000000004e-06, | |
| "loss": 1.4851, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.1722338204592902, | |
| "grad_norm": 5.128388404846191, | |
| "learning_rate": 2.7500000000000004e-06, | |
| "loss": 1.5832, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.17536534446764093, | |
| "grad_norm": 16.935827255249023, | |
| "learning_rate": 2.8000000000000003e-06, | |
| "loss": 1.6553, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.17849686847599164, | |
| "grad_norm": 3.664458990097046, | |
| "learning_rate": 2.85e-06, | |
| "loss": 1.5, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.18162839248434237, | |
| "grad_norm": 7.763802528381348, | |
| "learning_rate": 2.9e-06, | |
| "loss": 1.367, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.1847599164926931, | |
| "grad_norm": 3.2216155529022217, | |
| "learning_rate": 2.95e-06, | |
| "loss": 1.3863, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.18789144050104384, | |
| "grad_norm": 4.384445667266846, | |
| "learning_rate": 3e-06, | |
| "loss": 1.4247, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.19102296450939457, | |
| "grad_norm": 4.8080878257751465, | |
| "learning_rate": 3.05e-06, | |
| "loss": 1.3257, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.1941544885177453, | |
| "grad_norm": 4.154761791229248, | |
| "learning_rate": 3.1000000000000004e-06, | |
| "loss": 1.321, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.19728601252609604, | |
| "grad_norm": 6.4742112159729, | |
| "learning_rate": 3.1500000000000003e-06, | |
| "loss": 1.2823, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.20041753653444677, | |
| "grad_norm": 2.583422899246216, | |
| "learning_rate": 3.2000000000000003e-06, | |
| "loss": 1.2136, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.2035490605427975, | |
| "grad_norm": 4.1933488845825195, | |
| "learning_rate": 3.2500000000000002e-06, | |
| "loss": 1.1855, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.20668058455114824, | |
| "grad_norm": 4.11049747467041, | |
| "learning_rate": 3.3000000000000006e-06, | |
| "loss": 1.2389, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.20981210855949894, | |
| "grad_norm": 2.264458417892456, | |
| "learning_rate": 3.3500000000000005e-06, | |
| "loss": 1.0651, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.21294363256784968, | |
| "grad_norm": 2.5408174991607666, | |
| "learning_rate": 3.4000000000000005e-06, | |
| "loss": 1.1389, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.2160751565762004, | |
| "grad_norm": 7.82421350479126, | |
| "learning_rate": 3.45e-06, | |
| "loss": 1.0956, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.21920668058455114, | |
| "grad_norm": 3.070939064025879, | |
| "learning_rate": 3.5e-06, | |
| "loss": 1.0451, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.22233820459290188, | |
| "grad_norm": 2.6310527324676514, | |
| "learning_rate": 3.5500000000000003e-06, | |
| "loss": 1.0538, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.2254697286012526, | |
| "grad_norm": 7.630155563354492, | |
| "learning_rate": 3.6000000000000003e-06, | |
| "loss": 1.0052, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.22860125260960334, | |
| "grad_norm": 6.950636863708496, | |
| "learning_rate": 3.65e-06, | |
| "loss": 1.0473, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.23173277661795408, | |
| "grad_norm": 2.2703945636749268, | |
| "learning_rate": 3.7e-06, | |
| "loss": 1.0576, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.2348643006263048, | |
| "grad_norm": 3.3817710876464844, | |
| "learning_rate": 3.7500000000000005e-06, | |
| "loss": 1.0177, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.23799582463465555, | |
| "grad_norm": 7.266414642333984, | |
| "learning_rate": 3.8000000000000005e-06, | |
| "loss": 1.0645, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.24112734864300625, | |
| "grad_norm": 5.782608509063721, | |
| "learning_rate": 3.85e-06, | |
| "loss": 1.0162, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.24425887265135698, | |
| "grad_norm": 2.7938575744628906, | |
| "learning_rate": 3.900000000000001e-06, | |
| "loss": 0.9664, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.24739039665970772, | |
| "grad_norm": 6.681935787200928, | |
| "learning_rate": 3.95e-06, | |
| "loss": 0.953, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.25052192066805845, | |
| "grad_norm": 2.253279209136963, | |
| "learning_rate": 4.000000000000001e-06, | |
| "loss": 0.9568, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.2536534446764092, | |
| "grad_norm": 1.4875826835632324, | |
| "learning_rate": 4.05e-06, | |
| "loss": 0.9448, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.2567849686847599, | |
| "grad_norm": 2.4987940788269043, | |
| "learning_rate": 4.1e-06, | |
| "loss": 0.9393, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.2599164926931106, | |
| "grad_norm": 4.712948322296143, | |
| "learning_rate": 4.15e-06, | |
| "loss": 0.9532, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.2630480167014614, | |
| "grad_norm": 6.9030632972717285, | |
| "learning_rate": 4.2000000000000004e-06, | |
| "loss": 0.96, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.2661795407098121, | |
| "grad_norm": 3.4780967235565186, | |
| "learning_rate": 4.25e-06, | |
| "loss": 0.8993, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.26931106471816285, | |
| "grad_norm": 1.526064395904541, | |
| "learning_rate": 4.3e-06, | |
| "loss": 0.9021, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.27244258872651356, | |
| "grad_norm": 10.727686882019043, | |
| "learning_rate": 4.350000000000001e-06, | |
| "loss": 0.856, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.2755741127348643, | |
| "grad_norm": 12.483160972595215, | |
| "learning_rate": 4.4e-06, | |
| "loss": 0.9357, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.278705636743215, | |
| "grad_norm": 6.544492244720459, | |
| "learning_rate": 4.450000000000001e-06, | |
| "loss": 0.9168, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.2818371607515658, | |
| "grad_norm": 1.178139567375183, | |
| "learning_rate": 4.5e-06, | |
| "loss": 0.8748, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.2849686847599165, | |
| "grad_norm": 1.711506962776184, | |
| "learning_rate": 4.5500000000000005e-06, | |
| "loss": 0.8425, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.2881002087682672, | |
| "grad_norm": 3.281747341156006, | |
| "learning_rate": 4.600000000000001e-06, | |
| "loss": 0.8491, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.29123173277661796, | |
| "grad_norm": 2.2964377403259277, | |
| "learning_rate": 4.65e-06, | |
| "loss": 0.8038, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.29436325678496866, | |
| "grad_norm": 1.959700345993042, | |
| "learning_rate": 4.7e-06, | |
| "loss": 0.8439, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.2974947807933194, | |
| "grad_norm": 3.979384183883667, | |
| "learning_rate": 4.75e-06, | |
| "loss": 0.8839, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.30062630480167013, | |
| "grad_norm": 1.4721262454986572, | |
| "learning_rate": 4.800000000000001e-06, | |
| "loss": 0.845, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.3037578288100209, | |
| "grad_norm": 2.862248659133911, | |
| "learning_rate": 4.85e-06, | |
| "loss": 0.7748, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.3068893528183716, | |
| "grad_norm": 3.7439088821411133, | |
| "learning_rate": 4.9000000000000005e-06, | |
| "loss": 0.8145, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.31002087682672236, | |
| "grad_norm": 1.6654618978500366, | |
| "learning_rate": 4.95e-06, | |
| "loss": 0.8326, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.31315240083507306, | |
| "grad_norm": 7.8437581062316895, | |
| "learning_rate": 5e-06, | |
| "loss": 0.8666, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.3162839248434238, | |
| "grad_norm": 6.429738521575928, | |
| "learning_rate": 4.999996250830422e-06, | |
| "loss": 0.836, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.31941544885177453, | |
| "grad_norm": 2.6017794609069824, | |
| "learning_rate": 4.9999850033329326e-06, | |
| "loss": 0.7785, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.32254697286012524, | |
| "grad_norm": 1.0575449466705322, | |
| "learning_rate": 4.999966257541265e-06, | |
| "loss": 0.7639, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.325678496868476, | |
| "grad_norm": 2.6932010650634766, | |
| "learning_rate": 4.999940013511647e-06, | |
| "loss": 0.8214, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.3288100208768267, | |
| "grad_norm": 2.925288438796997, | |
| "learning_rate": 4.999906271322792e-06, | |
| "loss": 0.8797, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.33194154488517746, | |
| "grad_norm": 1.3570607900619507, | |
| "learning_rate": 4.9998650310759035e-06, | |
| "loss": 0.792, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.33507306889352817, | |
| "grad_norm": 5.126713752746582, | |
| "learning_rate": 4.999816292894676e-06, | |
| "loss": 0.8352, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.33820459290187893, | |
| "grad_norm": 1.8966432809829712, | |
| "learning_rate": 4.99976005692529e-06, | |
| "loss": 0.7663, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.34133611691022964, | |
| "grad_norm": 1.3100829124450684, | |
| "learning_rate": 4.999696323336418e-06, | |
| "loss": 0.771, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.3444676409185804, | |
| "grad_norm": 2.4025354385375977, | |
| "learning_rate": 4.999625092319218e-06, | |
| "loss": 0.7618, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.3475991649269311, | |
| "grad_norm": 1.130232810974121, | |
| "learning_rate": 4.999546364087334e-06, | |
| "loss": 0.7705, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.35073068893528186, | |
| "grad_norm": 3.430262327194214, | |
| "learning_rate": 4.999460138876901e-06, | |
| "loss": 0.77, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.35386221294363257, | |
| "grad_norm": 1.1272103786468506, | |
| "learning_rate": 4.999366416946536e-06, | |
| "loss": 0.7133, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.3569937369519833, | |
| "grad_norm": 1.1740471124649048, | |
| "learning_rate": 4.999265198577342e-06, | |
| "loss": 0.7684, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.36012526096033404, | |
| "grad_norm": 1.3138248920440674, | |
| "learning_rate": 4.999156484072907e-06, | |
| "loss": 0.7888, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.36325678496868474, | |
| "grad_norm": 1.061711311340332, | |
| "learning_rate": 4.999040273759304e-06, | |
| "loss": 0.7484, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.3663883089770355, | |
| "grad_norm": 1.4682390689849854, | |
| "learning_rate": 4.998916567985083e-06, | |
| "loss": 0.7296, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.3695198329853862, | |
| "grad_norm": 2.884068250656128, | |
| "learning_rate": 4.998785367121284e-06, | |
| "loss": 0.7662, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.37265135699373697, | |
| "grad_norm": 0.9812761545181274, | |
| "learning_rate": 4.9986466715614205e-06, | |
| "loss": 0.7307, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.3757828810020877, | |
| "grad_norm": 2.2237496376037598, | |
| "learning_rate": 4.998500481721484e-06, | |
| "loss": 0.6761, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.37891440501043844, | |
| "grad_norm": 1.4004178047180176, | |
| "learning_rate": 4.998346798039952e-06, | |
| "loss": 0.7505, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.38204592901878914, | |
| "grad_norm": 5.54975700378418, | |
| "learning_rate": 4.99818562097777e-06, | |
| "loss": 0.7615, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.38517745302713985, | |
| "grad_norm": 6.17140531539917, | |
| "learning_rate": 4.9980169510183624e-06, | |
| "loss": 0.7002, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.3883089770354906, | |
| "grad_norm": 4.974380016326904, | |
| "learning_rate": 4.997840788667628e-06, | |
| "loss": 0.7449, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.3914405010438413, | |
| "grad_norm": 1.4133399724960327, | |
| "learning_rate": 4.997657134453937e-06, | |
| "loss": 0.7442, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.3945720250521921, | |
| "grad_norm": 1.868915319442749, | |
| "learning_rate": 4.9974659889281295e-06, | |
| "loss": 0.7104, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.3977035490605428, | |
| "grad_norm": 1.2599350214004517, | |
| "learning_rate": 4.997267352663514e-06, | |
| "loss": 0.7385, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.40083507306889354, | |
| "grad_norm": 1.4353641271591187, | |
| "learning_rate": 4.997061226255869e-06, | |
| "loss": 0.7081, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.40396659707724425, | |
| "grad_norm": 3.2492141723632812, | |
| "learning_rate": 4.996847610323437e-06, | |
| "loss": 0.7859, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.407098121085595, | |
| "grad_norm": 9.599719047546387, | |
| "learning_rate": 4.996626505506923e-06, | |
| "loss": 0.7241, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.4102296450939457, | |
| "grad_norm": 10.053650856018066, | |
| "learning_rate": 4.996397912469494e-06, | |
| "loss": 0.6841, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.4133611691022965, | |
| "grad_norm": 1.323876976966858, | |
| "learning_rate": 4.996161831896777e-06, | |
| "loss": 0.7317, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.4164926931106472, | |
| "grad_norm": 1.4180598258972168, | |
| "learning_rate": 4.9959182644968594e-06, | |
| "loss": 0.692, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.4196242171189979, | |
| "grad_norm": 1.2194396257400513, | |
| "learning_rate": 4.99566721100028e-06, | |
| "loss": 0.7068, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.42275574112734865, | |
| "grad_norm": 1.0984960794448853, | |
| "learning_rate": 4.995408672160031e-06, | |
| "loss": 0.6946, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.42588726513569936, | |
| "grad_norm": 1.9341071844100952, | |
| "learning_rate": 4.995142648751561e-06, | |
| "loss": 0.7467, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.4290187891440501, | |
| "grad_norm": 1.9960932731628418, | |
| "learning_rate": 4.9948691415727594e-06, | |
| "loss": 0.7379, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.4321503131524008, | |
| "grad_norm": 0.8743917942047119, | |
| "learning_rate": 4.994588151443968e-06, | |
| "loss": 0.66, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.4352818371607516, | |
| "grad_norm": 0.8655261993408203, | |
| "learning_rate": 4.99429967920797e-06, | |
| "loss": 0.6646, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.4384133611691023, | |
| "grad_norm": 5.462070941925049, | |
| "learning_rate": 4.994003725729992e-06, | |
| "loss": 0.643, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.44154488517745305, | |
| "grad_norm": 2.1401469707489014, | |
| "learning_rate": 4.993700291897695e-06, | |
| "loss": 0.6639, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.44467640918580376, | |
| "grad_norm": 1.8219833374023438, | |
| "learning_rate": 4.9933893786211815e-06, | |
| "loss": 0.6673, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.44780793319415446, | |
| "grad_norm": 1.641079306602478, | |
| "learning_rate": 4.993070986832984e-06, | |
| "loss": 0.658, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.4509394572025052, | |
| "grad_norm": 1.1739819049835205, | |
| "learning_rate": 4.992745117488066e-06, | |
| "loss": 0.6826, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.45407098121085593, | |
| "grad_norm": 2.309185743331909, | |
| "learning_rate": 4.9924117715638185e-06, | |
| "loss": 0.6536, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.4572025052192067, | |
| "grad_norm": 1.09304940700531, | |
| "learning_rate": 4.99207095006006e-06, | |
| "loss": 0.721, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.4603340292275574, | |
| "grad_norm": 0.9056984186172485, | |
| "learning_rate": 4.991722653999025e-06, | |
| "loss": 0.7019, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.46346555323590816, | |
| "grad_norm": 1.8440625667572021, | |
| "learning_rate": 4.991366884425374e-06, | |
| "loss": 0.707, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.46659707724425886, | |
| "grad_norm": 1.2244676351547241, | |
| "learning_rate": 4.991003642406177e-06, | |
| "loss": 0.6407, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.4697286012526096, | |
| "grad_norm": 0.9258589744567871, | |
| "learning_rate": 4.99063292903092e-06, | |
| "loss": 0.6954, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.47286012526096033, | |
| "grad_norm": 4.176390647888184, | |
| "learning_rate": 4.990254745411496e-06, | |
| "loss": 0.6812, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.4759916492693111, | |
| "grad_norm": 1.4322530031204224, | |
| "learning_rate": 4.989869092682205e-06, | |
| "loss": 0.6808, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.4791231732776618, | |
| "grad_norm": 0.8017717003822327, | |
| "learning_rate": 4.989475971999748e-06, | |
| "loss": 0.687, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.4822546972860125, | |
| "grad_norm": 1.5641374588012695, | |
| "learning_rate": 4.989075384543228e-06, | |
| "loss": 0.6599, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.48538622129436326, | |
| "grad_norm": 1.1522141695022583, | |
| "learning_rate": 4.98866733151414e-06, | |
| "loss": 0.6546, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.48851774530271397, | |
| "grad_norm": 0.8593171238899231, | |
| "learning_rate": 4.988251814136372e-06, | |
| "loss": 0.6857, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.49164926931106473, | |
| "grad_norm": 2.668159246444702, | |
| "learning_rate": 4.9878288336562e-06, | |
| "loss": 0.661, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.49478079331941544, | |
| "grad_norm": 0.9953671097755432, | |
| "learning_rate": 4.987398391342285e-06, | |
| "loss": 0.6512, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.4979123173277662, | |
| "grad_norm": 1.042872667312622, | |
| "learning_rate": 4.986960488485667e-06, | |
| "loss": 0.6311, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.5010438413361169, | |
| "grad_norm": 0.9070663452148438, | |
| "learning_rate": 4.9865151263997645e-06, | |
| "loss": 0.675, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.5041753653444676, | |
| "grad_norm": 0.8460433483123779, | |
| "learning_rate": 4.986062306420367e-06, | |
| "loss": 0.6635, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.5073068893528184, | |
| "grad_norm": 1.2639834880828857, | |
| "learning_rate": 4.985602029905635e-06, | |
| "loss": 0.6327, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.5104384133611691, | |
| "grad_norm": 0.8775074481964111, | |
| "learning_rate": 4.985134298236091e-06, | |
| "loss": 0.644, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.5135699373695198, | |
| "grad_norm": 1.2031961679458618, | |
| "learning_rate": 4.98465911281462e-06, | |
| "loss": 0.6254, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.5167014613778705, | |
| "grad_norm": 0.892494797706604, | |
| "learning_rate": 4.984176475066463e-06, | |
| "loss": 0.7122, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.5198329853862212, | |
| "grad_norm": 2.7122485637664795, | |
| "learning_rate": 4.983686386439212e-06, | |
| "loss": 0.6679, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.5229645093945721, | |
| "grad_norm": 0.9344426989555359, | |
| "learning_rate": 4.983188848402806e-06, | |
| "loss": 0.6319, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.5260960334029228, | |
| "grad_norm": 1.4093577861785889, | |
| "learning_rate": 4.982683862449531e-06, | |
| "loss": 0.6425, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.5292275574112735, | |
| "grad_norm": 1.1285009384155273, | |
| "learning_rate": 4.982171430094007e-06, | |
| "loss": 0.6298, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.5323590814196242, | |
| "grad_norm": 1.952778935432434, | |
| "learning_rate": 4.981651552873193e-06, | |
| "loss": 0.7066, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.535490605427975, | |
| "grad_norm": 5.133765697479248, | |
| "learning_rate": 4.981124232346374e-06, | |
| "loss": 0.6634, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.5386221294363257, | |
| "grad_norm": 0.9770542979240417, | |
| "learning_rate": 4.980589470095161e-06, | |
| "loss": 0.7121, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.5417536534446764, | |
| "grad_norm": 0.8414323925971985, | |
| "learning_rate": 4.980047267723487e-06, | |
| "loss": 0.6397, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.5448851774530271, | |
| "grad_norm": 1.9173879623413086, | |
| "learning_rate": 4.979497626857596e-06, | |
| "loss": 0.6228, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.5480167014613778, | |
| "grad_norm": 1.0823363065719604, | |
| "learning_rate": 4.978940549146048e-06, | |
| "loss": 0.6475, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.5511482254697286, | |
| "grad_norm": 3.715353488922119, | |
| "learning_rate": 4.978376036259706e-06, | |
| "loss": 0.7127, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.5542797494780793, | |
| "grad_norm": 0.981584370136261, | |
| "learning_rate": 4.9778040898917325e-06, | |
| "loss": 0.6468, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.55741127348643, | |
| "grad_norm": 1.70566987991333, | |
| "learning_rate": 4.977224711757587e-06, | |
| "loss": 0.6476, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.5605427974947808, | |
| "grad_norm": 0.9217923283576965, | |
| "learning_rate": 4.976637903595019e-06, | |
| "loss": 0.6731, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.5636743215031316, | |
| "grad_norm": 0.8994677662849426, | |
| "learning_rate": 4.976043667164063e-06, | |
| "loss": 0.6562, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.5668058455114823, | |
| "grad_norm": 1.1613017320632935, | |
| "learning_rate": 4.975442004247034e-06, | |
| "loss": 0.6417, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.569937369519833, | |
| "grad_norm": 1.6041977405548096, | |
| "learning_rate": 4.974832916648521e-06, | |
| "loss": 0.6029, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.5730688935281837, | |
| "grad_norm": 1.7978405952453613, | |
| "learning_rate": 4.974216406195383e-06, | |
| "loss": 0.6269, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.5762004175365344, | |
| "grad_norm": 1.6021920442581177, | |
| "learning_rate": 4.973592474736739e-06, | |
| "loss": 0.6149, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.5793319415448852, | |
| "grad_norm": 0.8973568677902222, | |
| "learning_rate": 4.972961124143971e-06, | |
| "loss": 0.6648, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.5824634655532359, | |
| "grad_norm": 1.9432591199874878, | |
| "learning_rate": 4.972322356310711e-06, | |
| "loss": 0.6299, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.5855949895615866, | |
| "grad_norm": 4.457028388977051, | |
| "learning_rate": 4.971676173152839e-06, | |
| "loss": 0.656, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.5887265135699373, | |
| "grad_norm": 2.0989716053009033, | |
| "learning_rate": 4.971022576608473e-06, | |
| "loss": 0.6539, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.5918580375782881, | |
| "grad_norm": 1.0646967887878418, | |
| "learning_rate": 4.97036156863797e-06, | |
| "loss": 0.6727, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.5949895615866388, | |
| "grad_norm": 1.6522265672683716, | |
| "learning_rate": 4.969693151223914e-06, | |
| "loss": 0.6643, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.5981210855949896, | |
| "grad_norm": 1.7503505945205688, | |
| "learning_rate": 4.969017326371115e-06, | |
| "loss": 0.6402, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.6012526096033403, | |
| "grad_norm": 1.2341989278793335, | |
| "learning_rate": 4.968334096106597e-06, | |
| "loss": 0.6413, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.6043841336116911, | |
| "grad_norm": 3.089054584503174, | |
| "learning_rate": 4.967643462479597e-06, | |
| "loss": 0.6825, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.6075156576200418, | |
| "grad_norm": 2.711623430252075, | |
| "learning_rate": 4.966945427561557e-06, | |
| "loss": 0.65, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.6106471816283925, | |
| "grad_norm": 4.641184329986572, | |
| "learning_rate": 4.966239993446118e-06, | |
| "loss": 0.6229, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.6137787056367432, | |
| "grad_norm": 1.7984074354171753, | |
| "learning_rate": 4.965527162249114e-06, | |
| "loss": 0.6473, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.6169102296450939, | |
| "grad_norm": 1.1643115282058716, | |
| "learning_rate": 4.964806936108566e-06, | |
| "loss": 0.6404, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.6200417536534447, | |
| "grad_norm": 2.1877920627593994, | |
| "learning_rate": 4.9640793171846725e-06, | |
| "loss": 0.6185, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.6231732776617954, | |
| "grad_norm": 1.7970566749572754, | |
| "learning_rate": 4.963344307659807e-06, | |
| "loss": 0.634, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.6263048016701461, | |
| "grad_norm": 1.6014361381530762, | |
| "learning_rate": 4.96260190973851e-06, | |
| "loss": 0.6562, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.6294363256784968, | |
| "grad_norm": 0.8743320107460022, | |
| "learning_rate": 4.961852125647482e-06, | |
| "loss": 0.6133, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.6325678496868476, | |
| "grad_norm": 1.9526551961898804, | |
| "learning_rate": 4.961094957635578e-06, | |
| "loss": 0.6451, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.6356993736951984, | |
| "grad_norm": 3.6597347259521484, | |
| "learning_rate": 4.960330407973798e-06, | |
| "loss": 0.6386, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.6388308977035491, | |
| "grad_norm": 1.7180207967758179, | |
| "learning_rate": 4.959558478955283e-06, | |
| "loss": 0.6688, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.6419624217118998, | |
| "grad_norm": 0.9058470129966736, | |
| "learning_rate": 4.958779172895308e-06, | |
| "loss": 0.6161, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.6450939457202505, | |
| "grad_norm": 1.0031033754348755, | |
| "learning_rate": 4.957992492131274e-06, | |
| "loss": 0.6437, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.6482254697286013, | |
| "grad_norm": 1.5846725702285767, | |
| "learning_rate": 4.9571984390226985e-06, | |
| "loss": 0.6332, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.651356993736952, | |
| "grad_norm": 1.9951609373092651, | |
| "learning_rate": 4.956397015951215e-06, | |
| "loss": 0.636, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.6544885177453027, | |
| "grad_norm": 1.4122583866119385, | |
| "learning_rate": 4.95558822532056e-06, | |
| "loss": 0.6586, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.6576200417536534, | |
| "grad_norm": 1.2243481874465942, | |
| "learning_rate": 4.954772069556568e-06, | |
| "loss": 0.6313, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.6607515657620042, | |
| "grad_norm": 0.8756356835365295, | |
| "learning_rate": 4.953948551107164e-06, | |
| "loss": 0.6406, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.6638830897703549, | |
| "grad_norm": 2.9979734420776367, | |
| "learning_rate": 4.953117672442356e-06, | |
| "loss": 0.5803, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.6670146137787056, | |
| "grad_norm": 2.1859359741210938, | |
| "learning_rate": 4.952279436054229e-06, | |
| "loss": 0.6607, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.6701461377870563, | |
| "grad_norm": 0.6929755806922913, | |
| "learning_rate": 4.9514338444569346e-06, | |
| "loss": 0.5989, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.673277661795407, | |
| "grad_norm": 1.0361783504486084, | |
| "learning_rate": 4.950580900186685e-06, | |
| "loss": 0.6654, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.6764091858037579, | |
| "grad_norm": 1.210898518562317, | |
| "learning_rate": 4.9497206058017475e-06, | |
| "loss": 0.6213, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.6795407098121086, | |
| "grad_norm": 1.200990080833435, | |
| "learning_rate": 4.948852963882434e-06, | |
| "loss": 0.6654, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.6826722338204593, | |
| "grad_norm": 1.481831669807434, | |
| "learning_rate": 4.947977977031093e-06, | |
| "loss": 0.6474, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.68580375782881, | |
| "grad_norm": 0.9883334636688232, | |
| "learning_rate": 4.947095647872103e-06, | |
| "loss": 0.6735, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.6889352818371608, | |
| "grad_norm": 0.7436536550521851, | |
| "learning_rate": 4.946205979051868e-06, | |
| "loss": 0.6456, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.6920668058455115, | |
| "grad_norm": 0.9057570099830627, | |
| "learning_rate": 4.945308973238802e-06, | |
| "loss": 0.6228, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 0.6951983298538622, | |
| "grad_norm": 1.341081142425537, | |
| "learning_rate": 4.944404633123324e-06, | |
| "loss": 0.6417, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.6983298538622129, | |
| "grad_norm": 0.7958157062530518, | |
| "learning_rate": 4.943492961417859e-06, | |
| "loss": 0.6494, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 0.7014613778705637, | |
| "grad_norm": 1.216025471687317, | |
| "learning_rate": 4.9425739608568106e-06, | |
| "loss": 0.6566, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.7045929018789144, | |
| "grad_norm": 0.9774854779243469, | |
| "learning_rate": 4.9416476341965735e-06, | |
| "loss": 0.6171, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.7077244258872651, | |
| "grad_norm": 2.1562681198120117, | |
| "learning_rate": 4.940713984215512e-06, | |
| "loss": 0.629, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.7108559498956158, | |
| "grad_norm": 1.9521286487579346, | |
| "learning_rate": 4.9397730137139556e-06, | |
| "loss": 0.6475, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 0.7139874739039666, | |
| "grad_norm": 1.5749104022979736, | |
| "learning_rate": 4.9388247255141895e-06, | |
| "loss": 0.6053, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.7171189979123174, | |
| "grad_norm": 1.2008254528045654, | |
| "learning_rate": 4.937869122460449e-06, | |
| "loss": 0.6052, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 0.7202505219206681, | |
| "grad_norm": 1.0774102210998535, | |
| "learning_rate": 4.93690620741891e-06, | |
| "loss": 0.6099, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.7233820459290188, | |
| "grad_norm": 1.0929996967315674, | |
| "learning_rate": 4.935935983277675e-06, | |
| "loss": 0.6363, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.7265135699373695, | |
| "grad_norm": 0.8830653429031372, | |
| "learning_rate": 4.934958452946774e-06, | |
| "loss": 0.6136, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.7296450939457203, | |
| "grad_norm": 3.591218948364258, | |
| "learning_rate": 4.933973619358147e-06, | |
| "loss": 0.5962, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 0.732776617954071, | |
| "grad_norm": 2.5797672271728516, | |
| "learning_rate": 4.932981485465643e-06, | |
| "loss": 0.6405, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.7359081419624217, | |
| "grad_norm": 1.0467664003372192, | |
| "learning_rate": 4.9319820542450025e-06, | |
| "loss": 0.6155, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.7390396659707724, | |
| "grad_norm": 0.8099795579910278, | |
| "learning_rate": 4.930975328693856e-06, | |
| "loss": 0.5615, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.7421711899791231, | |
| "grad_norm": 0.8906702995300293, | |
| "learning_rate": 4.92996131183171e-06, | |
| "loss": 0.6501, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.7453027139874739, | |
| "grad_norm": 1.0871416330337524, | |
| "learning_rate": 4.928940006699944e-06, | |
| "loss": 0.6282, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.7484342379958246, | |
| "grad_norm": 1.3209614753723145, | |
| "learning_rate": 4.927911416361792e-06, | |
| "loss": 0.598, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 0.7515657620041754, | |
| "grad_norm": 1.2252682447433472, | |
| "learning_rate": 4.926875543902344e-06, | |
| "loss": 0.6433, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.7546972860125261, | |
| "grad_norm": 1.0569007396697998, | |
| "learning_rate": 4.9258323924285285e-06, | |
| "loss": 0.5927, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 0.7578288100208769, | |
| "grad_norm": 0.9309014081954956, | |
| "learning_rate": 4.924781965069106e-06, | |
| "loss": 0.5927, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.7609603340292276, | |
| "grad_norm": 1.0200378894805908, | |
| "learning_rate": 4.923724264974662e-06, | |
| "loss": 0.6064, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.7640918580375783, | |
| "grad_norm": 1.0533075332641602, | |
| "learning_rate": 4.922659295317593e-06, | |
| "loss": 0.6373, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.767223382045929, | |
| "grad_norm": 0.7889382839202881, | |
| "learning_rate": 4.921587059292102e-06, | |
| "loss": 0.5887, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.7703549060542797, | |
| "grad_norm": 0.7943588495254517, | |
| "learning_rate": 4.920507560114183e-06, | |
| "loss": 0.593, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.7734864300626305, | |
| "grad_norm": 0.8247205018997192, | |
| "learning_rate": 4.919420801021617e-06, | |
| "loss": 0.6151, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 0.7766179540709812, | |
| "grad_norm": 0.9979158043861389, | |
| "learning_rate": 4.91832678527396e-06, | |
| "loss": 0.6019, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.7797494780793319, | |
| "grad_norm": 0.9346868991851807, | |
| "learning_rate": 4.917225516152532e-06, | |
| "loss": 0.6098, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.7828810020876826, | |
| "grad_norm": 0.7487881183624268, | |
| "learning_rate": 4.916116996960408e-06, | |
| "loss": 0.5965, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.7860125260960334, | |
| "grad_norm": 0.821576714515686, | |
| "learning_rate": 4.915001231022411e-06, | |
| "loss": 0.6483, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 0.7891440501043842, | |
| "grad_norm": 1.0413196086883545, | |
| "learning_rate": 4.913878221685096e-06, | |
| "loss": 0.6108, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.7922755741127349, | |
| "grad_norm": 0.9560331702232361, | |
| "learning_rate": 4.912747972316745e-06, | |
| "loss": 0.5758, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 0.7954070981210856, | |
| "grad_norm": 0.8964638113975525, | |
| "learning_rate": 4.911610486307356e-06, | |
| "loss": 0.6432, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 0.7985386221294363, | |
| "grad_norm": 0.8418346047401428, | |
| "learning_rate": 4.910465767068631e-06, | |
| "loss": 0.6027, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.8016701461377871, | |
| "grad_norm": 1.792371153831482, | |
| "learning_rate": 4.909313818033966e-06, | |
| "loss": 0.6198, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.8048016701461378, | |
| "grad_norm": 1.036665439605713, | |
| "learning_rate": 4.908154642658446e-06, | |
| "loss": 0.6255, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 0.8079331941544885, | |
| "grad_norm": 0.7592151165008545, | |
| "learning_rate": 4.906988244418823e-06, | |
| "loss": 0.6035, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.8110647181628392, | |
| "grad_norm": 0.8843073844909668, | |
| "learning_rate": 4.90581462681352e-06, | |
| "loss": 0.6299, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 0.81419624217119, | |
| "grad_norm": 0.9489964246749878, | |
| "learning_rate": 4.9046337933626086e-06, | |
| "loss": 0.5869, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.8173277661795407, | |
| "grad_norm": 0.851691722869873, | |
| "learning_rate": 4.903445747607806e-06, | |
| "loss": 0.603, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 0.8204592901878914, | |
| "grad_norm": 1.3722106218338013, | |
| "learning_rate": 4.902250493112458e-06, | |
| "loss": 0.5939, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 0.8235908141962421, | |
| "grad_norm": 1.1002827882766724, | |
| "learning_rate": 4.901048033461537e-06, | |
| "loss": 0.6452, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 0.826722338204593, | |
| "grad_norm": 0.8428632020950317, | |
| "learning_rate": 4.89983837226162e-06, | |
| "loss": 0.5956, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.8298538622129437, | |
| "grad_norm": 0.7666584849357605, | |
| "learning_rate": 4.898621513140889e-06, | |
| "loss": 0.6067, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.8329853862212944, | |
| "grad_norm": 0.8413611054420471, | |
| "learning_rate": 4.897397459749113e-06, | |
| "loss": 0.5985, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 0.8361169102296451, | |
| "grad_norm": 2.3374335765838623, | |
| "learning_rate": 4.896166215757638e-06, | |
| "loss": 0.5885, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 0.8392484342379958, | |
| "grad_norm": 2.236640214920044, | |
| "learning_rate": 4.894927784859377e-06, | |
| "loss": 0.6408, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.8423799582463466, | |
| "grad_norm": 0.9715856313705444, | |
| "learning_rate": 4.893682170768802e-06, | |
| "loss": 0.5954, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 0.8455114822546973, | |
| "grad_norm": 1.0249912738800049, | |
| "learning_rate": 4.892429377221928e-06, | |
| "loss": 0.6186, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.848643006263048, | |
| "grad_norm": 1.255426049232483, | |
| "learning_rate": 4.891169407976302e-06, | |
| "loss": 0.6351, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 0.8517745302713987, | |
| "grad_norm": 0.9339559674263, | |
| "learning_rate": 4.889902266810995e-06, | |
| "loss": 0.5944, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.8549060542797495, | |
| "grad_norm": 1.2473429441452026, | |
| "learning_rate": 4.888627957526589e-06, | |
| "loss": 0.544, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 0.8580375782881002, | |
| "grad_norm": 1.0589442253112793, | |
| "learning_rate": 4.887346483945166e-06, | |
| "loss": 0.5543, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 0.8611691022964509, | |
| "grad_norm": 0.9844024777412415, | |
| "learning_rate": 4.886057849910294e-06, | |
| "loss": 0.5941, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.8643006263048016, | |
| "grad_norm": 2.88578200340271, | |
| "learning_rate": 4.8847620592870196e-06, | |
| "loss": 0.6124, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.8674321503131524, | |
| "grad_norm": 0.7496054172515869, | |
| "learning_rate": 4.8834591159618524e-06, | |
| "loss": 0.6006, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 0.8705636743215032, | |
| "grad_norm": 0.7403052449226379, | |
| "learning_rate": 4.88214902384276e-06, | |
| "loss": 0.5911, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 0.8736951983298539, | |
| "grad_norm": 0.9003771543502808, | |
| "learning_rate": 4.880831786859146e-06, | |
| "loss": 0.6347, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 0.8768267223382046, | |
| "grad_norm": 1.0345501899719238, | |
| "learning_rate": 4.879507408961847e-06, | |
| "loss": 0.6111, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.8799582463465553, | |
| "grad_norm": 1.4385879039764404, | |
| "learning_rate": 4.878175894123116e-06, | |
| "loss": 0.6454, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 0.8830897703549061, | |
| "grad_norm": 0.8469482064247131, | |
| "learning_rate": 4.8768372463366145e-06, | |
| "loss": 0.6163, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.8862212943632568, | |
| "grad_norm": 0.8859589695930481, | |
| "learning_rate": 4.875491469617395e-06, | |
| "loss": 0.6144, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 0.8893528183716075, | |
| "grad_norm": 1.8436834812164307, | |
| "learning_rate": 4.874138568001895e-06, | |
| "loss": 0.6275, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.8924843423799582, | |
| "grad_norm": 0.6646101474761963, | |
| "learning_rate": 4.87277854554792e-06, | |
| "loss": 0.615, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.8956158663883089, | |
| "grad_norm": 1.0070925951004028, | |
| "learning_rate": 4.871411406334633e-06, | |
| "loss": 0.5898, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.8987473903966597, | |
| "grad_norm": 0.9785194993019104, | |
| "learning_rate": 4.870037154462545e-06, | |
| "loss": 0.5992, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 0.9018789144050104, | |
| "grad_norm": 0.7244889736175537, | |
| "learning_rate": 4.868655794053497e-06, | |
| "loss": 0.6078, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.9050104384133612, | |
| "grad_norm": 1.4496444463729858, | |
| "learning_rate": 4.8672673292506535e-06, | |
| "loss": 0.5855, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 0.9081419624217119, | |
| "grad_norm": 1.8514957427978516, | |
| "learning_rate": 4.865871764218486e-06, | |
| "loss": 0.5707, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.9112734864300627, | |
| "grad_norm": 0.8439773321151733, | |
| "learning_rate": 4.864469103142763e-06, | |
| "loss": 0.5562, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 0.9144050104384134, | |
| "grad_norm": 0.8146086931228638, | |
| "learning_rate": 4.8630593502305355e-06, | |
| "loss": 0.6161, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.9175365344467641, | |
| "grad_norm": 0.8920315504074097, | |
| "learning_rate": 4.861642509710126e-06, | |
| "loss": 0.6139, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 0.9206680584551148, | |
| "grad_norm": 1.4980088472366333, | |
| "learning_rate": 4.860218585831116e-06, | |
| "loss": 0.6187, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 0.9237995824634656, | |
| "grad_norm": 0.9910127520561218, | |
| "learning_rate": 4.8587875828643285e-06, | |
| "loss": 0.5852, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.9269311064718163, | |
| "grad_norm": 0.819600522518158, | |
| "learning_rate": 4.857349505101823e-06, | |
| "loss": 0.6172, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 0.930062630480167, | |
| "grad_norm": 1.1059772968292236, | |
| "learning_rate": 4.855904356856878e-06, | |
| "loss": 0.5868, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 0.9331941544885177, | |
| "grad_norm": 1.2362196445465088, | |
| "learning_rate": 4.854452142463977e-06, | |
| "loss": 0.625, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 0.9363256784968684, | |
| "grad_norm": 0.9956470727920532, | |
| "learning_rate": 4.852992866278799e-06, | |
| "loss": 0.5923, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 0.9394572025052192, | |
| "grad_norm": 0.864109218120575, | |
| "learning_rate": 4.851526532678203e-06, | |
| "loss": 0.6315, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.94258872651357, | |
| "grad_norm": 0.8900614380836487, | |
| "learning_rate": 4.850053146060217e-06, | |
| "loss": 0.6128, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 0.9457202505219207, | |
| "grad_norm": 0.927254855632782, | |
| "learning_rate": 4.84857271084402e-06, | |
| "loss": 0.5955, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 0.9488517745302714, | |
| "grad_norm": 1.0046517848968506, | |
| "learning_rate": 4.847085231469935e-06, | |
| "loss": 0.6134, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 0.9519832985386222, | |
| "grad_norm": 0.734597384929657, | |
| "learning_rate": 4.8455907123994125e-06, | |
| "loss": 0.5927, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 0.9551148225469729, | |
| "grad_norm": 0.7338348031044006, | |
| "learning_rate": 4.844089158115016e-06, | |
| "loss": 0.5897, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.9582463465553236, | |
| "grad_norm": 0.9163988828659058, | |
| "learning_rate": 4.8425805731204106e-06, | |
| "loss": 0.6051, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 0.9613778705636743, | |
| "grad_norm": 1.050246238708496, | |
| "learning_rate": 4.84106496194035e-06, | |
| "loss": 0.5751, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 0.964509394572025, | |
| "grad_norm": 0.7637603878974915, | |
| "learning_rate": 4.83954232912066e-06, | |
| "loss": 0.5677, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 0.9676409185803758, | |
| "grad_norm": 0.7110525965690613, | |
| "learning_rate": 4.838012679228229e-06, | |
| "loss": 0.6051, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 0.9707724425887265, | |
| "grad_norm": 0.7662068605422974, | |
| "learning_rate": 4.836476016850988e-06, | |
| "loss": 0.59, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.9739039665970772, | |
| "grad_norm": 0.8907375335693359, | |
| "learning_rate": 4.834932346597906e-06, | |
| "loss": 0.5792, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 0.9770354906054279, | |
| "grad_norm": 0.8939849138259888, | |
| "learning_rate": 4.833381673098966e-06, | |
| "loss": 0.6062, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.9801670146137788, | |
| "grad_norm": 0.8878788948059082, | |
| "learning_rate": 4.8318240010051595e-06, | |
| "loss": 0.5694, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 0.9832985386221295, | |
| "grad_norm": 1.2523870468139648, | |
| "learning_rate": 4.830259334988468e-06, | |
| "loss": 0.5809, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 0.9864300626304802, | |
| "grad_norm": 1.0836797952651978, | |
| "learning_rate": 4.82868767974185e-06, | |
| "loss": 0.5949, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.9895615866388309, | |
| "grad_norm": 0.7985473871231079, | |
| "learning_rate": 4.827109039979226e-06, | |
| "loss": 0.6057, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 0.9926931106471816, | |
| "grad_norm": 1.042951226234436, | |
| "learning_rate": 4.825523420435469e-06, | |
| "loss": 0.6004, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 0.9958246346555324, | |
| "grad_norm": 0.7845115661621094, | |
| "learning_rate": 4.823930825866381e-06, | |
| "loss": 0.6161, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 0.9989561586638831, | |
| "grad_norm": 0.931854784488678, | |
| "learning_rate": 4.82233126104869e-06, | |
| "loss": 0.5912, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 0.931854784488678, | |
| "learning_rate": 4.8207247307800275e-06, | |
| "loss": 0.1914, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 1.0031315240083507, | |
| "grad_norm": 0.751028835773468, | |
| "learning_rate": 4.819111239878916e-06, | |
| "loss": 0.5802, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 1.0062630480167014, | |
| "grad_norm": 1.4943569898605347, | |
| "learning_rate": 4.817490793184758e-06, | |
| "loss": 0.613, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 1.0093945720250521, | |
| "grad_norm": 2.296318531036377, | |
| "learning_rate": 4.815863395557816e-06, | |
| "loss": 0.5453, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 1.0125260960334028, | |
| "grad_norm": 0.760101318359375, | |
| "learning_rate": 4.814229051879202e-06, | |
| "loss": 0.5302, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 1.0156576200417538, | |
| "grad_norm": 0.8145846128463745, | |
| "learning_rate": 4.812587767050861e-06, | |
| "loss": 0.5831, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 1.0187891440501045, | |
| "grad_norm": 0.9169796109199524, | |
| "learning_rate": 4.8109395459955565e-06, | |
| "loss": 0.5756, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 1.0219206680584552, | |
| "grad_norm": 0.8791524171829224, | |
| "learning_rate": 4.809284393656858e-06, | |
| "loss": 0.5988, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 1.0250521920668059, | |
| "grad_norm": 1.0184170007705688, | |
| "learning_rate": 4.807622314999122e-06, | |
| "loss": 0.5476, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 1.0281837160751566, | |
| "grad_norm": 0.8095184564590454, | |
| "learning_rate": 4.8059533150074805e-06, | |
| "loss": 0.5723, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 1.0313152400835073, | |
| "grad_norm": 0.7621930241584778, | |
| "learning_rate": 4.804277398687826e-06, | |
| "loss": 0.5841, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 1.034446764091858, | |
| "grad_norm": 3.729628324508667, | |
| "learning_rate": 4.802594571066791e-06, | |
| "loss": 0.5639, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 1.0375782881002087, | |
| "grad_norm": 1.6502974033355713, | |
| "learning_rate": 4.800904837191743e-06, | |
| "loss": 0.6024, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 1.0407098121085594, | |
| "grad_norm": 0.8031198978424072, | |
| "learning_rate": 4.799208202130762e-06, | |
| "loss": 0.5305, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 1.0438413361169103, | |
| "grad_norm": 0.939644992351532, | |
| "learning_rate": 4.797504670972623e-06, | |
| "loss": 0.5446, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 1.046972860125261, | |
| "grad_norm": 1.0589954853057861, | |
| "learning_rate": 4.795794248826789e-06, | |
| "loss": 0.5366, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 1.0501043841336117, | |
| "grad_norm": 0.9089614748954773, | |
| "learning_rate": 4.794076940823391e-06, | |
| "loss": 0.5795, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 1.0532359081419624, | |
| "grad_norm": 0.7732561230659485, | |
| "learning_rate": 4.792352752113212e-06, | |
| "loss": 0.5765, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 1.0563674321503131, | |
| "grad_norm": 1.811553955078125, | |
| "learning_rate": 4.790621687867672e-06, | |
| "loss": 0.561, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 1.0594989561586639, | |
| "grad_norm": 1.1930758953094482, | |
| "learning_rate": 4.788883753278813e-06, | |
| "loss": 0.5, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 1.0626304801670146, | |
| "grad_norm": 0.9551813006401062, | |
| "learning_rate": 4.787138953559285e-06, | |
| "loss": 0.5228, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 1.0657620041753653, | |
| "grad_norm": 0.9609586596488953, | |
| "learning_rate": 4.785387293942329e-06, | |
| "loss": 0.5827, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 1.068893528183716, | |
| "grad_norm": 0.8403449654579163, | |
| "learning_rate": 4.783628779681759e-06, | |
| "loss": 0.5585, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 1.072025052192067, | |
| "grad_norm": 0.9108251929283142, | |
| "learning_rate": 4.7818634160519496e-06, | |
| "loss": 0.6077, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 1.0751565762004176, | |
| "grad_norm": 0.9476898908615112, | |
| "learning_rate": 4.780091208347819e-06, | |
| "loss": 0.5493, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 1.0782881002087683, | |
| "grad_norm": 1.1943707466125488, | |
| "learning_rate": 4.778312161884813e-06, | |
| "loss": 0.5736, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 1.081419624217119, | |
| "grad_norm": 3.1342639923095703, | |
| "learning_rate": 4.77652628199889e-06, | |
| "loss": 0.5765, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 1.0845511482254697, | |
| "grad_norm": 2.7982125282287598, | |
| "learning_rate": 4.7747335740465015e-06, | |
| "loss": 0.6003, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 1.0876826722338204, | |
| "grad_norm": 1.5068914890289307, | |
| "learning_rate": 4.7729340434045815e-06, | |
| "loss": 0.5033, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 1.0908141962421711, | |
| "grad_norm": 0.8273429274559021, | |
| "learning_rate": 4.771127695470527e-06, | |
| "loss": 0.5309, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 1.0939457202505218, | |
| "grad_norm": 1.104974389076233, | |
| "learning_rate": 4.76931453566218e-06, | |
| "loss": 0.5244, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 1.0970772442588728, | |
| "grad_norm": 1.096509337425232, | |
| "learning_rate": 4.7674945694178166e-06, | |
| "loss": 0.5585, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 1.1002087682672235, | |
| "grad_norm": 1.0238200426101685, | |
| "learning_rate": 4.765667802196127e-06, | |
| "loss": 0.5589, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 1.1033402922755742, | |
| "grad_norm": 0.7515526413917542, | |
| "learning_rate": 4.763834239476197e-06, | |
| "loss": 0.5304, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 1.1064718162839249, | |
| "grad_norm": 1.0282566547393799, | |
| "learning_rate": 4.761993886757499e-06, | |
| "loss": 0.5476, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 1.1096033402922756, | |
| "grad_norm": 0.9962708950042725, | |
| "learning_rate": 4.760146749559868e-06, | |
| "loss": 0.5117, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 1.1127348643006263, | |
| "grad_norm": 0.7851671576499939, | |
| "learning_rate": 4.758292833423488e-06, | |
| "loss": 0.5542, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 1.115866388308977, | |
| "grad_norm": 0.8857759237289429, | |
| "learning_rate": 4.756432143908876e-06, | |
| "loss": 0.544, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 1.1189979123173277, | |
| "grad_norm": 0.9402740597724915, | |
| "learning_rate": 4.7545646865968645e-06, | |
| "loss": 0.5656, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 1.1221294363256784, | |
| "grad_norm": 0.8210407495498657, | |
| "learning_rate": 4.752690467088584e-06, | |
| "loss": 0.5733, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 1.1252609603340291, | |
| "grad_norm": 0.795684278011322, | |
| "learning_rate": 4.750809491005449e-06, | |
| "loss": 0.5678, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 1.12839248434238, | |
| "grad_norm": 0.8712463974952698, | |
| "learning_rate": 4.748921763989139e-06, | |
| "loss": 0.5777, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 1.1315240083507307, | |
| "grad_norm": 0.9810119867324829, | |
| "learning_rate": 4.747027291701578e-06, | |
| "loss": 0.5511, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 1.1346555323590815, | |
| "grad_norm": 0.81117844581604, | |
| "learning_rate": 4.745126079824926e-06, | |
| "loss": 0.5038, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 1.1377870563674322, | |
| "grad_norm": 0.7631494402885437, | |
| "learning_rate": 4.743218134061556e-06, | |
| "loss": 0.6272, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 1.1409185803757829, | |
| "grad_norm": 0.7601696252822876, | |
| "learning_rate": 4.741303460134038e-06, | |
| "loss": 0.571, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 1.1440501043841336, | |
| "grad_norm": 1.7977744340896606, | |
| "learning_rate": 4.7393820637851205e-06, | |
| "loss": 0.538, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 1.1471816283924843, | |
| "grad_norm": 2.022578001022339, | |
| "learning_rate": 4.737453950777718e-06, | |
| "loss": 0.5822, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 1.150313152400835, | |
| "grad_norm": 0.7586764693260193, | |
| "learning_rate": 4.735519126894885e-06, | |
| "loss": 0.5986, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 1.153444676409186, | |
| "grad_norm": 0.8970286846160889, | |
| "learning_rate": 4.733577597939812e-06, | |
| "loss": 0.542, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 1.1565762004175366, | |
| "grad_norm": 0.8546352982521057, | |
| "learning_rate": 4.731629369735793e-06, | |
| "loss": 0.5832, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 1.1597077244258873, | |
| "grad_norm": 0.9266164898872375, | |
| "learning_rate": 4.72967444812622e-06, | |
| "loss": 0.551, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 1.162839248434238, | |
| "grad_norm": 1.0413658618927002, | |
| "learning_rate": 4.7277128389745595e-06, | |
| "loss": 0.5866, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 1.1659707724425887, | |
| "grad_norm": 0.9312199950218201, | |
| "learning_rate": 4.7257445481643334e-06, | |
| "loss": 0.5723, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 1.1691022964509394, | |
| "grad_norm": 0.7389806509017944, | |
| "learning_rate": 4.723769581599109e-06, | |
| "loss": 0.5209, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 1.1722338204592901, | |
| "grad_norm": 3.053169012069702, | |
| "learning_rate": 4.721787945202472e-06, | |
| "loss": 0.6094, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 1.1753653444676408, | |
| "grad_norm": 1.288589596748352, | |
| "learning_rate": 4.719799644918017e-06, | |
| "loss": 0.5616, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 1.1784968684759916, | |
| "grad_norm": 0.7675042152404785, | |
| "learning_rate": 4.717804686709323e-06, | |
| "loss": 0.4963, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 1.1816283924843423, | |
| "grad_norm": 0.7246491312980652, | |
| "learning_rate": 4.715803076559938e-06, | |
| "loss": 0.5273, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 1.1847599164926932, | |
| "grad_norm": 0.8193361759185791, | |
| "learning_rate": 4.713794820473366e-06, | |
| "loss": 0.6107, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 1.187891440501044, | |
| "grad_norm": 0.9498510360717773, | |
| "learning_rate": 4.711779924473037e-06, | |
| "loss": 0.5421, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 1.1910229645093946, | |
| "grad_norm": 1.0479756593704224, | |
| "learning_rate": 4.709758394602305e-06, | |
| "loss": 0.5257, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 1.1941544885177453, | |
| "grad_norm": 0.907866895198822, | |
| "learning_rate": 4.707730236924413e-06, | |
| "loss": 0.5289, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 1.197286012526096, | |
| "grad_norm": 0.8861165642738342, | |
| "learning_rate": 4.705695457522488e-06, | |
| "loss": 0.5727, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 1.2004175365344467, | |
| "grad_norm": 0.7467761039733887, | |
| "learning_rate": 4.703654062499516e-06, | |
| "loss": 0.5602, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 1.2035490605427974, | |
| "grad_norm": 0.7456198334693909, | |
| "learning_rate": 4.701606057978325e-06, | |
| "loss": 0.5345, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 1.2066805845511483, | |
| "grad_norm": 1.9976060390472412, | |
| "learning_rate": 4.699551450101571e-06, | |
| "loss": 0.5504, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 1.209812108559499, | |
| "grad_norm": 1.5253807306289673, | |
| "learning_rate": 4.697490245031709e-06, | |
| "loss": 0.5568, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 1.2129436325678498, | |
| "grad_norm": 1.0786075592041016, | |
| "learning_rate": 4.6954224489509885e-06, | |
| "loss": 0.5564, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 1.2160751565762005, | |
| "grad_norm": 0.8385995030403137, | |
| "learning_rate": 4.693348068061422e-06, | |
| "loss": 0.5341, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 1.2192066805845512, | |
| "grad_norm": 0.8184949159622192, | |
| "learning_rate": 4.691267108584774e-06, | |
| "loss": 0.5614, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 1.2223382045929019, | |
| "grad_norm": 0.9964898824691772, | |
| "learning_rate": 4.68917957676254e-06, | |
| "loss": 0.5589, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 1.2254697286012526, | |
| "grad_norm": 1.0168914794921875, | |
| "learning_rate": 4.687085478855931e-06, | |
| "loss": 0.5892, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 1.2286012526096033, | |
| "grad_norm": 0.8841140866279602, | |
| "learning_rate": 4.684984821145846e-06, | |
| "loss": 0.5327, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 1.231732776617954, | |
| "grad_norm": 0.834431529045105, | |
| "learning_rate": 4.682877609932866e-06, | |
| "loss": 0.5594, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 1.2348643006263047, | |
| "grad_norm": 0.7256641983985901, | |
| "learning_rate": 4.6807638515372234e-06, | |
| "loss": 0.5443, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 1.2379958246346556, | |
| "grad_norm": 0.765096127986908, | |
| "learning_rate": 4.678643552298788e-06, | |
| "loss": 0.5439, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 1.2411273486430063, | |
| "grad_norm": 0.8760455846786499, | |
| "learning_rate": 4.676516718577051e-06, | |
| "loss": 0.5485, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 1.244258872651357, | |
| "grad_norm": 2.7111501693725586, | |
| "learning_rate": 4.674383356751099e-06, | |
| "loss": 0.5696, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 1.2473903966597077, | |
| "grad_norm": 1.0521738529205322, | |
| "learning_rate": 4.672243473219601e-06, | |
| "loss": 0.5503, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 1.2505219206680585, | |
| "grad_norm": 0.8909669518470764, | |
| "learning_rate": 4.670097074400785e-06, | |
| "loss": 0.5183, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 1.2536534446764092, | |
| "grad_norm": 0.7483847737312317, | |
| "learning_rate": 4.667944166732424e-06, | |
| "loss": 0.5669, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 1.2567849686847599, | |
| "grad_norm": 1.146997094154358, | |
| "learning_rate": 4.665784756671808e-06, | |
| "loss": 0.5464, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 1.2599164926931106, | |
| "grad_norm": 0.8998096585273743, | |
| "learning_rate": 4.663618850695733e-06, | |
| "loss": 0.5502, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 1.2630480167014615, | |
| "grad_norm": 0.8882688283920288, | |
| "learning_rate": 4.6614464553004795e-06, | |
| "loss": 0.5507, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 1.2661795407098122, | |
| "grad_norm": 0.8310684561729431, | |
| "learning_rate": 4.659267577001789e-06, | |
| "loss": 0.5164, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 1.269311064718163, | |
| "grad_norm": 0.9286114573478699, | |
| "learning_rate": 4.657082222334851e-06, | |
| "loss": 0.4813, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 1.2724425887265136, | |
| "grad_norm": 1.2394906282424927, | |
| "learning_rate": 4.654890397854275e-06, | |
| "loss": 0.5837, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 1.2755741127348643, | |
| "grad_norm": 4.00585412979126, | |
| "learning_rate": 4.652692110134079e-06, | |
| "loss": 0.5453, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 1.278705636743215, | |
| "grad_norm": 1.1667803525924683, | |
| "learning_rate": 4.650487365767667e-06, | |
| "loss": 0.5652, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 1.2818371607515657, | |
| "grad_norm": 0.9351289868354797, | |
| "learning_rate": 4.648276171367807e-06, | |
| "loss": 0.5576, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 1.2849686847599164, | |
| "grad_norm": 0.8107728958129883, | |
| "learning_rate": 4.646058533566614e-06, | |
| "loss": 0.5821, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 1.2881002087682671, | |
| "grad_norm": 0.7293011546134949, | |
| "learning_rate": 4.643834459015525e-06, | |
| "loss": 0.5363, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 1.2912317327766178, | |
| "grad_norm": 0.7550690770149231, | |
| "learning_rate": 4.641603954385289e-06, | |
| "loss": 0.53, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 1.2943632567849686, | |
| "grad_norm": 0.7626177072525024, | |
| "learning_rate": 4.639367026365938e-06, | |
| "loss": 0.5307, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 1.2974947807933195, | |
| "grad_norm": 1.0841096639633179, | |
| "learning_rate": 4.637123681666769e-06, | |
| "loss": 0.5162, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 1.3006263048016702, | |
| "grad_norm": 0.8814271092414856, | |
| "learning_rate": 4.634873927016326e-06, | |
| "loss": 0.5369, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 1.303757828810021, | |
| "grad_norm": 0.7402971982955933, | |
| "learning_rate": 4.632617769162378e-06, | |
| "loss": 0.5846, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 1.3068893528183716, | |
| "grad_norm": 0.8106061220169067, | |
| "learning_rate": 4.6303552148719e-06, | |
| "loss": 0.5289, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 1.3100208768267223, | |
| "grad_norm": 0.9241361618041992, | |
| "learning_rate": 4.628086270931053e-06, | |
| "loss": 0.5714, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 1.313152400835073, | |
| "grad_norm": 0.950332522392273, | |
| "learning_rate": 4.625810944145159e-06, | |
| "loss": 0.5817, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 1.316283924843424, | |
| "grad_norm": 0.9037718772888184, | |
| "learning_rate": 4.623529241338689e-06, | |
| "loss": 0.5902, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 1.3194154488517746, | |
| "grad_norm": 1.2110658884048462, | |
| "learning_rate": 4.621241169355234e-06, | |
| "loss": 0.561, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 1.3225469728601253, | |
| "grad_norm": 0.8582742214202881, | |
| "learning_rate": 4.618946735057491e-06, | |
| "loss": 0.5003, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 1.325678496868476, | |
| "grad_norm": 0.9203405976295471, | |
| "learning_rate": 4.6166459453272386e-06, | |
| "loss": 0.5639, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 1.3288100208768268, | |
| "grad_norm": 0.933721125125885, | |
| "learning_rate": 4.614338807065317e-06, | |
| "loss": 0.5766, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 1.3319415448851775, | |
| "grad_norm": 0.8435131311416626, | |
| "learning_rate": 4.612025327191608e-06, | |
| "loss": 0.5656, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 1.3350730688935282, | |
| "grad_norm": 0.795796811580658, | |
| "learning_rate": 4.609705512645015e-06, | |
| "loss": 0.4996, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 1.3382045929018789, | |
| "grad_norm": 0.8168228268623352, | |
| "learning_rate": 4.6073793703834404e-06, | |
| "loss": 0.5465, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 1.3413361169102296, | |
| "grad_norm": 0.8795569539070129, | |
| "learning_rate": 4.605046907383765e-06, | |
| "loss": 0.5407, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 1.3444676409185803, | |
| "grad_norm": 0.8504094481468201, | |
| "learning_rate": 4.6027081306418295e-06, | |
| "loss": 0.5589, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 1.347599164926931, | |
| "grad_norm": 1.485202431678772, | |
| "learning_rate": 4.600363047172409e-06, | |
| "loss": 0.5515, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 1.350730688935282, | |
| "grad_norm": 1.1156851053237915, | |
| "learning_rate": 4.598011664009197e-06, | |
| "loss": 0.5681, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 1.3538622129436326, | |
| "grad_norm": 0.8666876554489136, | |
| "learning_rate": 4.595653988204779e-06, | |
| "loss": 0.5451, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 1.3569937369519833, | |
| "grad_norm": 0.8192381858825684, | |
| "learning_rate": 4.593290026830619e-06, | |
| "loss": 0.5632, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 1.360125260960334, | |
| "grad_norm": 0.7994804978370667, | |
| "learning_rate": 4.590919786977029e-06, | |
| "loss": 0.5181, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 1.3632567849686847, | |
| "grad_norm": 0.8038607835769653, | |
| "learning_rate": 4.5885432757531535e-06, | |
| "loss": 0.5385, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 1.3663883089770354, | |
| "grad_norm": 0.7677503824234009, | |
| "learning_rate": 4.586160500286948e-06, | |
| "loss": 0.5455, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 1.3695198329853862, | |
| "grad_norm": 0.8293285369873047, | |
| "learning_rate": 4.583771467725157e-06, | |
| "loss": 0.5401, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 1.372651356993737, | |
| "grad_norm": 0.8607680797576904, | |
| "learning_rate": 4.581376185233289e-06, | |
| "loss": 0.5782, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 1.3757828810020878, | |
| "grad_norm": 0.8847081065177917, | |
| "learning_rate": 4.578974659995601e-06, | |
| "loss": 0.572, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 1.3789144050104385, | |
| "grad_norm": 0.7669641971588135, | |
| "learning_rate": 4.576566899215075e-06, | |
| "loss": 0.5655, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 1.3820459290187892, | |
| "grad_norm": 0.8514629006385803, | |
| "learning_rate": 4.5741529101133904e-06, | |
| "loss": 0.5218, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 1.38517745302714, | |
| "grad_norm": 0.8719842433929443, | |
| "learning_rate": 4.5717326999309145e-06, | |
| "loss": 0.5579, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 1.3883089770354906, | |
| "grad_norm": 1.1142809391021729, | |
| "learning_rate": 4.569306275926667e-06, | |
| "loss": 0.5535, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 1.3914405010438413, | |
| "grad_norm": 0.7392387986183167, | |
| "learning_rate": 4.566873645378309e-06, | |
| "loss": 0.5335, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 1.394572025052192, | |
| "grad_norm": 0.9066658616065979, | |
| "learning_rate": 4.564434815582117e-06, | |
| "loss": 0.5286, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 1.3977035490605427, | |
| "grad_norm": 0.8648932576179504, | |
| "learning_rate": 4.561989793852959e-06, | |
| "loss": 0.5008, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 1.4008350730688934, | |
| "grad_norm": 0.7768712043762207, | |
| "learning_rate": 4.559538587524276e-06, | |
| "loss": 0.5727, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 1.4039665970772441, | |
| "grad_norm": 0.7851182222366333, | |
| "learning_rate": 4.557081203948059e-06, | |
| "loss": 0.5731, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 1.407098121085595, | |
| "grad_norm": 0.8959861397743225, | |
| "learning_rate": 4.5546176504948255e-06, | |
| "loss": 0.5587, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 1.4102296450939458, | |
| "grad_norm": 1.0538026094436646, | |
| "learning_rate": 4.552147934553601e-06, | |
| "loss": 0.5808, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 1.4133611691022965, | |
| "grad_norm": 0.9887629151344299, | |
| "learning_rate": 4.54967206353189e-06, | |
| "loss": 0.5658, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 1.4164926931106472, | |
| "grad_norm": 0.9579302072525024, | |
| "learning_rate": 4.547190044855663e-06, | |
| "loss": 0.5092, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 1.4196242171189979, | |
| "grad_norm": 0.6993522047996521, | |
| "learning_rate": 4.544701885969326e-06, | |
| "loss": 0.5233, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 1.4227557411273486, | |
| "grad_norm": 0.8197568655014038, | |
| "learning_rate": 4.542207594335703e-06, | |
| "loss": 0.553, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 1.4258872651356993, | |
| "grad_norm": 2.921947717666626, | |
| "learning_rate": 4.53970717743601e-06, | |
| "loss": 0.4857, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 1.4290187891440502, | |
| "grad_norm": 1.3547242879867554, | |
| "learning_rate": 4.53720064276984e-06, | |
| "loss": 0.5676, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 1.432150313152401, | |
| "grad_norm": 1.4175567626953125, | |
| "learning_rate": 4.534687997855131e-06, | |
| "loss": 0.5164, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 1.4352818371607516, | |
| "grad_norm": 1.378146767616272, | |
| "learning_rate": 4.532169250228145e-06, | |
| "loss": 0.5429, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 1.4384133611691023, | |
| "grad_norm": 0.7811698317527771, | |
| "learning_rate": 4.529644407443456e-06, | |
| "loss": 0.524, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 1.441544885177453, | |
| "grad_norm": 1.1481678485870361, | |
| "learning_rate": 4.527113477073914e-06, | |
| "loss": 0.5513, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 1.4446764091858038, | |
| "grad_norm": 0.8450161218643188, | |
| "learning_rate": 4.5245764667106266e-06, | |
| "loss": 0.5632, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 1.4478079331941545, | |
| "grad_norm": 1.1582145690917969, | |
| "learning_rate": 4.522033383962941e-06, | |
| "loss": 0.5834, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 1.4509394572025052, | |
| "grad_norm": 1.0403447151184082, | |
| "learning_rate": 4.519484236458416e-06, | |
| "loss": 0.506, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 1.4540709812108559, | |
| "grad_norm": 0.7894920706748962, | |
| "learning_rate": 4.516929031842799e-06, | |
| "loss": 0.5526, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 1.4572025052192066, | |
| "grad_norm": 0.8092262744903564, | |
| "learning_rate": 4.51436777778001e-06, | |
| "loss": 0.5619, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 1.4603340292275573, | |
| "grad_norm": 0.9773806929588318, | |
| "learning_rate": 4.511800481952106e-06, | |
| "loss": 0.5179, | |
| "step": 467 | |
| }, | |
| { | |
| "epoch": 1.4634655532359082, | |
| "grad_norm": 1.018676519393921, | |
| "learning_rate": 4.509227152059271e-06, | |
| "loss": 0.5415, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 1.466597077244259, | |
| "grad_norm": 0.7457838654518127, | |
| "learning_rate": 4.506647795819784e-06, | |
| "loss": 0.5473, | |
| "step": 469 | |
| }, | |
| { | |
| "epoch": 1.4697286012526096, | |
| "grad_norm": 0.7826436161994934, | |
| "learning_rate": 4.50406242097e-06, | |
| "loss": 0.5526, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 1.4728601252609603, | |
| "grad_norm": 0.9492483139038086, | |
| "learning_rate": 4.501471035264328e-06, | |
| "loss": 0.5179, | |
| "step": 471 | |
| }, | |
| { | |
| "epoch": 1.475991649269311, | |
| "grad_norm": 0.93398517370224, | |
| "learning_rate": 4.4988736464752005e-06, | |
| "loss": 0.5195, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 1.4791231732776617, | |
| "grad_norm": 0.8396487832069397, | |
| "learning_rate": 4.496270262393061e-06, | |
| "loss": 0.5447, | |
| "step": 473 | |
| }, | |
| { | |
| "epoch": 1.4822546972860124, | |
| "grad_norm": 0.7450584173202515, | |
| "learning_rate": 4.4936608908263315e-06, | |
| "loss": 0.5207, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 1.4853862212943634, | |
| "grad_norm": 0.7887717485427856, | |
| "learning_rate": 4.491045539601392e-06, | |
| "loss": 0.523, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 1.488517745302714, | |
| "grad_norm": 1.2051388025283813, | |
| "learning_rate": 4.48842421656256e-06, | |
| "loss": 0.5402, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 1.4916492693110648, | |
| "grad_norm": 2.3103389739990234, | |
| "learning_rate": 4.485796929572063e-06, | |
| "loss": 0.5588, | |
| "step": 477 | |
| }, | |
| { | |
| "epoch": 1.4947807933194155, | |
| "grad_norm": 0.7473112344741821, | |
| "learning_rate": 4.483163686510016e-06, | |
| "loss": 0.5731, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 1.4979123173277662, | |
| "grad_norm": 0.7545126676559448, | |
| "learning_rate": 4.480524495274399e-06, | |
| "loss": 0.5536, | |
| "step": 479 | |
| }, | |
| { | |
| "epoch": 1.501043841336117, | |
| "grad_norm": 0.7801297903060913, | |
| "learning_rate": 4.477879363781033e-06, | |
| "loss": 0.5696, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 1.5041753653444676, | |
| "grad_norm": 0.7740563750267029, | |
| "learning_rate": 4.475228299963554e-06, | |
| "loss": 0.5526, | |
| "step": 481 | |
| }, | |
| { | |
| "epoch": 1.5073068893528183, | |
| "grad_norm": 0.8600060343742371, | |
| "learning_rate": 4.4725713117733936e-06, | |
| "loss": 0.5051, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 1.510438413361169, | |
| "grad_norm": 0.6934283971786499, | |
| "learning_rate": 4.46990840717975e-06, | |
| "loss": 0.5564, | |
| "step": 483 | |
| }, | |
| { | |
| "epoch": 1.5135699373695197, | |
| "grad_norm": 0.8927920460700989, | |
| "learning_rate": 4.46723959416957e-06, | |
| "loss": 0.5529, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 1.5167014613778704, | |
| "grad_norm": 0.9570988416671753, | |
| "learning_rate": 4.464564880747517e-06, | |
| "loss": 0.5661, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 1.5198329853862211, | |
| "grad_norm": 0.7229202389717102, | |
| "learning_rate": 4.461884274935956e-06, | |
| "loss": 0.5964, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 1.522964509394572, | |
| "grad_norm": 0.7367239594459534, | |
| "learning_rate": 4.4591977847749225e-06, | |
| "loss": 0.5455, | |
| "step": 487 | |
| }, | |
| { | |
| "epoch": 1.5260960334029228, | |
| "grad_norm": 0.8062120676040649, | |
| "learning_rate": 4.456505418322103e-06, | |
| "loss": 0.5735, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 1.5292275574112735, | |
| "grad_norm": 0.8854482769966125, | |
| "learning_rate": 4.453807183652808e-06, | |
| "loss": 0.5421, | |
| "step": 489 | |
| }, | |
| { | |
| "epoch": 1.5323590814196242, | |
| "grad_norm": 0.7518959045410156, | |
| "learning_rate": 4.451103088859951e-06, | |
| "loss": 0.5083, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 1.535490605427975, | |
| "grad_norm": 0.8621206879615784, | |
| "learning_rate": 4.448393142054016e-06, | |
| "loss": 0.4712, | |
| "step": 491 | |
| }, | |
| { | |
| "epoch": 1.5386221294363258, | |
| "grad_norm": 1.0618741512298584, | |
| "learning_rate": 4.445677351363046e-06, | |
| "loss": 0.5808, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 1.5417536534446765, | |
| "grad_norm": 0.8261345028877258, | |
| "learning_rate": 4.442955724932607e-06, | |
| "loss": 0.5625, | |
| "step": 493 | |
| }, | |
| { | |
| "epoch": 1.5448851774530272, | |
| "grad_norm": 0.7067139744758606, | |
| "learning_rate": 4.440228270925772e-06, | |
| "loss": 0.5661, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 1.548016701461378, | |
| "grad_norm": 0.9234416484832764, | |
| "learning_rate": 4.437494997523091e-06, | |
| "loss": 0.5428, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 1.5511482254697286, | |
| "grad_norm": 0.9273470044136047, | |
| "learning_rate": 4.434755912922567e-06, | |
| "loss": 0.5388, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 1.5542797494780793, | |
| "grad_norm": 1.0163263082504272, | |
| "learning_rate": 4.4320110253396345e-06, | |
| "loss": 0.5409, | |
| "step": 497 | |
| }, | |
| { | |
| "epoch": 1.55741127348643, | |
| "grad_norm": 0.9542096853256226, | |
| "learning_rate": 4.429260343007133e-06, | |
| "loss": 0.5329, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 1.5605427974947808, | |
| "grad_norm": 0.8076801896095276, | |
| "learning_rate": 4.426503874175283e-06, | |
| "loss": 0.5616, | |
| "step": 499 | |
| }, | |
| { | |
| "epoch": 1.5636743215031315, | |
| "grad_norm": 1.0063767433166504, | |
| "learning_rate": 4.423741627111658e-06, | |
| "loss": 0.5369, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 1.5668058455114822, | |
| "grad_norm": 1.040286898612976, | |
| "learning_rate": 4.420973610101166e-06, | |
| "loss": 0.5474, | |
| "step": 501 | |
| }, | |
| { | |
| "epoch": 1.5699373695198329, | |
| "grad_norm": 0.7832860946655273, | |
| "learning_rate": 4.4181998314460164e-06, | |
| "loss": 0.5486, | |
| "step": 502 | |
| }, | |
| { | |
| "epoch": 1.5730688935281836, | |
| "grad_norm": 0.8162257075309753, | |
| "learning_rate": 4.415420299465706e-06, | |
| "loss": 0.5054, | |
| "step": 503 | |
| }, | |
| { | |
| "epoch": 1.5762004175365343, | |
| "grad_norm": 0.9108433127403259, | |
| "learning_rate": 4.4126350224969814e-06, | |
| "loss": 0.5399, | |
| "step": 504 | |
| }, | |
| { | |
| "epoch": 1.5793319415448852, | |
| "grad_norm": 0.8002520799636841, | |
| "learning_rate": 4.409844008893824e-06, | |
| "loss": 0.5485, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 1.582463465553236, | |
| "grad_norm": 0.8543248772621155, | |
| "learning_rate": 4.407047267027423e-06, | |
| "loss": 0.4984, | |
| "step": 506 | |
| }, | |
| { | |
| "epoch": 1.5855949895615866, | |
| "grad_norm": 0.7154155373573303, | |
| "learning_rate": 4.404244805286141e-06, | |
| "loss": 0.5392, | |
| "step": 507 | |
| }, | |
| { | |
| "epoch": 1.5887265135699373, | |
| "grad_norm": 0.818553626537323, | |
| "learning_rate": 4.401436632075504e-06, | |
| "loss": 0.5178, | |
| "step": 508 | |
| }, | |
| { | |
| "epoch": 1.5918580375782883, | |
| "grad_norm": 0.7535017728805542, | |
| "learning_rate": 4.398622755818167e-06, | |
| "loss": 0.5446, | |
| "step": 509 | |
| }, | |
| { | |
| "epoch": 1.594989561586639, | |
| "grad_norm": 0.9328975677490234, | |
| "learning_rate": 4.395803184953889e-06, | |
| "loss": 0.5546, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 1.5981210855949897, | |
| "grad_norm": 0.7960026860237122, | |
| "learning_rate": 4.392977927939508e-06, | |
| "loss": 0.5451, | |
| "step": 511 | |
| }, | |
| { | |
| "epoch": 1.6012526096033404, | |
| "grad_norm": 0.9686267971992493, | |
| "learning_rate": 4.3901469932489195e-06, | |
| "loss": 0.5198, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 1.604384133611691, | |
| "grad_norm": 0.903137743473053, | |
| "learning_rate": 4.387310389373047e-06, | |
| "loss": 0.5395, | |
| "step": 513 | |
| }, | |
| { | |
| "epoch": 1.6075156576200418, | |
| "grad_norm": 1.0728516578674316, | |
| "learning_rate": 4.384468124819816e-06, | |
| "loss": 0.5843, | |
| "step": 514 | |
| }, | |
| { | |
| "epoch": 1.6106471816283925, | |
| "grad_norm": 1.0245436429977417, | |
| "learning_rate": 4.3816202081141345e-06, | |
| "loss": 0.5672, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 1.6137787056367432, | |
| "grad_norm": 0.9672732353210449, | |
| "learning_rate": 4.378766647797858e-06, | |
| "loss": 0.5369, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 1.616910229645094, | |
| "grad_norm": 0.9149513840675354, | |
| "learning_rate": 4.375907452429774e-06, | |
| "loss": 0.4628, | |
| "step": 517 | |
| }, | |
| { | |
| "epoch": 1.6200417536534446, | |
| "grad_norm": 0.7543843984603882, | |
| "learning_rate": 4.373042630585567e-06, | |
| "loss": 0.5344, | |
| "step": 518 | |
| }, | |
| { | |
| "epoch": 1.6231732776617953, | |
| "grad_norm": 0.7589017152786255, | |
| "learning_rate": 4.370172190857801e-06, | |
| "loss": 0.5672, | |
| "step": 519 | |
| }, | |
| { | |
| "epoch": 1.626304801670146, | |
| "grad_norm": 0.803040623664856, | |
| "learning_rate": 4.367296141855887e-06, | |
| "loss": 0.5313, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 1.6294363256784967, | |
| "grad_norm": 0.8305794596672058, | |
| "learning_rate": 4.3644144922060625e-06, | |
| "loss": 0.5754, | |
| "step": 521 | |
| }, | |
| { | |
| "epoch": 1.6325678496868476, | |
| "grad_norm": 1.0086486339569092, | |
| "learning_rate": 4.361527250551361e-06, | |
| "loss": 0.5433, | |
| "step": 522 | |
| }, | |
| { | |
| "epoch": 1.6356993736951984, | |
| "grad_norm": 0.7217550277709961, | |
| "learning_rate": 4.35863442555159e-06, | |
| "loss": 0.524, | |
| "step": 523 | |
| }, | |
| { | |
| "epoch": 1.638830897703549, | |
| "grad_norm": 0.7788524627685547, | |
| "learning_rate": 4.355736025883303e-06, | |
| "loss": 0.536, | |
| "step": 524 | |
| }, | |
| { | |
| "epoch": 1.6419624217118998, | |
| "grad_norm": 0.8460550904273987, | |
| "learning_rate": 4.352832060239774e-06, | |
| "loss": 0.5381, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 1.6450939457202505, | |
| "grad_norm": 0.7571215033531189, | |
| "learning_rate": 4.3499225373309675e-06, | |
| "loss": 0.541, | |
| "step": 526 | |
| }, | |
| { | |
| "epoch": 1.6482254697286014, | |
| "grad_norm": 0.7343226671218872, | |
| "learning_rate": 4.347007465883523e-06, | |
| "loss": 0.5147, | |
| "step": 527 | |
| }, | |
| { | |
| "epoch": 1.651356993736952, | |
| "grad_norm": 0.7271892428398132, | |
| "learning_rate": 4.3440868546407165e-06, | |
| "loss": 0.5311, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 1.6544885177453028, | |
| "grad_norm": 0.8166136741638184, | |
| "learning_rate": 4.341160712362442e-06, | |
| "loss": 0.5379, | |
| "step": 529 | |
| }, | |
| { | |
| "epoch": 1.6576200417536535, | |
| "grad_norm": 1.5985233783721924, | |
| "learning_rate": 4.338229047825182e-06, | |
| "loss": 0.5782, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 1.6607515657620042, | |
| "grad_norm": 0.7835702896118164, | |
| "learning_rate": 4.3352918698219835e-06, | |
| "loss": 0.525, | |
| "step": 531 | |
| }, | |
| { | |
| "epoch": 1.663883089770355, | |
| "grad_norm": 0.7278687953948975, | |
| "learning_rate": 4.332349187162428e-06, | |
| "loss": 0.5266, | |
| "step": 532 | |
| }, | |
| { | |
| "epoch": 1.6670146137787056, | |
| "grad_norm": 0.8240190148353577, | |
| "learning_rate": 4.329401008672608e-06, | |
| "loss": 0.5515, | |
| "step": 533 | |
| }, | |
| { | |
| "epoch": 1.6701461377870563, | |
| "grad_norm": 0.9447080492973328, | |
| "learning_rate": 4.326447343195102e-06, | |
| "loss": 0.5596, | |
| "step": 534 | |
| }, | |
| { | |
| "epoch": 1.673277661795407, | |
| "grad_norm": 0.7827372550964355, | |
| "learning_rate": 4.323488199588944e-06, | |
| "loss": 0.5466, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 1.6764091858037578, | |
| "grad_norm": 0.9252517223358154, | |
| "learning_rate": 4.320523586729599e-06, | |
| "loss": 0.5433, | |
| "step": 536 | |
| }, | |
| { | |
| "epoch": 1.6795407098121085, | |
| "grad_norm": 0.9437504410743713, | |
| "learning_rate": 4.317553513508934e-06, | |
| "loss": 0.5552, | |
| "step": 537 | |
| }, | |
| { | |
| "epoch": 1.6826722338204592, | |
| "grad_norm": 0.8972746133804321, | |
| "learning_rate": 4.3145779888351986e-06, | |
| "loss": 0.5259, | |
| "step": 538 | |
| }, | |
| { | |
| "epoch": 1.6858037578288099, | |
| "grad_norm": 0.8017446994781494, | |
| "learning_rate": 4.311597021632988e-06, | |
| "loss": 0.5263, | |
| "step": 539 | |
| }, | |
| { | |
| "epoch": 1.6889352818371608, | |
| "grad_norm": 0.7875497341156006, | |
| "learning_rate": 4.3086106208432235e-06, | |
| "loss": 0.5316, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 1.6920668058455115, | |
| "grad_norm": 0.8204905986785889, | |
| "learning_rate": 4.305618795423125e-06, | |
| "loss": 0.5506, | |
| "step": 541 | |
| }, | |
| { | |
| "epoch": 1.6951983298538622, | |
| "grad_norm": 0.888359785079956, | |
| "learning_rate": 4.30262155434618e-06, | |
| "loss": 0.4825, | |
| "step": 542 | |
| }, | |
| { | |
| "epoch": 1.698329853862213, | |
| "grad_norm": 1.1026058197021484, | |
| "learning_rate": 4.29961890660212e-06, | |
| "loss": 0.5321, | |
| "step": 543 | |
| }, | |
| { | |
| "epoch": 1.7014613778705638, | |
| "grad_norm": 0.7662535905838013, | |
| "learning_rate": 4.2966108611968945e-06, | |
| "loss": 0.5432, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 1.7045929018789145, | |
| "grad_norm": 1.1951749324798584, | |
| "learning_rate": 4.293597427152641e-06, | |
| "loss": 0.5123, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 1.7077244258872653, | |
| "grad_norm": 1.303183913230896, | |
| "learning_rate": 4.290578613507661e-06, | |
| "loss": 0.5346, | |
| "step": 546 | |
| }, | |
| { | |
| "epoch": 1.710855949895616, | |
| "grad_norm": 0.7653357982635498, | |
| "learning_rate": 4.287554429316387e-06, | |
| "loss": 0.5397, | |
| "step": 547 | |
| }, | |
| { | |
| "epoch": 1.7139874739039667, | |
| "grad_norm": 0.796215295791626, | |
| "learning_rate": 4.284524883649366e-06, | |
| "loss": 0.5421, | |
| "step": 548 | |
| }, | |
| { | |
| "epoch": 1.7171189979123174, | |
| "grad_norm": 0.7599332332611084, | |
| "learning_rate": 4.281489985593219e-06, | |
| "loss": 0.5289, | |
| "step": 549 | |
| }, | |
| { | |
| "epoch": 1.720250521920668, | |
| "grad_norm": 0.8029115796089172, | |
| "learning_rate": 4.2784497442506265e-06, | |
| "loss": 0.5409, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 1.7233820459290188, | |
| "grad_norm": 0.7194099426269531, | |
| "learning_rate": 4.275404168740291e-06, | |
| "loss": 0.5327, | |
| "step": 551 | |
| }, | |
| { | |
| "epoch": 1.7265135699373695, | |
| "grad_norm": 0.7960740923881531, | |
| "learning_rate": 4.272353268196917e-06, | |
| "loss": 0.4896, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 1.7296450939457202, | |
| "grad_norm": 0.9572116732597351, | |
| "learning_rate": 4.269297051771178e-06, | |
| "loss": 0.5402, | |
| "step": 553 | |
| }, | |
| { | |
| "epoch": 1.732776617954071, | |
| "grad_norm": 1.3604938983917236, | |
| "learning_rate": 4.266235528629695e-06, | |
| "loss": 0.5792, | |
| "step": 554 | |
| }, | |
| { | |
| "epoch": 1.7359081419624216, | |
| "grad_norm": 2.067286729812622, | |
| "learning_rate": 4.263168707955002e-06, | |
| "loss": 0.5033, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 1.7390396659707723, | |
| "grad_norm": 0.8031097054481506, | |
| "learning_rate": 4.260096598945523e-06, | |
| "loss": 0.5117, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 1.742171189979123, | |
| "grad_norm": 1.0241729021072388, | |
| "learning_rate": 4.257019210815546e-06, | |
| "loss": 0.5359, | |
| "step": 557 | |
| }, | |
| { | |
| "epoch": 1.745302713987474, | |
| "grad_norm": 0.7625218629837036, | |
| "learning_rate": 4.25393655279519e-06, | |
| "loss": 0.5625, | |
| "step": 558 | |
| }, | |
| { | |
| "epoch": 1.7484342379958246, | |
| "grad_norm": 0.8603503704071045, | |
| "learning_rate": 4.250848634130381e-06, | |
| "loss": 0.5043, | |
| "step": 559 | |
| }, | |
| { | |
| "epoch": 1.7515657620041754, | |
| "grad_norm": 0.9543750286102295, | |
| "learning_rate": 4.247755464082824e-06, | |
| "loss": 0.5364, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 1.754697286012526, | |
| "grad_norm": 0.9707463979721069, | |
| "learning_rate": 4.244657051929973e-06, | |
| "loss": 0.5184, | |
| "step": 561 | |
| }, | |
| { | |
| "epoch": 1.757828810020877, | |
| "grad_norm": 0.7491432428359985, | |
| "learning_rate": 4.241553406965008e-06, | |
| "loss": 0.559, | |
| "step": 562 | |
| }, | |
| { | |
| "epoch": 1.7609603340292277, | |
| "grad_norm": 0.7444972991943359, | |
| "learning_rate": 4.238444538496801e-06, | |
| "loss": 0.5327, | |
| "step": 563 | |
| }, | |
| { | |
| "epoch": 1.7640918580375784, | |
| "grad_norm": 2.7108678817749023, | |
| "learning_rate": 4.235330455849892e-06, | |
| "loss": 0.55, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 1.767223382045929, | |
| "grad_norm": 1.6716049909591675, | |
| "learning_rate": 4.232211168364459e-06, | |
| "loss": 0.5093, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 1.7703549060542798, | |
| "grad_norm": 0.7023475170135498, | |
| "learning_rate": 4.229086685396295e-06, | |
| "loss": 0.569, | |
| "step": 566 | |
| }, | |
| { | |
| "epoch": 1.7734864300626305, | |
| "grad_norm": 0.8596265316009521, | |
| "learning_rate": 4.225957016316771e-06, | |
| "loss": 0.5128, | |
| "step": 567 | |
| }, | |
| { | |
| "epoch": 1.7766179540709812, | |
| "grad_norm": 0.8110849857330322, | |
| "learning_rate": 4.222822170512816e-06, | |
| "loss": 0.5142, | |
| "step": 568 | |
| }, | |
| { | |
| "epoch": 1.779749478079332, | |
| "grad_norm": 0.7583725452423096, | |
| "learning_rate": 4.219682157386884e-06, | |
| "loss": 0.5584, | |
| "step": 569 | |
| }, | |
| { | |
| "epoch": 1.7828810020876826, | |
| "grad_norm": 0.787811279296875, | |
| "learning_rate": 4.21653698635693e-06, | |
| "loss": 0.5068, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 1.7860125260960333, | |
| "grad_norm": 0.8298993110656738, | |
| "learning_rate": 4.213386666856375e-06, | |
| "loss": 0.5496, | |
| "step": 571 | |
| }, | |
| { | |
| "epoch": 1.789144050104384, | |
| "grad_norm": 0.8999841213226318, | |
| "learning_rate": 4.210231208334087e-06, | |
| "loss": 0.5454, | |
| "step": 572 | |
| }, | |
| { | |
| "epoch": 1.7922755741127347, | |
| "grad_norm": 4.264521598815918, | |
| "learning_rate": 4.207070620254345e-06, | |
| "loss": 0.5486, | |
| "step": 573 | |
| }, | |
| { | |
| "epoch": 1.7954070981210855, | |
| "grad_norm": 0.8517448306083679, | |
| "learning_rate": 4.203904912096812e-06, | |
| "loss": 0.5566, | |
| "step": 574 | |
| }, | |
| { | |
| "epoch": 1.7985386221294362, | |
| "grad_norm": 0.9230182766914368, | |
| "learning_rate": 4.200734093356511e-06, | |
| "loss": 0.4964, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 1.801670146137787, | |
| "grad_norm": 1.224039912223816, | |
| "learning_rate": 4.197558173543791e-06, | |
| "loss": 0.5356, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 1.8048016701461378, | |
| "grad_norm": 0.9998573660850525, | |
| "learning_rate": 4.194377162184301e-06, | |
| "loss": 0.5334, | |
| "step": 577 | |
| }, | |
| { | |
| "epoch": 1.8079331941544885, | |
| "grad_norm": 0.865521252155304, | |
| "learning_rate": 4.191191068818963e-06, | |
| "loss": 0.5036, | |
| "step": 578 | |
| }, | |
| { | |
| "epoch": 1.8110647181628392, | |
| "grad_norm": 0.8048138618469238, | |
| "learning_rate": 4.18799990300394e-06, | |
| "loss": 0.4979, | |
| "step": 579 | |
| }, | |
| { | |
| "epoch": 1.8141962421711901, | |
| "grad_norm": 0.717815637588501, | |
| "learning_rate": 4.184803674310609e-06, | |
| "loss": 0.5623, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 1.8173277661795408, | |
| "grad_norm": 0.8403327465057373, | |
| "learning_rate": 4.1816023923255335e-06, | |
| "loss": 0.5055, | |
| "step": 581 | |
| }, | |
| { | |
| "epoch": 1.8204592901878915, | |
| "grad_norm": 0.7298995852470398, | |
| "learning_rate": 4.178396066650432e-06, | |
| "loss": 0.5641, | |
| "step": 582 | |
| }, | |
| { | |
| "epoch": 1.8235908141962422, | |
| "grad_norm": 0.9469727873802185, | |
| "learning_rate": 4.1751847069021516e-06, | |
| "loss": 0.5557, | |
| "step": 583 | |
| }, | |
| { | |
| "epoch": 1.826722338204593, | |
| "grad_norm": 0.8641784191131592, | |
| "learning_rate": 4.1719683227126386e-06, | |
| "loss": 0.5153, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 1.8298538622129437, | |
| "grad_norm": 0.7316668629646301, | |
| "learning_rate": 4.168746923728908e-06, | |
| "loss": 0.4988, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 1.8329853862212944, | |
| "grad_norm": 0.8795468807220459, | |
| "learning_rate": 4.165520519613017e-06, | |
| "loss": 0.5483, | |
| "step": 586 | |
| }, | |
| { | |
| "epoch": 1.836116910229645, | |
| "grad_norm": 0.7323560118675232, | |
| "learning_rate": 4.162289120042034e-06, | |
| "loss": 0.5194, | |
| "step": 587 | |
| }, | |
| { | |
| "epoch": 1.8392484342379958, | |
| "grad_norm": 0.8217021822929382, | |
| "learning_rate": 4.159052734708013e-06, | |
| "loss": 0.532, | |
| "step": 588 | |
| }, | |
| { | |
| "epoch": 1.8423799582463465, | |
| "grad_norm": 0.7669674754142761, | |
| "learning_rate": 4.155811373317958e-06, | |
| "loss": 0.541, | |
| "step": 589 | |
| }, | |
| { | |
| "epoch": 1.8455114822546972, | |
| "grad_norm": 0.8312156200408936, | |
| "learning_rate": 4.152565045593801e-06, | |
| "loss": 0.5298, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 1.848643006263048, | |
| "grad_norm": 0.8967565298080444, | |
| "learning_rate": 4.1493137612723665e-06, | |
| "loss": 0.51, | |
| "step": 591 | |
| }, | |
| { | |
| "epoch": 1.8517745302713986, | |
| "grad_norm": 0.8706664443016052, | |
| "learning_rate": 4.14605753010535e-06, | |
| "loss": 0.4941, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 1.8549060542797495, | |
| "grad_norm": 0.7585753798484802, | |
| "learning_rate": 4.14279636185928e-06, | |
| "loss": 0.5161, | |
| "step": 593 | |
| }, | |
| { | |
| "epoch": 1.8580375782881002, | |
| "grad_norm": 0.7495241165161133, | |
| "learning_rate": 4.1395302663154954e-06, | |
| "loss": 0.5388, | |
| "step": 594 | |
| }, | |
| { | |
| "epoch": 1.861169102296451, | |
| "grad_norm": 1.0746862888336182, | |
| "learning_rate": 4.136259253270114e-06, | |
| "loss": 0.4976, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 1.8643006263048016, | |
| "grad_norm": 0.872309684753418, | |
| "learning_rate": 4.132983332534e-06, | |
| "loss": 0.559, | |
| "step": 596 | |
| }, | |
| { | |
| "epoch": 1.8674321503131524, | |
| "grad_norm": 0.8759891986846924, | |
| "learning_rate": 4.1297025139327405e-06, | |
| "loss": 0.5436, | |
| "step": 597 | |
| }, | |
| { | |
| "epoch": 1.8705636743215033, | |
| "grad_norm": 1.1044493913650513, | |
| "learning_rate": 4.126416807306611e-06, | |
| "loss": 0.5476, | |
| "step": 598 | |
| }, | |
| { | |
| "epoch": 1.873695198329854, | |
| "grad_norm": 0.8340442180633545, | |
| "learning_rate": 4.123126222510549e-06, | |
| "loss": 0.4592, | |
| "step": 599 | |
| }, | |
| { | |
| "epoch": 1.8768267223382047, | |
| "grad_norm": 0.8331449031829834, | |
| "learning_rate": 4.119830769414123e-06, | |
| "loss": 0.5219, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 1.8799582463465554, | |
| "grad_norm": 1.0862973928451538, | |
| "learning_rate": 4.116530457901503e-06, | |
| "loss": 0.5159, | |
| "step": 601 | |
| }, | |
| { | |
| "epoch": 1.883089770354906, | |
| "grad_norm": 0.8524414300918579, | |
| "learning_rate": 4.113225297871431e-06, | |
| "loss": 0.5502, | |
| "step": 602 | |
| }, | |
| { | |
| "epoch": 1.8862212943632568, | |
| "grad_norm": 1.4945416450500488, | |
| "learning_rate": 4.10991529923719e-06, | |
| "loss": 0.5627, | |
| "step": 603 | |
| }, | |
| { | |
| "epoch": 1.8893528183716075, | |
| "grad_norm": 1.5518157482147217, | |
| "learning_rate": 4.10660047192658e-06, | |
| "loss": 0.5517, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 1.8924843423799582, | |
| "grad_norm": 2.56638765335083, | |
| "learning_rate": 4.103280825881878e-06, | |
| "loss": 0.5422, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 1.895615866388309, | |
| "grad_norm": 0.867254912853241, | |
| "learning_rate": 4.099956371059817e-06, | |
| "loss": 0.4991, | |
| "step": 606 | |
| }, | |
| { | |
| "epoch": 1.8987473903966596, | |
| "grad_norm": 0.9555892944335938, | |
| "learning_rate": 4.096627117431554e-06, | |
| "loss": 0.5339, | |
| "step": 607 | |
| }, | |
| { | |
| "epoch": 1.9018789144050103, | |
| "grad_norm": 0.7905483245849609, | |
| "learning_rate": 4.093293074982638e-06, | |
| "loss": 0.5168, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 1.905010438413361, | |
| "grad_norm": 0.7500227093696594, | |
| "learning_rate": 4.089954253712981e-06, | |
| "loss": 0.5096, | |
| "step": 609 | |
| }, | |
| { | |
| "epoch": 1.9081419624217117, | |
| "grad_norm": 0.8458324074745178, | |
| "learning_rate": 4.086610663636828e-06, | |
| "loss": 0.5296, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 1.9112734864300627, | |
| "grad_norm": 0.7392706871032715, | |
| "learning_rate": 4.08326231478273e-06, | |
| "loss": 0.5305, | |
| "step": 611 | |
| }, | |
| { | |
| "epoch": 1.9144050104384134, | |
| "grad_norm": 0.8113343715667725, | |
| "learning_rate": 4.079909217193508e-06, | |
| "loss": 0.5044, | |
| "step": 612 | |
| }, | |
| { | |
| "epoch": 1.917536534446764, | |
| "grad_norm": 0.7637801766395569, | |
| "learning_rate": 4.076551380926226e-06, | |
| "loss": 0.5298, | |
| "step": 613 | |
| }, | |
| { | |
| "epoch": 1.9206680584551148, | |
| "grad_norm": 1.0523375272750854, | |
| "learning_rate": 4.073188816052164e-06, | |
| "loss": 0.5111, | |
| "step": 614 | |
| }, | |
| { | |
| "epoch": 1.9237995824634657, | |
| "grad_norm": 0.8224868774414062, | |
| "learning_rate": 4.069821532656781e-06, | |
| "loss": 0.5178, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 1.9269311064718164, | |
| "grad_norm": 0.7270777821540833, | |
| "learning_rate": 4.066449540839693e-06, | |
| "loss": 0.5307, | |
| "step": 616 | |
| }, | |
| { | |
| "epoch": 1.9300626304801671, | |
| "grad_norm": 0.7214602828025818, | |
| "learning_rate": 4.063072850714631e-06, | |
| "loss": 0.5171, | |
| "step": 617 | |
| }, | |
| { | |
| "epoch": 1.9331941544885178, | |
| "grad_norm": 0.7333671450614929, | |
| "learning_rate": 4.059691472409426e-06, | |
| "loss": 0.56, | |
| "step": 618 | |
| }, | |
| { | |
| "epoch": 1.9363256784968685, | |
| "grad_norm": 0.9166824221611023, | |
| "learning_rate": 4.056305416065964e-06, | |
| "loss": 0.5388, | |
| "step": 619 | |
| }, | |
| { | |
| "epoch": 1.9394572025052192, | |
| "grad_norm": 0.7743303775787354, | |
| "learning_rate": 4.052914691840167e-06, | |
| "loss": 0.5134, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 1.94258872651357, | |
| "grad_norm": 0.704097330570221, | |
| "learning_rate": 4.0495193099019524e-06, | |
| "loss": 0.4926, | |
| "step": 621 | |
| }, | |
| { | |
| "epoch": 1.9457202505219207, | |
| "grad_norm": 0.8508503437042236, | |
| "learning_rate": 4.046119280435212e-06, | |
| "loss": 0.5008, | |
| "step": 622 | |
| }, | |
| { | |
| "epoch": 1.9488517745302714, | |
| "grad_norm": 0.725933313369751, | |
| "learning_rate": 4.042714613637775e-06, | |
| "loss": 0.5549, | |
| "step": 623 | |
| }, | |
| { | |
| "epoch": 1.951983298538622, | |
| "grad_norm": 0.8919175863265991, | |
| "learning_rate": 4.039305319721381e-06, | |
| "loss": 0.5183, | |
| "step": 624 | |
| }, | |
| { | |
| "epoch": 1.9551148225469728, | |
| "grad_norm": 0.827919065952301, | |
| "learning_rate": 4.035891408911644e-06, | |
| "loss": 0.5624, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 1.9582463465553235, | |
| "grad_norm": 0.7415187358856201, | |
| "learning_rate": 4.032472891448032e-06, | |
| "loss": 0.5454, | |
| "step": 626 | |
| }, | |
| { | |
| "epoch": 1.9613778705636742, | |
| "grad_norm": 0.7675788998603821, | |
| "learning_rate": 4.029049777583824e-06, | |
| "loss": 0.5361, | |
| "step": 627 | |
| }, | |
| { | |
| "epoch": 1.964509394572025, | |
| "grad_norm": 0.8464030623435974, | |
| "learning_rate": 4.025622077586088e-06, | |
| "loss": 0.5295, | |
| "step": 628 | |
| }, | |
| { | |
| "epoch": 1.9676409185803758, | |
| "grad_norm": 0.7641633749008179, | |
| "learning_rate": 4.022189801735646e-06, | |
| "loss": 0.55, | |
| "step": 629 | |
| }, | |
| { | |
| "epoch": 1.9707724425887265, | |
| "grad_norm": 0.7813227772712708, | |
| "learning_rate": 4.018752960327048e-06, | |
| "loss": 0.5587, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 1.9739039665970772, | |
| "grad_norm": 0.7576701641082764, | |
| "learning_rate": 4.015311563668533e-06, | |
| "loss": 0.5413, | |
| "step": 631 | |
| }, | |
| { | |
| "epoch": 1.977035490605428, | |
| "grad_norm": 0.6949650049209595, | |
| "learning_rate": 4.011865622082004e-06, | |
| "loss": 0.5344, | |
| "step": 632 | |
| }, | |
| { | |
| "epoch": 1.9801670146137789, | |
| "grad_norm": 0.9009145498275757, | |
| "learning_rate": 4.008415145902997e-06, | |
| "loss": 0.5233, | |
| "step": 633 | |
| }, | |
| { | |
| "epoch": 1.9832985386221296, | |
| "grad_norm": 0.7635822892189026, | |
| "learning_rate": 4.004960145480651e-06, | |
| "loss": 0.4981, | |
| "step": 634 | |
| }, | |
| { | |
| "epoch": 1.9864300626304803, | |
| "grad_norm": 0.8916334509849548, | |
| "learning_rate": 4.0015006311776685e-06, | |
| "loss": 0.5311, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 1.989561586638831, | |
| "grad_norm": 0.7197673320770264, | |
| "learning_rate": 3.998036613370295e-06, | |
| "loss": 0.5361, | |
| "step": 636 | |
| }, | |
| { | |
| "epoch": 1.9926931106471817, | |
| "grad_norm": 0.8391228914260864, | |
| "learning_rate": 3.994568102448284e-06, | |
| "loss": 0.5473, | |
| "step": 637 | |
| }, | |
| { | |
| "epoch": 1.9958246346555324, | |
| "grad_norm": 0.9371750950813293, | |
| "learning_rate": 3.991095108814862e-06, | |
| "loss": 0.5303, | |
| "step": 638 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 1914, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 6, | |
| "save_steps": 319, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.9343570238741414e+19, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |