| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.0, | |
| "eval_steps": 100, | |
| "global_step": 3897, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0, | |
| "grad_norm": 45.099327087402344, | |
| "learning_rate": 0.00019948678470618425, | |
| "loss": 2.0895, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 34.96247100830078, | |
| "learning_rate": 0.0001989735694123685, | |
| "loss": 2.3707, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 16.490638732910156, | |
| "learning_rate": 0.00019846035411855275, | |
| "loss": 1.3701, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 0.11159859597682953, | |
| "learning_rate": 0.000197947138824737, | |
| "loss": 1.0582, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 0.12606161832809448, | |
| "learning_rate": 0.00019743392353092123, | |
| "loss": 1.1387, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 4.384986400604248, | |
| "learning_rate": 0.00019692070823710547, | |
| "loss": 2.2113, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 0.7131040692329407, | |
| "learning_rate": 0.00019640749294328974, | |
| "loss": 0.6345, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 5.308964252471924, | |
| "learning_rate": 0.00019589427764947397, | |
| "loss": 1.0728, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 0.04876876622438431, | |
| "learning_rate": 0.0001953810623556582, | |
| "loss": 1.6609, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 8.89411735534668, | |
| "learning_rate": 0.00019486784706184245, | |
| "loss": 1.8662, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 0.05204994976520538, | |
| "learning_rate": 0.00019435463176802672, | |
| "loss": 0.5977, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 34.89448165893555, | |
| "learning_rate": 0.00019384141647421096, | |
| "loss": 2.2815, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 15.786956787109375, | |
| "learning_rate": 0.0001933282011803952, | |
| "loss": 1.6381, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 29.863065719604492, | |
| "learning_rate": 0.00019281498588657943, | |
| "loss": 0.7649, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 16.698570251464844, | |
| "learning_rate": 0.00019230177059276367, | |
| "loss": 1.1147, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 34.869384765625, | |
| "learning_rate": 0.00019178855529894794, | |
| "loss": 1.4108, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 0.3193562924861908, | |
| "learning_rate": 0.00019127534000513218, | |
| "loss": 1.0166, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 23.633769989013672, | |
| "learning_rate": 0.00019076212471131642, | |
| "loss": 1.0537, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 0.03772430121898651, | |
| "learning_rate": 0.00019024890941750065, | |
| "loss": 0.6254, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 2.246229648590088, | |
| "learning_rate": 0.0001897356941236849, | |
| "loss": 2.0714, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 2.4526054859161377, | |
| "learning_rate": 0.00018922247882986913, | |
| "loss": 1.0681, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 1.3229526281356812, | |
| "learning_rate": 0.00018870926353605337, | |
| "loss": 0.6116, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 0.042341507971286774, | |
| "learning_rate": 0.0001881960482422376, | |
| "loss": 1.4669, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 35.0599479675293, | |
| "learning_rate": 0.00018768283294842188, | |
| "loss": 0.6897, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 0.21465550363063812, | |
| "learning_rate": 0.00018716961765460611, | |
| "loss": 1.8696, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 34.034175872802734, | |
| "learning_rate": 0.00018665640236079035, | |
| "loss": 1.5576, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 34.895172119140625, | |
| "learning_rate": 0.0001861431870669746, | |
| "loss": 1.6542, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 0.30317962169647217, | |
| "learning_rate": 0.00018562997177315883, | |
| "loss": 0.9893, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 31.76571273803711, | |
| "learning_rate": 0.0001851167564793431, | |
| "loss": 2.43, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 3.9265520572662354, | |
| "learning_rate": 0.00018460354118552733, | |
| "loss": 0.6246, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 2.8710033893585205, | |
| "learning_rate": 0.00018409032589171157, | |
| "loss": 1.2509, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 28.637104034423828, | |
| "learning_rate": 0.0001835771105978958, | |
| "loss": 1.4902, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 4.821115970611572, | |
| "learning_rate": 0.00018306389530408008, | |
| "loss": 0.7693, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 0.466738224029541, | |
| "learning_rate": 0.00018255068001026432, | |
| "loss": 1.9563, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 34.78805160522461, | |
| "learning_rate": 0.00018203746471644856, | |
| "loss": 1.4862, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 5.344199180603027, | |
| "learning_rate": 0.0001815242494226328, | |
| "loss": 1.1602, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 0.4355538785457611, | |
| "learning_rate": 0.00018101103412881703, | |
| "loss": 1.2546, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 0.7162923216819763, | |
| "learning_rate": 0.0001804978188350013, | |
| "loss": 3.0478, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 0.09367480874061584, | |
| "learning_rate": 0.00017998460354118554, | |
| "loss": 1.0921, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 0.20555052161216736, | |
| "learning_rate": 0.00017947138824736978, | |
| "loss": 0.0523, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 0.08092090487480164, | |
| "learning_rate": 0.00017895817295355402, | |
| "loss": 1.5604, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 0.32685595750808716, | |
| "learning_rate": 0.00017844495765973828, | |
| "loss": 0.5272, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 0.5422254800796509, | |
| "learning_rate": 0.00017793174236592252, | |
| "loss": 3.4396, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 35.279212951660156, | |
| "learning_rate": 0.00017741852707210676, | |
| "loss": 0.8554, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 26.898597717285156, | |
| "learning_rate": 0.000176905311778291, | |
| "loss": 0.6776, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 0.39926862716674805, | |
| "learning_rate": 0.00017639209648447526, | |
| "loss": 2.0551, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 0.39992421865463257, | |
| "learning_rate": 0.0001758788811906595, | |
| "loss": 0.6316, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 0.15786024928092957, | |
| "learning_rate": 0.00017536566589684374, | |
| "loss": 2.5885, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 35.12531280517578, | |
| "learning_rate": 0.00017485245060302798, | |
| "loss": 1.5028, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 1.712053656578064, | |
| "learning_rate": 0.00017433923530921222, | |
| "loss": 0.7324, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 28.602834701538086, | |
| "learning_rate": 0.00017382602001539648, | |
| "loss": 2.7516, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 8.03342342376709, | |
| "learning_rate": 0.00017331280472158072, | |
| "loss": 0.3827, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 33.861934661865234, | |
| "learning_rate": 0.00017279958942776496, | |
| "loss": 1.1201, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 0.0627857968211174, | |
| "learning_rate": 0.0001722863741339492, | |
| "loss": 0.6096, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 34.3585319519043, | |
| "learning_rate": 0.00017177315884013344, | |
| "loss": 1.4903, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 33.19193649291992, | |
| "learning_rate": 0.00017125994354631768, | |
| "loss": 1.4301, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 34.12609100341797, | |
| "learning_rate": 0.00017074672825250194, | |
| "loss": 1.1228, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 0.07128182798624039, | |
| "learning_rate": 0.00017023351295868618, | |
| "loss": 0.8195, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 34.57550048828125, | |
| "learning_rate": 0.00016972029766487042, | |
| "loss": 1.9291, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 60.265777587890625, | |
| "learning_rate": 0.00016920708237105466, | |
| "loss": 2.2235, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 33.86371612548828, | |
| "learning_rate": 0.0001686938670772389, | |
| "loss": 1.101, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 1.018540620803833, | |
| "learning_rate": 0.00016818065178342314, | |
| "loss": 1.0946, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 0.05546221882104874, | |
| "learning_rate": 0.00016766743648960738, | |
| "loss": 1.3699, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 0.4712983965873718, | |
| "learning_rate": 0.00016715422119579164, | |
| "loss": 2.366, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 0.1462753266096115, | |
| "learning_rate": 0.00016664100590197588, | |
| "loss": 1.1087, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 0.0955154225230217, | |
| "learning_rate": 0.00016612779060816012, | |
| "loss": 0.0171, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 35.19109344482422, | |
| "learning_rate": 0.00016561457531434436, | |
| "loss": 1.6907, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 0.16912657022476196, | |
| "learning_rate": 0.00016510136002052862, | |
| "loss": 0.0055, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 34.529685974121094, | |
| "learning_rate": 0.00016458814472671286, | |
| "loss": 2.8261, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 13.726387023925781, | |
| "learning_rate": 0.0001640749294328971, | |
| "loss": 1.7874, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 29.94759750366211, | |
| "learning_rate": 0.00016356171413908134, | |
| "loss": 1.3416, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 0.5144930481910706, | |
| "learning_rate": 0.0001630484988452656, | |
| "loss": 0.7381, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 0.4536816477775574, | |
| "learning_rate": 0.00016253528355144985, | |
| "loss": 1.7744, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 34.2490119934082, | |
| "learning_rate": 0.00016202206825763408, | |
| "loss": 1.6362, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 0.7601956725120544, | |
| "learning_rate": 0.00016150885296381832, | |
| "loss": 1.0899, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 0.16101330518722534, | |
| "learning_rate": 0.00016099563767000256, | |
| "loss": 1.4228, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 2.7051339149475098, | |
| "learning_rate": 0.00016048242237618683, | |
| "loss": 1.2515, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 34.69645690917969, | |
| "learning_rate": 0.00015996920708237107, | |
| "loss": 1.1568, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 0.5005641579627991, | |
| "learning_rate": 0.0001594559917885553, | |
| "loss": 1.8697, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 0.9808595776557922, | |
| "learning_rate": 0.00015894277649473954, | |
| "loss": 1.867, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 34.12071228027344, | |
| "learning_rate": 0.0001584295612009238, | |
| "loss": 1.2028, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 1.4215242862701416, | |
| "learning_rate": 0.00015791634590710805, | |
| "loss": 1.5028, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 32.7039794921875, | |
| "learning_rate": 0.0001574031306132923, | |
| "loss": 1.851, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 11.424134254455566, | |
| "learning_rate": 0.00015688991531947653, | |
| "loss": 1.1995, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 18.550819396972656, | |
| "learning_rate": 0.0001563767000256608, | |
| "loss": 0.5795, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 25.513269424438477, | |
| "learning_rate": 0.00015586348473184503, | |
| "loss": 0.6537, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 0.043114595115184784, | |
| "learning_rate": 0.00015535026943802927, | |
| "loss": 1.4115, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 0.06105157732963562, | |
| "learning_rate": 0.0001548370541442135, | |
| "loss": 0.5029, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 6.027977466583252, | |
| "learning_rate": 0.00015432383885039775, | |
| "loss": 1.1377, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 9.911833763122559, | |
| "learning_rate": 0.000153810623556582, | |
| "loss": 0.6785, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 0.08647739142179489, | |
| "learning_rate": 0.00015329740826276625, | |
| "loss": 0.3428, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 0.06115918606519699, | |
| "learning_rate": 0.0001527841929689505, | |
| "loss": 0.5422, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 33.05713653564453, | |
| "learning_rate": 0.00015227097767513473, | |
| "loss": 0.9532, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 2.755101442337036, | |
| "learning_rate": 0.00015175776238131897, | |
| "loss": 0.8933, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 1.09266197681427, | |
| "learning_rate": 0.0001512445470875032, | |
| "loss": 1.3081, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 1.8350268602371216, | |
| "learning_rate": 0.00015073133179368745, | |
| "loss": 1.6349, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 28.465572357177734, | |
| "learning_rate": 0.00015021811649987168, | |
| "loss": 1.4432, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 24.18708038330078, | |
| "learning_rate": 0.00014970490120605595, | |
| "loss": 0.8608, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 0.06248585879802704, | |
| "learning_rate": 0.0001491916859122402, | |
| "loss": 0.7754, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 0.02755439281463623, | |
| "learning_rate": 0.00014867847061842443, | |
| "loss": 0.3644, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 0.03344785049557686, | |
| "learning_rate": 0.00014816525532460867, | |
| "loss": 2.7486, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 4.339983940124512, | |
| "learning_rate": 0.0001476520400307929, | |
| "loss": 0.8171, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 4.660573959350586, | |
| "learning_rate": 0.00014713882473697717, | |
| "loss": 2.1821, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 21.810422897338867, | |
| "learning_rate": 0.0001466256094431614, | |
| "loss": 1.0611, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 33.206607818603516, | |
| "learning_rate": 0.00014611239414934565, | |
| "loss": 1.5671, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 0.678447425365448, | |
| "learning_rate": 0.0001455991788555299, | |
| "loss": 1.0065, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 32.59206771850586, | |
| "learning_rate": 0.00014508596356171415, | |
| "loss": 1.8422, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 28.002819061279297, | |
| "learning_rate": 0.0001445727482678984, | |
| "loss": 1.1739, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 4.193221092224121, | |
| "learning_rate": 0.00014405953297408263, | |
| "loss": 0.431, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 34.987388610839844, | |
| "learning_rate": 0.00014354631768026687, | |
| "loss": 1.1974, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 0.1439794898033142, | |
| "learning_rate": 0.0001430331023864511, | |
| "loss": 2.1824, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 0.07825344055891037, | |
| "learning_rate": 0.00014251988709263537, | |
| "loss": 1.103, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 34.14964294433594, | |
| "learning_rate": 0.0001420066717988196, | |
| "loss": 2.1521, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 1.839199185371399, | |
| "learning_rate": 0.00014149345650500385, | |
| "loss": 0.0536, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 34.573585510253906, | |
| "learning_rate": 0.0001409802412111881, | |
| "loss": 1.5661, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 0.41036948561668396, | |
| "learning_rate": 0.00014046702591737236, | |
| "loss": 1.1817, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 0.08495642244815826, | |
| "learning_rate": 0.0001399538106235566, | |
| "loss": 1.5704, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 3.0547990798950195, | |
| "learning_rate": 0.00013944059532974083, | |
| "loss": 1.5402, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 7.8949713706970215, | |
| "learning_rate": 0.00013892738003592507, | |
| "loss": 0.4595, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 24.375001907348633, | |
| "learning_rate": 0.00013841416474210934, | |
| "loss": 0.7511, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 12.163422584533691, | |
| "learning_rate": 0.00013790094944829358, | |
| "loss": 0.5726, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 34.362632751464844, | |
| "learning_rate": 0.00013738773415447782, | |
| "loss": 1.3018, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 21.605195999145508, | |
| "learning_rate": 0.00013687451886066205, | |
| "loss": 1.2047, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 4.356873035430908, | |
| "learning_rate": 0.0001363613035668463, | |
| "loss": 1.3787, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 12.701129913330078, | |
| "learning_rate": 0.00013584808827303056, | |
| "loss": 1.0334, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 7.2957892417907715, | |
| "learning_rate": 0.0001353348729792148, | |
| "loss": 0.2939, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 0.25156155228614807, | |
| "learning_rate": 0.00013482165768539904, | |
| "loss": 0.7811, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 0.8677279353141785, | |
| "learning_rate": 0.00013430844239158328, | |
| "loss": 1.5475, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 0.13093720376491547, | |
| "learning_rate": 0.00013379522709776754, | |
| "loss": 0.7902, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 0.2733534276485443, | |
| "learning_rate": 0.00013328201180395178, | |
| "loss": 0.9059, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 34.48268508911133, | |
| "learning_rate": 0.00013276879651013602, | |
| "loss": 1.5932, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 33.39665603637695, | |
| "learning_rate": 0.00013225558121632026, | |
| "loss": 2.4436, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 0.6858634948730469, | |
| "learning_rate": 0.0001317423659225045, | |
| "loss": 0.8113, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 32.40642166137695, | |
| "learning_rate": 0.00013122915062868873, | |
| "loss": 2.1392, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 0.05664900690317154, | |
| "learning_rate": 0.00013071593533487297, | |
| "loss": 0.8724, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 30.9477481842041, | |
| "learning_rate": 0.0001302027200410572, | |
| "loss": 1.2746, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 2.9625468254089355, | |
| "learning_rate": 0.00012968950474724145, | |
| "loss": 2.0471, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 6.3701629638671875, | |
| "learning_rate": 0.00012917628945342572, | |
| "loss": 0.9569, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 5.0901007652282715, | |
| "learning_rate": 0.00012866307415960996, | |
| "loss": 0.8391, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 2.827472448348999, | |
| "learning_rate": 0.0001281498588657942, | |
| "loss": 0.5071, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 0.24257418513298035, | |
| "learning_rate": 0.00012763664357197843, | |
| "loss": 0.0258, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 34.39240264892578, | |
| "learning_rate": 0.0001271234282781627, | |
| "loss": 1.6826, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 0.06999631226062775, | |
| "learning_rate": 0.00012661021298434694, | |
| "loss": 2.854, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 0.7731397151947021, | |
| "learning_rate": 0.00012609699769053118, | |
| "loss": 1.5991, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 0.7090578675270081, | |
| "learning_rate": 0.00012558378239671542, | |
| "loss": 1.1829, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 33.13046646118164, | |
| "learning_rate": 0.00012507056710289968, | |
| "loss": 1.4736, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 0.6540558338165283, | |
| "learning_rate": 0.00012455735180908392, | |
| "loss": 1.6983, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 1.6605318784713745, | |
| "learning_rate": 0.00012404413651526816, | |
| "loss": 1.5005, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 31.07722282409668, | |
| "learning_rate": 0.0001235309212214524, | |
| "loss": 1.4853, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 1.135640025138855, | |
| "learning_rate": 0.00012301770592763664, | |
| "loss": 0.9949, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 0.14291101694107056, | |
| "learning_rate": 0.0001225044906338209, | |
| "loss": 1.7112, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 0.07842687517404556, | |
| "learning_rate": 0.00012199127534000514, | |
| "loss": 0.677, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 0.6385894417762756, | |
| "learning_rate": 0.00012147806004618938, | |
| "loss": 1.0463, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 34.2146110534668, | |
| "learning_rate": 0.00012096484475237362, | |
| "loss": 2.0908, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 0.15750457346439362, | |
| "learning_rate": 0.00012045162945855788, | |
| "loss": 1.6537, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 0.15273036062717438, | |
| "learning_rate": 0.00011993841416474212, | |
| "loss": 1.8717, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 4.405086517333984, | |
| "learning_rate": 0.00011942519887092636, | |
| "loss": 0.3942, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 2.592953681945801, | |
| "learning_rate": 0.0001189119835771106, | |
| "loss": 0.3217, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 0.5546024441719055, | |
| "learning_rate": 0.00011839876828329485, | |
| "loss": 1.0285, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 0.09479659050703049, | |
| "learning_rate": 0.00011788555298947909, | |
| "loss": 1.7519, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 0.09771596640348434, | |
| "learning_rate": 0.00011737233769566333, | |
| "loss": 1.2341, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 33.37510681152344, | |
| "learning_rate": 0.00011685912240184758, | |
| "loss": 3.3626, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 32.650489807128906, | |
| "learning_rate": 0.00011634590710803182, | |
| "loss": 0.9869, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 32.52039337158203, | |
| "learning_rate": 0.00011583269181421607, | |
| "loss": 1.8117, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 15.11254596710205, | |
| "learning_rate": 0.00011531947652040031, | |
| "loss": 1.288, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 3.6993765830993652, | |
| "learning_rate": 0.00011480626122658455, | |
| "loss": 0.863, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 34.809791564941406, | |
| "learning_rate": 0.00011429304593276879, | |
| "loss": 0.6192, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 0.10077141970396042, | |
| "learning_rate": 0.00011377983063895306, | |
| "loss": 0.571, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 0.07107020914554596, | |
| "learning_rate": 0.0001132666153451373, | |
| "loss": 0.996, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 33.04214096069336, | |
| "learning_rate": 0.00011275340005132153, | |
| "loss": 2.6481, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 32.640541076660156, | |
| "learning_rate": 0.00011224018475750577, | |
| "loss": 2.1778, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 1.7503925561904907, | |
| "learning_rate": 0.00011172696946369004, | |
| "loss": 1.5489, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 0.348849892616272, | |
| "learning_rate": 0.00011121375416987428, | |
| "loss": 1.1267, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 0.8353062272071838, | |
| "learning_rate": 0.00011070053887605852, | |
| "loss": 0.0233, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 0.4115784466266632, | |
| "learning_rate": 0.00011018732358224275, | |
| "loss": 1.3895, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 0.09716140478849411, | |
| "learning_rate": 0.00010967410828842699, | |
| "loss": 0.9971, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 0.07956521958112717, | |
| "learning_rate": 0.00010916089299461126, | |
| "loss": 0.4633, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 33.34837341308594, | |
| "learning_rate": 0.0001086476777007955, | |
| "loss": 0.5945, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 33.278682708740234, | |
| "learning_rate": 0.00010813446240697974, | |
| "loss": 1.0137, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 32.34352111816406, | |
| "learning_rate": 0.00010762124711316398, | |
| "loss": 2.238, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 0.8423387408256531, | |
| "learning_rate": 0.00010710803181934823, | |
| "loss": 0.0119, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 0.43691009283065796, | |
| "learning_rate": 0.00010659481652553247, | |
| "loss": 0.7785, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 0.27386584877967834, | |
| "learning_rate": 0.0001060816012317167, | |
| "loss": 1.4121, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 0.035622984170913696, | |
| "learning_rate": 0.00010556838593790094, | |
| "loss": 1.8601, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 0.6981528401374817, | |
| "learning_rate": 0.00010505517064408518, | |
| "loss": 1.2077, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 32.1064567565918, | |
| "learning_rate": 0.00010454195535026945, | |
| "loss": 1.5198, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 30.88555908203125, | |
| "learning_rate": 0.00010402874005645369, | |
| "loss": 2.0069, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 0.03494291752576828, | |
| "learning_rate": 0.00010351552476263793, | |
| "loss": 0.3124, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 1.9848475456237793, | |
| "learning_rate": 0.00010300230946882216, | |
| "loss": 0.8303, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 31.325864791870117, | |
| "learning_rate": 0.00010248909417500643, | |
| "loss": 0.9595, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 0.025403378531336784, | |
| "learning_rate": 0.00010197587888119067, | |
| "loss": 0.362, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 1.1182492971420288, | |
| "learning_rate": 0.00010146266358737491, | |
| "loss": 1.4183, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 0.0197270717471838, | |
| "learning_rate": 0.00010094944829355915, | |
| "loss": 0.9907, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 1.5864602327346802, | |
| "learning_rate": 0.00010043623299974341, | |
| "loss": 0.9328, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 0.019860761240124702, | |
| "learning_rate": 9.992301770592765e-05, | |
| "loss": 1.2027, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 0.27393853664398193, | |
| "learning_rate": 9.940980241211189e-05, | |
| "loss": 1.2231, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 0.06520062685012817, | |
| "learning_rate": 9.889658711829613e-05, | |
| "loss": 1.1252, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 0.10603094100952148, | |
| "learning_rate": 9.838337182448038e-05, | |
| "loss": 0.5362, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 0.21443594992160797, | |
| "learning_rate": 9.787015653066462e-05, | |
| "loss": 1.6563, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 0.15189635753631592, | |
| "learning_rate": 9.735694123684886e-05, | |
| "loss": 1.5657, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 32.36518478393555, | |
| "learning_rate": 9.68437259430331e-05, | |
| "loss": 2.1702, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 0.2264794558286667, | |
| "learning_rate": 9.633051064921735e-05, | |
| "loss": 1.2663, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 30.200082778930664, | |
| "learning_rate": 9.581729535540159e-05, | |
| "loss": 2.2151, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 3.5600388050079346, | |
| "learning_rate": 9.530408006158584e-05, | |
| "loss": 0.7643, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 0.1124127209186554, | |
| "learning_rate": 9.479086476777008e-05, | |
| "loss": 0.679, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 28.294071197509766, | |
| "learning_rate": 9.427764947395433e-05, | |
| "loss": 1.5172, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 26.79751205444336, | |
| "learning_rate": 9.376443418013857e-05, | |
| "loss": 0.4192, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 0.1510336697101593, | |
| "learning_rate": 9.325121888632282e-05, | |
| "loss": 2.1576, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 0.32501623034477234, | |
| "learning_rate": 9.273800359250706e-05, | |
| "loss": 1.1192, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 26.4614315032959, | |
| "learning_rate": 9.22247882986913e-05, | |
| "loss": 0.7766, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 0.20047006011009216, | |
| "learning_rate": 9.171157300487555e-05, | |
| "loss": 0.6836, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 10.277036666870117, | |
| "learning_rate": 9.119835771105979e-05, | |
| "loss": 0.4951, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 9.455358505249023, | |
| "learning_rate": 9.068514241724404e-05, | |
| "loss": 0.2494, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 5.371951103210449, | |
| "learning_rate": 9.017192712342828e-05, | |
| "loss": 0.7672, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 4.935911655426025, | |
| "learning_rate": 8.965871182961254e-05, | |
| "loss": 0.7815, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 0.05824290215969086, | |
| "learning_rate": 8.914549653579677e-05, | |
| "loss": 0.4563, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 29.326370239257812, | |
| "learning_rate": 8.863228124198101e-05, | |
| "loss": 0.9615, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 30.599393844604492, | |
| "learning_rate": 8.811906594816526e-05, | |
| "loss": 0.617, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 0.6331340074539185, | |
| "learning_rate": 8.76058506543495e-05, | |
| "loss": 1.0741, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 34.11085891723633, | |
| "learning_rate": 8.709263536053374e-05, | |
| "loss": 1.9807, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 0.4941723942756653, | |
| "learning_rate": 8.6579420066718e-05, | |
| "loss": 1.6712, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 0.8858903646469116, | |
| "learning_rate": 8.606620477290223e-05, | |
| "loss": 1.1382, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 0.4389359951019287, | |
| "learning_rate": 8.555298947908647e-05, | |
| "loss": 0.0166, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 0.27877870202064514, | |
| "learning_rate": 8.503977418527072e-05, | |
| "loss": 1.8697, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 0.0968392938375473, | |
| "learning_rate": 8.452655889145496e-05, | |
| "loss": 0.9307, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 0.07415905594825745, | |
| "learning_rate": 8.401334359763922e-05, | |
| "loss": 1.3474, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 32.0526008605957, | |
| "learning_rate": 8.350012830382345e-05, | |
| "loss": 1.9036, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 31.857463836669922, | |
| "learning_rate": 8.29869130100077e-05, | |
| "loss": 1.7843, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 0.16152064502239227, | |
| "learning_rate": 8.247369771619195e-05, | |
| "loss": 1.138, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 31.4908504486084, | |
| "learning_rate": 8.19604824223762e-05, | |
| "loss": 1.348, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 0.8940775990486145, | |
| "learning_rate": 8.144726712856044e-05, | |
| "loss": 1.4317, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 1.3364328145980835, | |
| "learning_rate": 8.093405183474469e-05, | |
| "loss": 1.3774, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 32.3911018371582, | |
| "learning_rate": 8.042083654092893e-05, | |
| "loss": 1.8806, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 0.0908791646361351, | |
| "learning_rate": 7.990762124711317e-05, | |
| "loss": 0.3828, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 0.7057339549064636, | |
| "learning_rate": 7.939440595329742e-05, | |
| "loss": 1.8282, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 0.9124455451965332, | |
| "learning_rate": 7.888119065948166e-05, | |
| "loss": 0.8194, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 0.6995695233345032, | |
| "learning_rate": 7.83679753656659e-05, | |
| "loss": 2.4169, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 30.961750030517578, | |
| "learning_rate": 7.785476007185013e-05, | |
| "loss": 1.0782, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 30.930931091308594, | |
| "learning_rate": 7.734154477803439e-05, | |
| "loss": 1.773, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 2.427703857421875, | |
| "learning_rate": 7.682832948421863e-05, | |
| "loss": 1.1838, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 28.99689292907715, | |
| "learning_rate": 7.631511419040288e-05, | |
| "loss": 0.9657, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 0.0632254034280777, | |
| "learning_rate": 7.580189889658712e-05, | |
| "loss": 0.9273, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 0.06876543164253235, | |
| "learning_rate": 7.528868360277137e-05, | |
| "loss": 0.6686, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 0.059902504086494446, | |
| "learning_rate": 7.477546830895561e-05, | |
| "loss": 0.6242, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 30.04041290283203, | |
| "learning_rate": 7.426225301513986e-05, | |
| "loss": 2.3915, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 2.442535400390625, | |
| "learning_rate": 7.37490377213241e-05, | |
| "loss": 0.5622, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 0.0483098030090332, | |
| "learning_rate": 7.323582242750834e-05, | |
| "loss": 1.5706, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 0.048585113137960434, | |
| "learning_rate": 7.272260713369259e-05, | |
| "loss": 0.8918, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 29.45044708251953, | |
| "learning_rate": 7.220939183987683e-05, | |
| "loss": 1.4446, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 0.03927851840853691, | |
| "learning_rate": 7.169617654606108e-05, | |
| "loss": 0.6336, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 5.8573174476623535, | |
| "learning_rate": 7.118296125224532e-05, | |
| "loss": 0.5349, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 0.04964911565184593, | |
| "learning_rate": 7.066974595842957e-05, | |
| "loss": 1.471, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 1.5833439826965332, | |
| "learning_rate": 7.015653066461381e-05, | |
| "loss": 0.3127, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 0.6209262609481812, | |
| "learning_rate": 6.964331537079806e-05, | |
| "loss": 0.3781, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 34.23354721069336, | |
| "learning_rate": 6.91301000769823e-05, | |
| "loss": 1.8894, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 0.40863272547721863, | |
| "learning_rate": 6.861688478316654e-05, | |
| "loss": 0.8985, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 0.26213815808296204, | |
| "learning_rate": 6.810366948935078e-05, | |
| "loss": 1.038, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 32.263370513916016, | |
| "learning_rate": 6.759045419553503e-05, | |
| "loss": 2.2679, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 34.00870895385742, | |
| "learning_rate": 6.707723890171927e-05, | |
| "loss": 2.075, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 0.4284890294075012, | |
| "learning_rate": 6.656402360790351e-05, | |
| "loss": 0.7597, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 0.3407612144947052, | |
| "learning_rate": 6.605080831408776e-05, | |
| "loss": 0.4383, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 0.467129111289978, | |
| "learning_rate": 6.5537593020272e-05, | |
| "loss": 1.2687, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 0.13389898836612701, | |
| "learning_rate": 6.502437772645625e-05, | |
| "loss": 2.1633, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 0.13675318658351898, | |
| "learning_rate": 6.451116243264049e-05, | |
| "loss": 0.7836, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 0.5350820422172546, | |
| "learning_rate": 6.399794713882474e-05, | |
| "loss": 0.7649, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 0.48448631167411804, | |
| "learning_rate": 6.348473184500898e-05, | |
| "loss": 1.38, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 0.5778862237930298, | |
| "learning_rate": 6.297151655119323e-05, | |
| "loss": 2.2181, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 1.1682463884353638, | |
| "learning_rate": 6.245830125737747e-05, | |
| "loss": 1.0876, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 31.393110275268555, | |
| "learning_rate": 6.194508596356173e-05, | |
| "loss": 1.1128, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 0.6889828443527222, | |
| "learning_rate": 6.143187066974596e-05, | |
| "loss": 0.7252, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 0.5311009883880615, | |
| "learning_rate": 6.0918655375930204e-05, | |
| "loss": 0.3959, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 0.43320441246032715, | |
| "learning_rate": 6.040544008211445e-05, | |
| "loss": 1.2466, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 0.06079118326306343, | |
| "learning_rate": 5.989222478829869e-05, | |
| "loss": 2.0125, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 0.042944520711898804, | |
| "learning_rate": 5.937900949448294e-05, | |
| "loss": 0.779, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 0.9245113730430603, | |
| "learning_rate": 5.886579420066718e-05, | |
| "loss": 0.7976, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 0.039791032671928406, | |
| "learning_rate": 5.835257890685143e-05, | |
| "loss": 1.4881, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 0.6234956979751587, | |
| "learning_rate": 5.783936361303567e-05, | |
| "loss": 0.6912, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 0.9020193219184875, | |
| "learning_rate": 5.7326148319219916e-05, | |
| "loss": 0.7522, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 0.034982066601514816, | |
| "learning_rate": 5.681293302540416e-05, | |
| "loss": 1.2431, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 0.04073556512594223, | |
| "learning_rate": 5.6299717731588407e-05, | |
| "loss": 1.1053, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 35.25714874267578, | |
| "learning_rate": 5.5786502437772645e-05, | |
| "loss": 1.4825, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 0.7252044081687927, | |
| "learning_rate": 5.52732871439569e-05, | |
| "loss": 2.1285, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 31.510704040527344, | |
| "learning_rate": 5.4760071850141136e-05, | |
| "loss": 0.7914, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 0.44585466384887695, | |
| "learning_rate": 5.4246856556325375e-05, | |
| "loss": 1.2094, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 0.2099919468164444, | |
| "learning_rate": 5.373364126250963e-05, | |
| "loss": 2.496, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 30.501766204833984, | |
| "learning_rate": 5.3220425968693866e-05, | |
| "loss": 1.0997, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 0.36132165789604187, | |
| "learning_rate": 5.270721067487812e-05, | |
| "loss": 1.4378, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 1.6164754629135132, | |
| "learning_rate": 5.219399538106236e-05, | |
| "loss": 1.1191, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 0.9825451374053955, | |
| "learning_rate": 5.16807800872466e-05, | |
| "loss": 0.861, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 0.28137707710266113, | |
| "learning_rate": 5.116756479343084e-05, | |
| "loss": 0.4387, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 0.2143826186656952, | |
| "learning_rate": 5.0654349499615094e-05, | |
| "loss": 0.9561, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 0.25483042001724243, | |
| "learning_rate": 5.014113420579933e-05, | |
| "loss": 1.4271, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 0.36998990178108215, | |
| "learning_rate": 4.962791891198358e-05, | |
| "loss": 0.8621, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 0.08702105283737183, | |
| "learning_rate": 4.9114703618167824e-05, | |
| "loss": 0.4751, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 0.455074667930603, | |
| "learning_rate": 4.860148832435207e-05, | |
| "loss": 2.7804, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 0.41690900921821594, | |
| "learning_rate": 4.8088273030536315e-05, | |
| "loss": 0.444, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 0.0824534222483635, | |
| "learning_rate": 4.757505773672056e-05, | |
| "loss": 2.2328, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 0.7097424268722534, | |
| "learning_rate": 4.70618424429048e-05, | |
| "loss": 1.232, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 32.93928527832031, | |
| "learning_rate": 4.6548627149089045e-05, | |
| "loss": 2.1723, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 1.5564465522766113, | |
| "learning_rate": 4.603541185527329e-05, | |
| "loss": 1.3908, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 1.1192642450332642, | |
| "learning_rate": 4.5522196561457536e-05, | |
| "loss": 0.2986, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 0.09087738394737244, | |
| "learning_rate": 4.5008981267641774e-05, | |
| "loss": 1.3239, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 31.66700553894043, | |
| "learning_rate": 4.449576597382602e-05, | |
| "loss": 1.9121, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 0.08067402243614197, | |
| "learning_rate": 4.3982550680010266e-05, | |
| "loss": 1.4794, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 31.138818740844727, | |
| "learning_rate": 4.346933538619451e-05, | |
| "loss": 1.3591, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 30.739635467529297, | |
| "learning_rate": 4.2956120092378757e-05, | |
| "loss": 2.1444, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 29.64994239807129, | |
| "learning_rate": 4.2442904798563e-05, | |
| "loss": 1.1349, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 0.06835217773914337, | |
| "learning_rate": 4.192968950474724e-05, | |
| "loss": 1.1995, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 0.07994609326124191, | |
| "learning_rate": 4.1416474210931486e-05, | |
| "loss": 1.106, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 0.077377088367939, | |
| "learning_rate": 4.090325891711573e-05, | |
| "loss": 0.5515, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 0.07298921793699265, | |
| "learning_rate": 4.039004362329998e-05, | |
| "loss": 1.3587, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 2.3832900524139404, | |
| "learning_rate": 3.987682832948422e-05, | |
| "loss": 0.5702, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 1.3254412412643433, | |
| "learning_rate": 3.936361303566847e-05, | |
| "loss": 0.6837, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 31.57373046875, | |
| "learning_rate": 3.885039774185271e-05, | |
| "loss": 1.4529, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 31.78739356994629, | |
| "learning_rate": 3.833718244803695e-05, | |
| "loss": 1.1592, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 0.885814368724823, | |
| "learning_rate": 3.78239671542212e-05, | |
| "loss": 2.0692, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 0.11006342619657516, | |
| "learning_rate": 3.7310751860405444e-05, | |
| "loss": 1.3088, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 0.7606900930404663, | |
| "learning_rate": 3.679753656658968e-05, | |
| "loss": 0.7503, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 34.01801681518555, | |
| "learning_rate": 3.628432127277393e-05, | |
| "loss": 2.3927, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 31.359045028686523, | |
| "learning_rate": 3.5771105978958174e-05, | |
| "loss": 1.7152, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 0.4896736741065979, | |
| "learning_rate": 3.525789068514242e-05, | |
| "loss": 1.1745, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 33.61710739135742, | |
| "learning_rate": 3.4744675391326665e-05, | |
| "loss": 2.6372, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 0.2002750039100647, | |
| "learning_rate": 3.423146009751091e-05, | |
| "loss": 0.3958, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 0.5248939394950867, | |
| "learning_rate": 3.3718244803695156e-05, | |
| "loss": 1.1428, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 0.20893482863903046, | |
| "learning_rate": 3.32050295098794e-05, | |
| "loss": 1.1695, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 31.339996337890625, | |
| "learning_rate": 3.269181421606364e-05, | |
| "loss": 1.9207, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 0.20348432660102844, | |
| "learning_rate": 3.2178598922247886e-05, | |
| "loss": 1.587, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 31.108985900878906, | |
| "learning_rate": 3.1665383628432124e-05, | |
| "loss": 1.3677, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 29.706878662109375, | |
| "learning_rate": 3.115216833461637e-05, | |
| "loss": 2.0426, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 1.0215935707092285, | |
| "learning_rate": 3.0638953040800616e-05, | |
| "loss": 1.7041, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 0.23811456561088562, | |
| "learning_rate": 3.012573774698486e-05, | |
| "loss": 1.2077, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 0.21963796019554138, | |
| "learning_rate": 2.9612522453169107e-05, | |
| "loss": 1.1672, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 1.149311900138855, | |
| "learning_rate": 2.9099307159353352e-05, | |
| "loss": 1.6651, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 1.1534388065338135, | |
| "learning_rate": 2.8586091865537594e-05, | |
| "loss": 1.2245, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 1.3270301818847656, | |
| "learning_rate": 2.807287657172184e-05, | |
| "loss": 0.6268, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 0.2713325619697571, | |
| "learning_rate": 2.7559661277906085e-05, | |
| "loss": 1.6962, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 0.25903043150901794, | |
| "learning_rate": 2.704644598409033e-05, | |
| "loss": 0.4256, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 0.456488698720932, | |
| "learning_rate": 2.6533230690274573e-05, | |
| "loss": 0.7452, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 30.572643280029297, | |
| "learning_rate": 2.6020015396458812e-05, | |
| "loss": 1.9218, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 31.96410369873047, | |
| "learning_rate": 2.5506800102643057e-05, | |
| "loss": 2.2211, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 0.5312764644622803, | |
| "learning_rate": 2.4993584808827306e-05, | |
| "loss": 0.6983, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 0.23434561491012573, | |
| "learning_rate": 2.448036951501155e-05, | |
| "loss": 0.7466, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 36.59482192993164, | |
| "learning_rate": 2.3967154221195794e-05, | |
| "loss": 1.6694, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 32.04636001586914, | |
| "learning_rate": 2.3453938927380036e-05, | |
| "loss": 1.6496, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 0.2819555699825287, | |
| "learning_rate": 2.294072363356428e-05, | |
| "loss": 1.3424, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 0.46713873744010925, | |
| "learning_rate": 2.2427508339748527e-05, | |
| "loss": 0.0149, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 0.5181002020835876, | |
| "learning_rate": 2.1914293045932773e-05, | |
| "loss": 0.7906, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 0.20827196538448334, | |
| "learning_rate": 2.140107775211701e-05, | |
| "loss": 0.4505, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 32.157432556152344, | |
| "learning_rate": 2.0887862458301257e-05, | |
| "loss": 2.5997, | |
| "step": 3490 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 0.1746426224708557, | |
| "learning_rate": 2.0374647164485502e-05, | |
| "loss": 1.6267, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 31.728593826293945, | |
| "learning_rate": 1.9861431870669748e-05, | |
| "loss": 0.8132, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 0.8428208231925964, | |
| "learning_rate": 1.9348216576853994e-05, | |
| "loss": 1.4408, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 0.6710448861122131, | |
| "learning_rate": 1.8835001283038236e-05, | |
| "loss": 0.7472, | |
| "step": 3530 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 0.15114788711071014, | |
| "learning_rate": 1.8321785989222478e-05, | |
| "loss": 0.3549, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 0.48726460337638855, | |
| "learning_rate": 1.7808570695406723e-05, | |
| "loss": 1.0424, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 31.90473747253418, | |
| "learning_rate": 1.729535540159097e-05, | |
| "loss": 1.6014, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 0.11974696815013885, | |
| "learning_rate": 1.678214010777521e-05, | |
| "loss": 1.367, | |
| "step": 3570 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 31.7264461517334, | |
| "learning_rate": 1.6268924813959457e-05, | |
| "loss": 0.8339, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 0.9428460001945496, | |
| "learning_rate": 1.5755709520143702e-05, | |
| "loss": 1.2831, | |
| "step": 3590 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 0.5526345372200012, | |
| "learning_rate": 1.5242494226327944e-05, | |
| "loss": 0.7811, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 0.4995884299278259, | |
| "learning_rate": 1.4729278932512188e-05, | |
| "loss": 0.454, | |
| "step": 3610 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 32.1862907409668, | |
| "learning_rate": 1.4216063638696434e-05, | |
| "loss": 1.7057, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 30.95826530456543, | |
| "learning_rate": 1.3702848344880677e-05, | |
| "loss": 1.6207, | |
| "step": 3630 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 0.10005848109722137, | |
| "learning_rate": 1.3189633051064923e-05, | |
| "loss": 0.8695, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 31.5998477935791, | |
| "learning_rate": 1.2676417757249169e-05, | |
| "loss": 1.9645, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 0.797754168510437, | |
| "learning_rate": 1.216320246343341e-05, | |
| "loss": 1.1224, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 0.10280079394578934, | |
| "learning_rate": 1.1649987169617656e-05, | |
| "loss": 0.8117, | |
| "step": 3670 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 31.812061309814453, | |
| "learning_rate": 1.1136771875801898e-05, | |
| "loss": 0.5172, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 0.08260977268218994, | |
| "learning_rate": 1.0623556581986144e-05, | |
| "loss": 1.6816, | |
| "step": 3690 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 0.6431272029876709, | |
| "learning_rate": 1.0110341288170388e-05, | |
| "loss": 2.2201, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 0.5408867597579956, | |
| "learning_rate": 9.597125994354632e-06, | |
| "loss": 0.9963, | |
| "step": 3710 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 0.6773453950881958, | |
| "learning_rate": 9.083910700538877e-06, | |
| "loss": 2.3436, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 33.85165786743164, | |
| "learning_rate": 8.570695406723121e-06, | |
| "loss": 1.229, | |
| "step": 3730 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 0.09623867273330688, | |
| "learning_rate": 8.057480112907365e-06, | |
| "loss": 0.9852, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 0.6665917038917542, | |
| "learning_rate": 7.5442648190916095e-06, | |
| "loss": 0.9811, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 0.1049143448472023, | |
| "learning_rate": 7.031049525275854e-06, | |
| "loss": 1.6639, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 1.028926968574524, | |
| "learning_rate": 6.517834231460097e-06, | |
| "loss": 0.6745, | |
| "step": 3770 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 0.5345853567123413, | |
| "learning_rate": 6.004618937644342e-06, | |
| "loss": 1.9351, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 0.09490037709474564, | |
| "learning_rate": 5.4914036438285865e-06, | |
| "loss": 1.0257, | |
| "step": 3790 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 0.7808061838150024, | |
| "learning_rate": 4.978188350012831e-06, | |
| "loss": 1.6728, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 0.4944613575935364, | |
| "learning_rate": 4.464973056197075e-06, | |
| "loss": 1.6118, | |
| "step": 3810 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 1.0724005699157715, | |
| "learning_rate": 3.951757762381319e-06, | |
| "loss": 1.352, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 0.8290554285049438, | |
| "learning_rate": 3.4385424685655636e-06, | |
| "loss": 2.4513, | |
| "step": 3830 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 0.532361626625061, | |
| "learning_rate": 2.9253271747498074e-06, | |
| "loss": 0.412, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 0.1319778561592102, | |
| "learning_rate": 2.412111880934052e-06, | |
| "loss": 0.7912, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 0.6048825979232788, | |
| "learning_rate": 1.8988965871182964e-06, | |
| "loss": 0.9813, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 0.4283379018306732, | |
| "learning_rate": 1.3856812933025406e-06, | |
| "loss": 0.7374, | |
| "step": 3870 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 0.7439431548118591, | |
| "learning_rate": 8.724659994867848e-07, | |
| "loss": 0.3349, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 0.49157199263572693, | |
| "learning_rate": 3.5925070567102904e-07, | |
| "loss": 1.4183, | |
| "step": 3890 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "step": 3897, | |
| "total_flos": 9.121574551619174e+17, | |
| "train_loss": 1.2556480791153037, | |
| "train_runtime": 633.8251, | |
| "train_samples_per_second": 18.444, | |
| "train_steps_per_second": 6.148 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 3897, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 1000, | |
| "total_flos": 9.121574551619174e+17, | |
| "train_batch_size": 3, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |