| { | |
| "best_global_step": 50000, | |
| "best_metric": 63.090843200683594, | |
| "best_model_checkpoint": "tinybert_base_train_kd/checkpoint-50000", | |
| "epoch": 25.0, | |
| "eval_steps": 10000, | |
| "global_step": 59550, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.20990764063811923, | |
| "grad_norm": 1038.087890625, | |
| "learning_rate": 4.9900000000000005e-06, | |
| "loss": 6028.7805, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.41981528127623846, | |
| "grad_norm": 1000.9192504882812, | |
| "learning_rate": 9.990000000000001e-06, | |
| "loss": 3688.4922, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.6297229219143576, | |
| "grad_norm": 1443.9014892578125, | |
| "learning_rate": 1.499e-05, | |
| "loss": 2203.5858, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.8396305625524769, | |
| "grad_norm": 2327.7587890625, | |
| "learning_rate": 1.999e-05, | |
| "loss": 1477.7162, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 1.0495382031905962, | |
| "grad_norm": 1928.0552978515625, | |
| "learning_rate": 2.4990000000000003e-05, | |
| "loss": 1145.1917, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 1.2594458438287153, | |
| "grad_norm": 1427.1600341796875, | |
| "learning_rate": 2.9990000000000003e-05, | |
| "loss": 941.4575, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 1.4693534844668346, | |
| "grad_norm": 832.8273315429688, | |
| "learning_rate": 3.499e-05, | |
| "loss": 812.8064, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 1.6792611251049538, | |
| "grad_norm": 880.94140625, | |
| "learning_rate": 3.999e-05, | |
| "loss": 722.3781, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 1.8891687657430731, | |
| "grad_norm": 458.8153381347656, | |
| "learning_rate": 4.499e-05, | |
| "loss": 657.1395, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 2.0990764063811924, | |
| "grad_norm": 192.89340209960938, | |
| "learning_rate": 4.999e-05, | |
| "loss": 608.8174, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 2.3089840470193117, | |
| "grad_norm": 173.0674285888672, | |
| "learning_rate": 5.499000000000001e-05, | |
| "loss": 575.3464, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 2.5188916876574305, | |
| "grad_norm": 119.65222930908203, | |
| "learning_rate": 5.999e-05, | |
| "loss": 550.9334, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 2.72879932829555, | |
| "grad_norm": 163.03219604492188, | |
| "learning_rate": 6.499000000000001e-05, | |
| "loss": 532.3316, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 2.938706968933669, | |
| "grad_norm": 103.94256591796875, | |
| "learning_rate": 6.999e-05, | |
| "loss": 516.1778, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 3.1486146095717884, | |
| "grad_norm": 137.8463134765625, | |
| "learning_rate": 7.499e-05, | |
| "loss": 502.3519, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 3.3585222502099077, | |
| "grad_norm": 93.18783569335938, | |
| "learning_rate": 7.999000000000001e-05, | |
| "loss": 493.4963, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 3.568429890848027, | |
| "grad_norm": 81.264404296875, | |
| "learning_rate": 8.499e-05, | |
| "loss": 486.5324, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 3.7783375314861463, | |
| "grad_norm": 96.80731201171875, | |
| "learning_rate": 8.999000000000001e-05, | |
| "loss": 477.9452, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 3.988245172124265, | |
| "grad_norm": 99.205322265625, | |
| "learning_rate": 9.499e-05, | |
| "loss": 463.8162, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 4.198152812762385, | |
| "grad_norm": 118.66422271728516, | |
| "learning_rate": 9.999000000000001e-05, | |
| "loss": 447.3158, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 4.198152812762385, | |
| "eval_accuracy": 0.16583655638721168, | |
| "eval_loss": 428.842041015625, | |
| "eval_runtime": 3.6172, | |
| "eval_samples_per_second": 132.423, | |
| "eval_steps_per_second": 1.382, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 4.408060453400504, | |
| "grad_norm": 144.82168579101562, | |
| "learning_rate": 9.899293642785066e-05, | |
| "loss": 430.4443, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 4.617968094038623, | |
| "grad_norm": 166.46974182128906, | |
| "learning_rate": 9.798385469223008e-05, | |
| "loss": 416.3207, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 4.827875734676742, | |
| "grad_norm": 232.3704071044922, | |
| "learning_rate": 9.69747729566095e-05, | |
| "loss": 403.2591, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 5.037783375314861, | |
| "grad_norm": 254.5363311767578, | |
| "learning_rate": 9.59656912209889e-05, | |
| "loss": 390.4413, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 5.247691015952981, | |
| "grad_norm": 270.5792236328125, | |
| "learning_rate": 9.495660948536832e-05, | |
| "loss": 377.3261, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 5.4575986565911, | |
| "grad_norm": 244.77581787109375, | |
| "learning_rate": 9.394752774974774e-05, | |
| "loss": 365.2654, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 5.667506297229219, | |
| "grad_norm": 267.1288757324219, | |
| "learning_rate": 9.293844601412714e-05, | |
| "loss": 352.7486, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 5.877413937867338, | |
| "grad_norm": 315.1538391113281, | |
| "learning_rate": 9.192936427850656e-05, | |
| "loss": 338.1035, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 6.087321578505458, | |
| "grad_norm": 344.386962890625, | |
| "learning_rate": 9.092028254288598e-05, | |
| "loss": 323.3387, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 6.297229219143577, | |
| "grad_norm": 345.59320068359375, | |
| "learning_rate": 8.991120080726539e-05, | |
| "loss": 308.7323, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 6.507136859781696, | |
| "grad_norm": 373.200439453125, | |
| "learning_rate": 8.890211907164481e-05, | |
| "loss": 293.0822, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 6.717044500419815, | |
| "grad_norm": 387.7672119140625, | |
| "learning_rate": 8.789303733602423e-05, | |
| "loss": 276.3089, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 6.926952141057934, | |
| "grad_norm": 380.2518615722656, | |
| "learning_rate": 8.688395560040363e-05, | |
| "loss": 257.2038, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 7.136859781696054, | |
| "grad_norm": 306.2325134277344, | |
| "learning_rate": 8.587487386478305e-05, | |
| "loss": 221.4098, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 7.346767422334173, | |
| "grad_norm": 244.85455322265625, | |
| "learning_rate": 8.486579212916247e-05, | |
| "loss": 200.0532, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 7.5566750629722925, | |
| "grad_norm": 243.97576904296875, | |
| "learning_rate": 8.385671039354188e-05, | |
| "loss": 186.9741, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 7.766582703610411, | |
| "grad_norm": 201.4752655029297, | |
| "learning_rate": 8.28476286579213e-05, | |
| "loss": 176.1845, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 7.97649034424853, | |
| "grad_norm": 180.40164184570312, | |
| "learning_rate": 8.183854692230071e-05, | |
| "loss": 167.2202, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 8.18639798488665, | |
| "grad_norm": 169.40943908691406, | |
| "learning_rate": 8.082946518668012e-05, | |
| "loss": 158.6278, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 8.39630562552477, | |
| "grad_norm": 142.1228790283203, | |
| "learning_rate": 7.982038345105954e-05, | |
| "loss": 151.7919, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 8.39630562552477, | |
| "eval_accuracy": 0.48157916888801594, | |
| "eval_loss": 135.78590393066406, | |
| "eval_runtime": 3.5122, | |
| "eval_samples_per_second": 136.38, | |
| "eval_steps_per_second": 1.424, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 8.606213266162888, | |
| "grad_norm": 155.93728637695312, | |
| "learning_rate": 7.881130171543896e-05, | |
| "loss": 145.7355, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 8.816120906801007, | |
| "grad_norm": 139.52723693847656, | |
| "learning_rate": 7.780221997981836e-05, | |
| "loss": 140.6628, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 9.026028547439127, | |
| "grad_norm": 146.34320068359375, | |
| "learning_rate": 7.679313824419778e-05, | |
| "loss": 135.8976, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 9.235936188077247, | |
| "grad_norm": 129.9792938232422, | |
| "learning_rate": 7.57840565085772e-05, | |
| "loss": 131.474, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 9.445843828715365, | |
| "grad_norm": 122.61577606201172, | |
| "learning_rate": 7.477497477295662e-05, | |
| "loss": 127.8668, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 9.655751469353484, | |
| "grad_norm": 130.75668334960938, | |
| "learning_rate": 7.376589303733603e-05, | |
| "loss": 124.6147, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 9.865659109991604, | |
| "grad_norm": 130.1267547607422, | |
| "learning_rate": 7.275681130171544e-05, | |
| "loss": 121.6354, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 10.075566750629722, | |
| "grad_norm": 138.63009643554688, | |
| "learning_rate": 7.174772956609486e-05, | |
| "loss": 118.5567, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 10.285474391267842, | |
| "grad_norm": 128.6032257080078, | |
| "learning_rate": 7.073864783047427e-05, | |
| "loss": 115.905, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 10.495382031905962, | |
| "grad_norm": 117.67427825927734, | |
| "learning_rate": 6.972956609485369e-05, | |
| "loss": 113.5025, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 10.705289672544081, | |
| "grad_norm": 132.61322021484375, | |
| "learning_rate": 6.872048435923311e-05, | |
| "loss": 111.2961, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 10.9151973131822, | |
| "grad_norm": 124.02651977539062, | |
| "learning_rate": 6.771140262361251e-05, | |
| "loss": 109.3261, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 11.125104953820319, | |
| "grad_norm": 116.80554962158203, | |
| "learning_rate": 6.670232088799193e-05, | |
| "loss": 106.9513, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 11.335012594458439, | |
| "grad_norm": 120.98764038085938, | |
| "learning_rate": 6.569323915237135e-05, | |
| "loss": 105.1095, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 11.544920235096557, | |
| "grad_norm": 109.36351776123047, | |
| "learning_rate": 6.468415741675076e-05, | |
| "loss": 103.5172, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 11.754827875734676, | |
| "grad_norm": 115.52029418945312, | |
| "learning_rate": 6.367507568113018e-05, | |
| "loss": 101.8046, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 11.964735516372796, | |
| "grad_norm": 110.40377044677734, | |
| "learning_rate": 6.26659939455096e-05, | |
| "loss": 100.335, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 12.174643157010916, | |
| "grad_norm": 106.47163391113281, | |
| "learning_rate": 6.1656912209889e-05, | |
| "loss": 98.6005, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 12.384550797649034, | |
| "grad_norm": 114.08397674560547, | |
| "learning_rate": 6.064783047426842e-05, | |
| "loss": 97.2174, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 12.594458438287154, | |
| "grad_norm": 108.48992156982422, | |
| "learning_rate": 5.963874873864783e-05, | |
| "loss": 95.9257, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 12.594458438287154, | |
| "eval_accuracy": 0.5308223969555362, | |
| "eval_loss": 84.8065414428711, | |
| "eval_runtime": 3.6726, | |
| "eval_samples_per_second": 130.426, | |
| "eval_steps_per_second": 1.361, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 12.804366078925273, | |
| "grad_norm": 116.85979461669922, | |
| "learning_rate": 5.8629667003027243e-05, | |
| "loss": 94.7173, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 13.014273719563391, | |
| "grad_norm": 101.37419891357422, | |
| "learning_rate": 5.762058526740667e-05, | |
| "loss": 93.6031, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 13.224181360201511, | |
| "grad_norm": 101.53105926513672, | |
| "learning_rate": 5.661150353178608e-05, | |
| "loss": 92.2381, | |
| "step": 31500 | |
| }, | |
| { | |
| "epoch": 13.43408900083963, | |
| "grad_norm": 131.7721405029297, | |
| "learning_rate": 5.5602421796165494e-05, | |
| "loss": 91.2291, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 13.64399664147775, | |
| "grad_norm": 147.26918029785156, | |
| "learning_rate": 5.4593340060544906e-05, | |
| "loss": 90.265, | |
| "step": 32500 | |
| }, | |
| { | |
| "epoch": 13.853904282115868, | |
| "grad_norm": 107.17750549316406, | |
| "learning_rate": 5.358425832492432e-05, | |
| "loss": 89.2605, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 14.063811922753988, | |
| "grad_norm": 135.6392364501953, | |
| "learning_rate": 5.257517658930373e-05, | |
| "loss": 88.1679, | |
| "step": 33500 | |
| }, | |
| { | |
| "epoch": 14.273719563392108, | |
| "grad_norm": 131.0419158935547, | |
| "learning_rate": 5.1566094853683156e-05, | |
| "loss": 87.1749, | |
| "step": 34000 | |
| }, | |
| { | |
| "epoch": 14.483627204030226, | |
| "grad_norm": 110.07307434082031, | |
| "learning_rate": 5.055701311806257e-05, | |
| "loss": 86.2937, | |
| "step": 34500 | |
| }, | |
| { | |
| "epoch": 14.693534844668346, | |
| "grad_norm": 107.72781372070312, | |
| "learning_rate": 4.954793138244198e-05, | |
| "loss": 85.5873, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 14.903442485306465, | |
| "grad_norm": 91.38671112060547, | |
| "learning_rate": 4.853884964682139e-05, | |
| "loss": 84.7775, | |
| "step": 35500 | |
| }, | |
| { | |
| "epoch": 15.113350125944585, | |
| "grad_norm": 107.55229187011719, | |
| "learning_rate": 4.752976791120081e-05, | |
| "loss": 83.9663, | |
| "step": 36000 | |
| }, | |
| { | |
| "epoch": 15.323257766582703, | |
| "grad_norm": 93.78266906738281, | |
| "learning_rate": 4.6520686175580225e-05, | |
| "loss": 83.2307, | |
| "step": 36500 | |
| }, | |
| { | |
| "epoch": 15.533165407220823, | |
| "grad_norm": 94.07606506347656, | |
| "learning_rate": 4.551160443995964e-05, | |
| "loss": 82.4276, | |
| "step": 37000 | |
| }, | |
| { | |
| "epoch": 15.743073047858942, | |
| "grad_norm": 89.64346313476562, | |
| "learning_rate": 4.4502522704339056e-05, | |
| "loss": 81.7749, | |
| "step": 37500 | |
| }, | |
| { | |
| "epoch": 15.95298068849706, | |
| "grad_norm": 90.450927734375, | |
| "learning_rate": 4.349344096871847e-05, | |
| "loss": 81.2509, | |
| "step": 38000 | |
| }, | |
| { | |
| "epoch": 16.162888329135182, | |
| "grad_norm": 104.07128143310547, | |
| "learning_rate": 4.248435923309788e-05, | |
| "loss": 80.4306, | |
| "step": 38500 | |
| }, | |
| { | |
| "epoch": 16.3727959697733, | |
| "grad_norm": 97.70896911621094, | |
| "learning_rate": 4.14752774974773e-05, | |
| "loss": 79.7886, | |
| "step": 39000 | |
| }, | |
| { | |
| "epoch": 16.582703610411418, | |
| "grad_norm": 107.2480697631836, | |
| "learning_rate": 4.046619576185671e-05, | |
| "loss": 79.2786, | |
| "step": 39500 | |
| }, | |
| { | |
| "epoch": 16.79261125104954, | |
| "grad_norm": 94.23933410644531, | |
| "learning_rate": 3.9457114026236124e-05, | |
| "loss": 78.7736, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 16.79261125104954, | |
| "eval_accuracy": 0.5467926432656723, | |
| "eval_loss": 70.87545013427734, | |
| "eval_runtime": 2.7701, | |
| "eval_samples_per_second": 172.917, | |
| "eval_steps_per_second": 1.805, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 17.002518891687657, | |
| "grad_norm": 95.26321411132812, | |
| "learning_rate": 3.844803229061554e-05, | |
| "loss": 78.2565, | |
| "step": 40500 | |
| }, | |
| { | |
| "epoch": 17.212426532325775, | |
| "grad_norm": 97.84280395507812, | |
| "learning_rate": 3.7438950554994956e-05, | |
| "loss": 77.6301, | |
| "step": 41000 | |
| }, | |
| { | |
| "epoch": 17.422334172963897, | |
| "grad_norm": 118.63774871826172, | |
| "learning_rate": 3.642986881937437e-05, | |
| "loss": 77.0761, | |
| "step": 41500 | |
| }, | |
| { | |
| "epoch": 17.632241813602015, | |
| "grad_norm": 120.7638168334961, | |
| "learning_rate": 3.542078708375379e-05, | |
| "loss": 76.71, | |
| "step": 42000 | |
| }, | |
| { | |
| "epoch": 17.842149454240133, | |
| "grad_norm": 96.10567474365234, | |
| "learning_rate": 3.44117053481332e-05, | |
| "loss": 76.231, | |
| "step": 42500 | |
| }, | |
| { | |
| "epoch": 18.052057094878254, | |
| "grad_norm": 109.37013244628906, | |
| "learning_rate": 3.340262361251261e-05, | |
| "loss": 75.7584, | |
| "step": 43000 | |
| }, | |
| { | |
| "epoch": 18.261964735516372, | |
| "grad_norm": 81.04769134521484, | |
| "learning_rate": 3.239354187689203e-05, | |
| "loss": 75.1399, | |
| "step": 43500 | |
| }, | |
| { | |
| "epoch": 18.471872376154494, | |
| "grad_norm": 107.45811462402344, | |
| "learning_rate": 3.138446014127144e-05, | |
| "loss": 74.9177, | |
| "step": 44000 | |
| }, | |
| { | |
| "epoch": 18.68178001679261, | |
| "grad_norm": 91.4131088256836, | |
| "learning_rate": 3.0375378405650862e-05, | |
| "loss": 74.4926, | |
| "step": 44500 | |
| }, | |
| { | |
| "epoch": 18.89168765743073, | |
| "grad_norm": 103.88982391357422, | |
| "learning_rate": 2.9366296670030274e-05, | |
| "loss": 74.1645, | |
| "step": 45000 | |
| }, | |
| { | |
| "epoch": 19.10159529806885, | |
| "grad_norm": 101.31675720214844, | |
| "learning_rate": 2.8357214934409686e-05, | |
| "loss": 73.6817, | |
| "step": 45500 | |
| }, | |
| { | |
| "epoch": 19.31150293870697, | |
| "grad_norm": 97.94842529296875, | |
| "learning_rate": 2.7348133198789106e-05, | |
| "loss": 73.2026, | |
| "step": 46000 | |
| }, | |
| { | |
| "epoch": 19.521410579345087, | |
| "grad_norm": 85.70989990234375, | |
| "learning_rate": 2.6339051463168518e-05, | |
| "loss": 72.9578, | |
| "step": 46500 | |
| }, | |
| { | |
| "epoch": 19.73131821998321, | |
| "grad_norm": 93.49505615234375, | |
| "learning_rate": 2.532996972754793e-05, | |
| "loss": 72.7161, | |
| "step": 47000 | |
| }, | |
| { | |
| "epoch": 19.941225860621326, | |
| "grad_norm": 89.21786499023438, | |
| "learning_rate": 2.4320887991927346e-05, | |
| "loss": 72.3488, | |
| "step": 47500 | |
| }, | |
| { | |
| "epoch": 20.151133501259444, | |
| "grad_norm": 86.77904510498047, | |
| "learning_rate": 2.331180625630676e-05, | |
| "loss": 71.9474, | |
| "step": 48000 | |
| }, | |
| { | |
| "epoch": 20.361041141897566, | |
| "grad_norm": 104.81689453125, | |
| "learning_rate": 2.2302724520686177e-05, | |
| "loss": 71.6556, | |
| "step": 48500 | |
| }, | |
| { | |
| "epoch": 20.570948782535684, | |
| "grad_norm": 85.92950439453125, | |
| "learning_rate": 2.129364278506559e-05, | |
| "loss": 71.3194, | |
| "step": 49000 | |
| }, | |
| { | |
| "epoch": 20.7808564231738, | |
| "grad_norm": 86.256103515625, | |
| "learning_rate": 2.0284561049445005e-05, | |
| "loss": 71.1533, | |
| "step": 49500 | |
| }, | |
| { | |
| "epoch": 20.990764063811923, | |
| "grad_norm": 92.44305419921875, | |
| "learning_rate": 1.927547931382442e-05, | |
| "loss": 70.9704, | |
| "step": 50000 | |
| }, | |
| { | |
| "epoch": 20.990764063811923, | |
| "eval_accuracy": 0.5510415790506631, | |
| "eval_loss": 63.090843200683594, | |
| "eval_runtime": 2.8803, | |
| "eval_samples_per_second": 166.303, | |
| "eval_steps_per_second": 1.736, | |
| "step": 50000 | |
| }, | |
| { | |
| "epoch": 21.20067170445004, | |
| "grad_norm": 80.83171844482422, | |
| "learning_rate": 1.8266397578203836e-05, | |
| "loss": 70.5263, | |
| "step": 50500 | |
| }, | |
| { | |
| "epoch": 21.410579345088163, | |
| "grad_norm": 86.68183898925781, | |
| "learning_rate": 1.725731584258325e-05, | |
| "loss": 70.3711, | |
| "step": 51000 | |
| }, | |
| { | |
| "epoch": 21.62048698572628, | |
| "grad_norm": 75.10340118408203, | |
| "learning_rate": 1.6248234106962664e-05, | |
| "loss": 70.0563, | |
| "step": 51500 | |
| }, | |
| { | |
| "epoch": 21.8303946263644, | |
| "grad_norm": 80.75882720947266, | |
| "learning_rate": 1.523915237134208e-05, | |
| "loss": 69.9151, | |
| "step": 52000 | |
| }, | |
| { | |
| "epoch": 22.04030226700252, | |
| "grad_norm": 80.31729888916016, | |
| "learning_rate": 1.4230070635721492e-05, | |
| "loss": 69.7056, | |
| "step": 52500 | |
| }, | |
| { | |
| "epoch": 22.250209907640638, | |
| "grad_norm": 79.79283905029297, | |
| "learning_rate": 1.3220988900100908e-05, | |
| "loss": 69.3878, | |
| "step": 53000 | |
| }, | |
| { | |
| "epoch": 22.460117548278756, | |
| "grad_norm": 88.36482238769531, | |
| "learning_rate": 1.2211907164480324e-05, | |
| "loss": 69.2763, | |
| "step": 53500 | |
| }, | |
| { | |
| "epoch": 22.670025188916878, | |
| "grad_norm": 82.19640350341797, | |
| "learning_rate": 1.1202825428859738e-05, | |
| "loss": 69.1046, | |
| "step": 54000 | |
| }, | |
| { | |
| "epoch": 22.879932829554996, | |
| "grad_norm": 77.26073455810547, | |
| "learning_rate": 1.0193743693239152e-05, | |
| "loss": 68.8851, | |
| "step": 54500 | |
| }, | |
| { | |
| "epoch": 23.089840470193113, | |
| "grad_norm": 89.86582946777344, | |
| "learning_rate": 9.184661957618567e-06, | |
| "loss": 68.7541, | |
| "step": 55000 | |
| }, | |
| { | |
| "epoch": 23.299748110831235, | |
| "grad_norm": 70.37703704833984, | |
| "learning_rate": 8.175580221997981e-06, | |
| "loss": 68.5219, | |
| "step": 55500 | |
| }, | |
| { | |
| "epoch": 23.509655751469353, | |
| "grad_norm": 71.59619140625, | |
| "learning_rate": 7.166498486377397e-06, | |
| "loss": 68.3648, | |
| "step": 56000 | |
| }, | |
| { | |
| "epoch": 23.719563392107474, | |
| "grad_norm": 69.95408630371094, | |
| "learning_rate": 6.157416750756812e-06, | |
| "loss": 68.3101, | |
| "step": 56500 | |
| }, | |
| { | |
| "epoch": 23.929471032745592, | |
| "grad_norm": 71.7118148803711, | |
| "learning_rate": 5.148335015136227e-06, | |
| "loss": 68.1746, | |
| "step": 57000 | |
| }, | |
| { | |
| "epoch": 24.13937867338371, | |
| "grad_norm": 67.84393310546875, | |
| "learning_rate": 4.139253279515641e-06, | |
| "loss": 67.9893, | |
| "step": 57500 | |
| }, | |
| { | |
| "epoch": 24.349286314021832, | |
| "grad_norm": 70.03734588623047, | |
| "learning_rate": 3.130171543895056e-06, | |
| "loss": 67.9278, | |
| "step": 58000 | |
| }, | |
| { | |
| "epoch": 24.55919395465995, | |
| "grad_norm": 67.93866729736328, | |
| "learning_rate": 2.1210898082744702e-06, | |
| "loss": 67.8202, | |
| "step": 58500 | |
| }, | |
| { | |
| "epoch": 24.769101595298068, | |
| "grad_norm": 61.62685775756836, | |
| "learning_rate": 1.112008072653885e-06, | |
| "loss": 67.7249, | |
| "step": 59000 | |
| }, | |
| { | |
| "epoch": 24.97900923593619, | |
| "grad_norm": 63.67091751098633, | |
| "learning_rate": 1.0292633703329971e-07, | |
| "loss": 67.6809, | |
| "step": 59500 | |
| }, | |
| { | |
| "epoch": 25.0, | |
| "step": 59550, | |
| "total_flos": 7.577162542158336e+17, | |
| "train_loss": 301.8068682845954, | |
| "train_runtime": 49685.0337, | |
| "train_samples_per_second": 115.044, | |
| "train_steps_per_second": 1.199 | |
| } | |
| ], | |
| "logging_steps": 500, | |
| "max_steps": 59550, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 25, | |
| "save_steps": 10000, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 7.577162542158336e+17, | |
| "train_batch_size": 96, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |