| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 19.51219512195122, | |
| "eval_steps": 100, | |
| "global_step": 500, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.1951219512195122, | |
| "grad_norm": 3.5239181518554688, | |
| "learning_rate": 4.999451708687114e-05, | |
| "loss": 1.1522, | |
| "num_input_tokens_seen": 2231104, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.3902439024390244, | |
| "grad_norm": 1.2978184223175049, | |
| "learning_rate": 4.997807075247146e-05, | |
| "loss": 0.8861, | |
| "num_input_tokens_seen": 4468608, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.5853658536585366, | |
| "grad_norm": 1.0176712274551392, | |
| "learning_rate": 4.995066821070679e-05, | |
| "loss": 0.8351, | |
| "num_input_tokens_seen": 6779584, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.7804878048780488, | |
| "grad_norm": 0.6517723798751831, | |
| "learning_rate": 4.991232148123761e-05, | |
| "loss": 0.7797, | |
| "num_input_tokens_seen": 9106880, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.975609756097561, | |
| "grad_norm": 0.6675965785980225, | |
| "learning_rate": 4.9863047384206835e-05, | |
| "loss": 0.7575, | |
| "num_input_tokens_seen": 11435136, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 1.170731707317073, | |
| "grad_norm": 1.225784182548523, | |
| "learning_rate": 4.980286753286195e-05, | |
| "loss": 0.7345, | |
| "num_input_tokens_seen": 13738880, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 1.3658536585365852, | |
| "grad_norm": 0.4991011321544647, | |
| "learning_rate": 4.9731808324074717e-05, | |
| "loss": 0.7443, | |
| "num_input_tokens_seen": 16001920, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 1.5609756097560976, | |
| "grad_norm": 0.7908358573913574, | |
| "learning_rate": 4.964990092676263e-05, | |
| "loss": 0.7211, | |
| "num_input_tokens_seen": 18282944, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 1.7560975609756098, | |
| "grad_norm": 0.9223726391792297, | |
| "learning_rate": 4.9557181268217227e-05, | |
| "loss": 0.7306, | |
| "num_input_tokens_seen": 20575296, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 1.951219512195122, | |
| "grad_norm": 0.8891558051109314, | |
| "learning_rate": 4.9453690018345144e-05, | |
| "loss": 0.7039, | |
| "num_input_tokens_seen": 22849856, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 2.1463414634146343, | |
| "grad_norm": 1.3324096202850342, | |
| "learning_rate": 4.933947257182901e-05, | |
| "loss": 0.7044, | |
| "num_input_tokens_seen": 25176896, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 2.341463414634146, | |
| "grad_norm": 1.0357320308685303, | |
| "learning_rate": 4.9214579028215776e-05, | |
| "loss": 0.6768, | |
| "num_input_tokens_seen": 27457472, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 2.5365853658536586, | |
| "grad_norm": 0.6939280033111572, | |
| "learning_rate": 4.907906416994146e-05, | |
| "loss": 0.6906, | |
| "num_input_tokens_seen": 29772544, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 2.7317073170731705, | |
| "grad_norm": 0.7578678727149963, | |
| "learning_rate": 4.893298743830168e-05, | |
| "loss": 0.6595, | |
| "num_input_tokens_seen": 32013760, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 2.926829268292683, | |
| "grad_norm": 0.548279345035553, | |
| "learning_rate": 4.877641290737884e-05, | |
| "loss": 0.6514, | |
| "num_input_tokens_seen": 34342848, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 3.1219512195121952, | |
| "grad_norm": 1.1982917785644531, | |
| "learning_rate": 4.860940925593703e-05, | |
| "loss": 0.6498, | |
| "num_input_tokens_seen": 36592000, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 3.317073170731707, | |
| "grad_norm": 1.0446078777313232, | |
| "learning_rate": 4.843204973729729e-05, | |
| "loss": 0.6537, | |
| "num_input_tokens_seen": 38878528, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 3.5121951219512195, | |
| "grad_norm": 0.7469108700752258, | |
| "learning_rate": 4.8244412147206284e-05, | |
| "loss": 0.5996, | |
| "num_input_tokens_seen": 41156672, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 3.7073170731707314, | |
| "grad_norm": 0.7562854886054993, | |
| "learning_rate": 4.8046578789712515e-05, | |
| "loss": 0.6289, | |
| "num_input_tokens_seen": 43475456, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 3.902439024390244, | |
| "grad_norm": 0.7673454880714417, | |
| "learning_rate": 4.783863644106502e-05, | |
| "loss": 0.6607, | |
| "num_input_tokens_seen": 45798272, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 3.902439024390244, | |
| "eval_loss": 0.6353042125701904, | |
| "eval_runtime": 174.6543, | |
| "eval_samples_per_second": 12.07, | |
| "eval_steps_per_second": 0.504, | |
| "num_input_tokens_seen": 45798272, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 4.097560975609756, | |
| "grad_norm": 0.6962050199508667, | |
| "learning_rate": 4.762067631165049e-05, | |
| "loss": 0.5924, | |
| "num_input_tokens_seen": 48035456, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 4.2926829268292686, | |
| "grad_norm": 1.080012559890747, | |
| "learning_rate": 4.7392794005985326e-05, | |
| "loss": 0.6107, | |
| "num_input_tokens_seen": 50295680, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 4.487804878048781, | |
| "grad_norm": 0.6142390966415405, | |
| "learning_rate": 4.715508948078037e-05, | |
| "loss": 0.6016, | |
| "num_input_tokens_seen": 52611072, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 4.682926829268292, | |
| "grad_norm": 0.7695103883743286, | |
| "learning_rate": 4.690766700109659e-05, | |
| "loss": 0.5946, | |
| "num_input_tokens_seen": 54817920, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 4.878048780487805, | |
| "grad_norm": 0.7141166925430298, | |
| "learning_rate": 4.665063509461097e-05, | |
| "loss": 0.5926, | |
| "num_input_tokens_seen": 57135552, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 5.073170731707317, | |
| "grad_norm": 0.9515812993049622, | |
| "learning_rate": 4.638410650401267e-05, | |
| "loss": 0.5986, | |
| "num_input_tokens_seen": 59455680, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 5.2682926829268295, | |
| "grad_norm": 1.5192577838897705, | |
| "learning_rate": 4.610819813755038e-05, | |
| "loss": 0.5619, | |
| "num_input_tokens_seen": 61782208, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 5.463414634146342, | |
| "grad_norm": 1.0420284271240234, | |
| "learning_rate": 4.5823031017752485e-05, | |
| "loss": 0.5739, | |
| "num_input_tokens_seen": 64051520, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 5.658536585365853, | |
| "grad_norm": 0.6723448038101196, | |
| "learning_rate": 4.5528730228342605e-05, | |
| "loss": 0.5673, | |
| "num_input_tokens_seen": 66320192, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 5.853658536585366, | |
| "grad_norm": 0.8727495074272156, | |
| "learning_rate": 4.522542485937369e-05, | |
| "loss": 0.5607, | |
| "num_input_tokens_seen": 68535424, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 6.048780487804878, | |
| "grad_norm": 1.1162189245224, | |
| "learning_rate": 4.491324795060491e-05, | |
| "loss": 0.5711, | |
| "num_input_tokens_seen": 70861632, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 6.2439024390243905, | |
| "grad_norm": 1.0553621053695679, | |
| "learning_rate": 4.4592336433146e-05, | |
| "loss": 0.5437, | |
| "num_input_tokens_seen": 73124480, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 6.439024390243903, | |
| "grad_norm": 1.105603575706482, | |
| "learning_rate": 4.426283106939474e-05, | |
| "loss": 0.5437, | |
| "num_input_tokens_seen": 75439232, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 6.634146341463414, | |
| "grad_norm": 0.9295886158943176, | |
| "learning_rate": 4.3924876391293915e-05, | |
| "loss": 0.5321, | |
| "num_input_tokens_seen": 77689536, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 6.829268292682927, | |
| "grad_norm": 0.9514282941818237, | |
| "learning_rate": 4.357862063693486e-05, | |
| "loss": 0.5075, | |
| "num_input_tokens_seen": 80000000, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 7.024390243902439, | |
| "grad_norm": 0.641543447971344, | |
| "learning_rate": 4.3224215685535294e-05, | |
| "loss": 0.5246, | |
| "num_input_tokens_seen": 82294336, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 7.219512195121951, | |
| "grad_norm": 0.6666741967201233, | |
| "learning_rate": 4.2861816990820084e-05, | |
| "loss": 0.5021, | |
| "num_input_tokens_seen": 84609856, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 7.414634146341464, | |
| "grad_norm": 0.9410921931266785, | |
| "learning_rate": 4.249158351283414e-05, | |
| "loss": 0.4766, | |
| "num_input_tokens_seen": 86956608, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 7.609756097560975, | |
| "grad_norm": 0.9510065913200378, | |
| "learning_rate": 4.211367764821722e-05, | |
| "loss": 0.4922, | |
| "num_input_tokens_seen": 89199040, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 7.804878048780488, | |
| "grad_norm": 0.9778364896774292, | |
| "learning_rate": 4.172826515897146e-05, | |
| "loss": 0.5149, | |
| "num_input_tokens_seen": 91464512, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 7.804878048780488, | |
| "eval_loss": 0.6269230246543884, | |
| "eval_runtime": 173.5455, | |
| "eval_samples_per_second": 12.147, | |
| "eval_steps_per_second": 0.507, | |
| "num_input_tokens_seen": 91464512, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "grad_norm": 1.1030495166778564, | |
| "learning_rate": 4.133551509975264e-05, | |
| "loss": 0.5019, | |
| "num_input_tokens_seen": 93698368, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 8.195121951219512, | |
| "grad_norm": 0.8688378930091858, | |
| "learning_rate": 4.093559974371725e-05, | |
| "loss": 0.4275, | |
| "num_input_tokens_seen": 96008256, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 8.390243902439025, | |
| "grad_norm": 0.9601365327835083, | |
| "learning_rate": 4.052869450695776e-05, | |
| "loss": 0.4545, | |
| "num_input_tokens_seen": 98232832, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 8.585365853658537, | |
| "grad_norm": 0.9640222191810608, | |
| "learning_rate": 4.011497787155938e-05, | |
| "loss": 0.4577, | |
| "num_input_tokens_seen": 100520128, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 8.78048780487805, | |
| "grad_norm": 0.7299013137817383, | |
| "learning_rate": 3.969463130731183e-05, | |
| "loss": 0.4589, | |
| "num_input_tokens_seen": 102814976, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 8.975609756097562, | |
| "grad_norm": 0.6459980607032776, | |
| "learning_rate": 3.92678391921108e-05, | |
| "loss": 0.4482, | |
| "num_input_tokens_seen": 105142528, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 9.170731707317072, | |
| "grad_norm": 0.8188865184783936, | |
| "learning_rate": 3.883478873108361e-05, | |
| "loss": 0.4301, | |
| "num_input_tokens_seen": 107394880, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 9.365853658536585, | |
| "grad_norm": 0.7741889953613281, | |
| "learning_rate": 3.8395669874474915e-05, | |
| "loss": 0.3852, | |
| "num_input_tokens_seen": 109706560, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 9.560975609756097, | |
| "grad_norm": 0.7010043263435364, | |
| "learning_rate": 3.795067523432826e-05, | |
| "loss": 0.3886, | |
| "num_input_tokens_seen": 111993664, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 9.75609756097561, | |
| "grad_norm": 0.5523029565811157, | |
| "learning_rate": 3.7500000000000003e-05, | |
| "loss": 0.4028, | |
| "num_input_tokens_seen": 114255232, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 9.951219512195122, | |
| "grad_norm": 0.5837187170982361, | |
| "learning_rate": 3.704384185254288e-05, | |
| "loss": 0.3847, | |
| "num_input_tokens_seen": 116578560, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 10.146341463414634, | |
| "grad_norm": 0.7814375162124634, | |
| "learning_rate": 3.6582400877996546e-05, | |
| "loss": 0.3516, | |
| "num_input_tokens_seen": 118836800, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 10.341463414634147, | |
| "grad_norm": 0.9329513907432556, | |
| "learning_rate": 3.611587947962319e-05, | |
| "loss": 0.3504, | |
| "num_input_tokens_seen": 121148608, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 10.536585365853659, | |
| "grad_norm": 0.5801968574523926, | |
| "learning_rate": 3.564448228912682e-05, | |
| "loss": 0.3303, | |
| "num_input_tokens_seen": 123490176, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 10.731707317073171, | |
| "grad_norm": 0.5544772744178772, | |
| "learning_rate": 3.516841607689501e-05, | |
| "loss": 0.336, | |
| "num_input_tokens_seen": 125741376, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 10.926829268292684, | |
| "grad_norm": 0.5629393458366394, | |
| "learning_rate": 3.4687889661302576e-05, | |
| "loss": 0.337, | |
| "num_input_tokens_seen": 127996480, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 11.121951219512194, | |
| "grad_norm": 0.6132429242134094, | |
| "learning_rate": 3.4203113817116957e-05, | |
| "loss": 0.3102, | |
| "num_input_tokens_seen": 130286208, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 11.317073170731707, | |
| "grad_norm": 0.6265811920166016, | |
| "learning_rate": 3.3714301183045385e-05, | |
| "loss": 0.2777, | |
| "num_input_tokens_seen": 132600256, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 11.512195121951219, | |
| "grad_norm": 0.7565940618515015, | |
| "learning_rate": 3.322166616846458e-05, | |
| "loss": 0.2875, | |
| "num_input_tokens_seen": 134908928, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 11.707317073170731, | |
| "grad_norm": 0.8063022494316101, | |
| "learning_rate": 3.272542485937369e-05, | |
| "loss": 0.2768, | |
| "num_input_tokens_seen": 137181248, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 11.707317073170731, | |
| "eval_loss": 0.718697726726532, | |
| "eval_runtime": 173.4793, | |
| "eval_samples_per_second": 12.151, | |
| "eval_steps_per_second": 0.507, | |
| "num_input_tokens_seen": 137181248, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 11.902439024390244, | |
| "grad_norm": 0.7160292267799377, | |
| "learning_rate": 3.222579492361179e-05, | |
| "loss": 0.2823, | |
| "num_input_tokens_seen": 139385728, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 12.097560975609756, | |
| "grad_norm": 0.6414353251457214, | |
| "learning_rate": 3.172299551538164e-05, | |
| "loss": 0.27, | |
| "num_input_tokens_seen": 141670400, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 12.292682926829269, | |
| "grad_norm": 0.6131883859634399, | |
| "learning_rate": 3.121724717912138e-05, | |
| "loss": 0.2308, | |
| "num_input_tokens_seen": 143944512, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 12.487804878048781, | |
| "grad_norm": 0.46672749519348145, | |
| "learning_rate": 3.0708771752766394e-05, | |
| "loss": 0.2236, | |
| "num_input_tokens_seen": 146200128, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 12.682926829268293, | |
| "grad_norm": 0.48453348875045776, | |
| "learning_rate": 3.0197792270443982e-05, | |
| "loss": 0.2227, | |
| "num_input_tokens_seen": 148494912, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 12.878048780487806, | |
| "grad_norm": 0.5639292597770691, | |
| "learning_rate": 2.9684532864643122e-05, | |
| "loss": 0.2236, | |
| "num_input_tokens_seen": 150765632, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 13.073170731707316, | |
| "grad_norm": 0.45188578963279724, | |
| "learning_rate": 2.916921866790256e-05, | |
| "loss": 0.2193, | |
| "num_input_tokens_seen": 153025984, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 13.268292682926829, | |
| "grad_norm": 0.5362280011177063, | |
| "learning_rate": 2.8652075714060295e-05, | |
| "loss": 0.1804, | |
| "num_input_tokens_seen": 155359744, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 13.463414634146341, | |
| "grad_norm": 0.46796223521232605, | |
| "learning_rate": 2.8133330839107608e-05, | |
| "loss": 0.1763, | |
| "num_input_tokens_seen": 157653760, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 13.658536585365853, | |
| "grad_norm": 0.6022464036941528, | |
| "learning_rate": 2.761321158169134e-05, | |
| "loss": 0.1785, | |
| "num_input_tokens_seen": 159970112, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 13.853658536585366, | |
| "grad_norm": 0.6007694005966187, | |
| "learning_rate": 2.7091946083307896e-05, | |
| "loss": 0.1847, | |
| "num_input_tokens_seen": 162266048, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 14.048780487804878, | |
| "grad_norm": 0.5330171585083008, | |
| "learning_rate": 2.656976298823284e-05, | |
| "loss": 0.1765, | |
| "num_input_tokens_seen": 164502784, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 14.24390243902439, | |
| "grad_norm": 0.4508293569087982, | |
| "learning_rate": 2.604689134322999e-05, | |
| "loss": 0.1433, | |
| "num_input_tokens_seen": 166793792, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 14.439024390243903, | |
| "grad_norm": 0.4898134469985962, | |
| "learning_rate": 2.5523560497083926e-05, | |
| "loss": 0.1436, | |
| "num_input_tokens_seen": 169105536, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 14.634146341463415, | |
| "grad_norm": 0.5252612829208374, | |
| "learning_rate": 2.5e-05, | |
| "loss": 0.1456, | |
| "num_input_tokens_seen": 171380160, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 14.829268292682928, | |
| "grad_norm": 0.45608317852020264, | |
| "learning_rate": 2.447643950291608e-05, | |
| "loss": 0.1361, | |
| "num_input_tokens_seen": 173636544, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 15.024390243902438, | |
| "grad_norm": 0.5009132623672485, | |
| "learning_rate": 2.3953108656770016e-05, | |
| "loss": 0.1381, | |
| "num_input_tokens_seen": 175912832, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 15.21951219512195, | |
| "grad_norm": 0.4505337178707123, | |
| "learning_rate": 2.3430237011767167e-05, | |
| "loss": 0.1121, | |
| "num_input_tokens_seen": 178157824, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 15.414634146341463, | |
| "grad_norm": 0.43384668231010437, | |
| "learning_rate": 2.2908053916692117e-05, | |
| "loss": 0.1103, | |
| "num_input_tokens_seen": 180468736, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 15.609756097560975, | |
| "grad_norm": 0.4433673620223999, | |
| "learning_rate": 2.238678841830867e-05, | |
| "loss": 0.1126, | |
| "num_input_tokens_seen": 182788736, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 15.609756097560975, | |
| "eval_loss": 0.8370733261108398, | |
| "eval_runtime": 173.2462, | |
| "eval_samples_per_second": 12.168, | |
| "eval_steps_per_second": 0.508, | |
| "num_input_tokens_seen": 182788736, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 15.804878048780488, | |
| "grad_norm": 0.39896300435066223, | |
| "learning_rate": 2.186666916089239e-05, | |
| "loss": 0.1114, | |
| "num_input_tokens_seen": 185050944, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 16.0, | |
| "grad_norm": 0.3975840210914612, | |
| "learning_rate": 2.1347924285939714e-05, | |
| "loss": 0.1135, | |
| "num_input_tokens_seen": 187289920, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 16.195121951219512, | |
| "grad_norm": 0.3381832242012024, | |
| "learning_rate": 2.0830781332097446e-05, | |
| "loss": 0.0935, | |
| "num_input_tokens_seen": 189562944, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 16.390243902439025, | |
| "grad_norm": 0.3201667070388794, | |
| "learning_rate": 2.031546713535688e-05, | |
| "loss": 0.0851, | |
| "num_input_tokens_seen": 191813696, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 16.585365853658537, | |
| "grad_norm": 0.38780859112739563, | |
| "learning_rate": 1.980220772955602e-05, | |
| "loss": 0.0897, | |
| "num_input_tokens_seen": 194094144, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 16.78048780487805, | |
| "grad_norm": 0.31540414690971375, | |
| "learning_rate": 1.9291228247233605e-05, | |
| "loss": 0.0886, | |
| "num_input_tokens_seen": 196385984, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 16.975609756097562, | |
| "grad_norm": 0.5678117275238037, | |
| "learning_rate": 1.8782752820878634e-05, | |
| "loss": 0.0899, | |
| "num_input_tokens_seen": 198700032, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 17.170731707317074, | |
| "grad_norm": 0.4425772726535797, | |
| "learning_rate": 1.827700448461836e-05, | |
| "loss": 0.0753, | |
| "num_input_tokens_seen": 201036480, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 17.365853658536587, | |
| "grad_norm": 0.32113131880760193, | |
| "learning_rate": 1.7774205076388206e-05, | |
| "loss": 0.0724, | |
| "num_input_tokens_seen": 203331520, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 17.5609756097561, | |
| "grad_norm": 0.2662774920463562, | |
| "learning_rate": 1.7274575140626318e-05, | |
| "loss": 0.0711, | |
| "num_input_tokens_seen": 205627072, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 17.75609756097561, | |
| "grad_norm": 0.27264896035194397, | |
| "learning_rate": 1.677833383153542e-05, | |
| "loss": 0.0719, | |
| "num_input_tokens_seen": 207845568, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 17.951219512195124, | |
| "grad_norm": 0.29512110352516174, | |
| "learning_rate": 1.6285698816954624e-05, | |
| "loss": 0.0704, | |
| "num_input_tokens_seen": 210180736, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 18.146341463414632, | |
| "grad_norm": 0.252193421125412, | |
| "learning_rate": 1.5796886182883053e-05, | |
| "loss": 0.0666, | |
| "num_input_tokens_seen": 212454016, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 18.341463414634145, | |
| "grad_norm": 0.23161688446998596, | |
| "learning_rate": 1.5312110338697426e-05, | |
| "loss": 0.0564, | |
| "num_input_tokens_seen": 214714240, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 18.536585365853657, | |
| "grad_norm": 0.25446033477783203, | |
| "learning_rate": 1.4831583923104999e-05, | |
| "loss": 0.0617, | |
| "num_input_tokens_seen": 217009664, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 18.73170731707317, | |
| "grad_norm": 0.2264089584350586, | |
| "learning_rate": 1.4355517710873184e-05, | |
| "loss": 0.0587, | |
| "num_input_tokens_seen": 219270336, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 18.926829268292682, | |
| "grad_norm": 0.21758581697940826, | |
| "learning_rate": 1.388412052037682e-05, | |
| "loss": 0.0573, | |
| "num_input_tokens_seen": 221604736, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 19.121951219512194, | |
| "grad_norm": 0.21286001801490784, | |
| "learning_rate": 1.3417599122003464e-05, | |
| "loss": 0.0537, | |
| "num_input_tokens_seen": 223843968, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 19.317073170731707, | |
| "grad_norm": 0.19537770748138428, | |
| "learning_rate": 1.2956158147457115e-05, | |
| "loss": 0.0505, | |
| "num_input_tokens_seen": 226198144, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 19.51219512195122, | |
| "grad_norm": 0.18156716227531433, | |
| "learning_rate": 1.2500000000000006e-05, | |
| "loss": 0.0494, | |
| "num_input_tokens_seen": 228521280, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 19.51219512195122, | |
| "eval_loss": 0.9114091396331787, | |
| "eval_runtime": 173.5404, | |
| "eval_samples_per_second": 12.147, | |
| "eval_steps_per_second": 0.507, | |
| "num_input_tokens_seen": 228521280, | |
| "step": 500 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 750, | |
| "num_input_tokens_seen": 228521280, | |
| "num_train_epochs": 30, | |
| "save_steps": 100, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 9.694721607846593e+18, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |