| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.994161801501251, | |
| "eval_steps": 500, | |
| "global_step": 447, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.006672226855713094, | |
| "grad_norm": 0.3732840765814082, | |
| "learning_rate": 2.2222222222222224e-07, | |
| "loss": 1.6647, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.013344453711426188, | |
| "grad_norm": 0.383618969512302, | |
| "learning_rate": 4.444444444444445e-07, | |
| "loss": 1.7073, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.020016680567139282, | |
| "grad_norm": 0.40769314152645725, | |
| "learning_rate": 6.666666666666667e-07, | |
| "loss": 1.7505, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.026688907422852376, | |
| "grad_norm": 0.4016260533661679, | |
| "learning_rate": 8.88888888888889e-07, | |
| "loss": 1.6952, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.03336113427856547, | |
| "grad_norm": 0.40174340733938796, | |
| "learning_rate": 1.111111111111111e-06, | |
| "loss": 1.7844, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.040033361134278564, | |
| "grad_norm": 0.40847133525152324, | |
| "learning_rate": 1.3333333333333334e-06, | |
| "loss": 1.7765, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.04670558798999166, | |
| "grad_norm": 0.40604889051688275, | |
| "learning_rate": 1.5555555555555558e-06, | |
| "loss": 1.6995, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.05337781484570475, | |
| "grad_norm": 0.3894313779958969, | |
| "learning_rate": 1.777777777777778e-06, | |
| "loss": 1.7128, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.060050041701417846, | |
| "grad_norm": 0.38627201738251066, | |
| "learning_rate": 2.0000000000000003e-06, | |
| "loss": 1.7789, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.06672226855713094, | |
| "grad_norm": 0.379007955799278, | |
| "learning_rate": 2.222222222222222e-06, | |
| "loss": 1.6907, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.07339449541284404, | |
| "grad_norm": 0.40293473844915795, | |
| "learning_rate": 2.4444444444444447e-06, | |
| "loss": 1.712, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.08006672226855713, | |
| "grad_norm": 0.3796921388422985, | |
| "learning_rate": 2.666666666666667e-06, | |
| "loss": 1.717, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.08673894912427023, | |
| "grad_norm": 0.363416530641245, | |
| "learning_rate": 2.888888888888889e-06, | |
| "loss": 1.6075, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.09341117597998332, | |
| "grad_norm": 0.42180934217352195, | |
| "learning_rate": 3.1111111111111116e-06, | |
| "loss": 1.7974, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.10008340283569642, | |
| "grad_norm": 0.4120160113974808, | |
| "learning_rate": 3.3333333333333333e-06, | |
| "loss": 1.7485, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.1067556296914095, | |
| "grad_norm": 0.39278536183126506, | |
| "learning_rate": 3.555555555555556e-06, | |
| "loss": 1.5707, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.1134278565471226, | |
| "grad_norm": 0.4167599840383631, | |
| "learning_rate": 3.777777777777778e-06, | |
| "loss": 1.613, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.12010008340283569, | |
| "grad_norm": 0.4238875949015934, | |
| "learning_rate": 4.000000000000001e-06, | |
| "loss": 1.6816, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.1267723102585488, | |
| "grad_norm": 0.4720616555526343, | |
| "learning_rate": 4.222222222222223e-06, | |
| "loss": 1.7593, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.13344453711426188, | |
| "grad_norm": 0.4525263926368364, | |
| "learning_rate": 4.444444444444444e-06, | |
| "loss": 1.775, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.14011676396997497, | |
| "grad_norm": 0.48266597054805727, | |
| "learning_rate": 4.666666666666667e-06, | |
| "loss": 1.791, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.14678899082568808, | |
| "grad_norm": 0.45939880890244206, | |
| "learning_rate": 4.888888888888889e-06, | |
| "loss": 1.7338, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.15346121768140117, | |
| "grad_norm": 0.47764567898261373, | |
| "learning_rate": 5.1111111111111115e-06, | |
| "loss": 1.7483, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.16013344453711426, | |
| "grad_norm": 0.5017631959610884, | |
| "learning_rate": 5.333333333333334e-06, | |
| "loss": 1.7481, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.16680567139282734, | |
| "grad_norm": 0.4893567689357815, | |
| "learning_rate": 5.555555555555557e-06, | |
| "loss": 1.7136, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.17347789824854046, | |
| "grad_norm": 0.5105291717073872, | |
| "learning_rate": 5.777777777777778e-06, | |
| "loss": 1.7602, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.18015012510425354, | |
| "grad_norm": 0.46974781533253324, | |
| "learning_rate": 6e-06, | |
| "loss": 1.6153, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.18682235195996663, | |
| "grad_norm": 0.5383405971129123, | |
| "learning_rate": 6.222222222222223e-06, | |
| "loss": 1.6602, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.19349457881567975, | |
| "grad_norm": 0.5952748064288939, | |
| "learning_rate": 6.444444444444445e-06, | |
| "loss": 1.7177, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.20016680567139283, | |
| "grad_norm": 0.5399245254089347, | |
| "learning_rate": 6.666666666666667e-06, | |
| "loss": 1.7092, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.20683903252710592, | |
| "grad_norm": 0.5598787056956456, | |
| "learning_rate": 6.88888888888889e-06, | |
| "loss": 1.67, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.213511259382819, | |
| "grad_norm": 0.5992687712634374, | |
| "learning_rate": 7.111111111111112e-06, | |
| "loss": 1.7355, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.22018348623853212, | |
| "grad_norm": 0.6046798477314653, | |
| "learning_rate": 7.333333333333333e-06, | |
| "loss": 1.7322, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.2268557130942452, | |
| "grad_norm": 0.6320084975028277, | |
| "learning_rate": 7.555555555555556e-06, | |
| "loss": 1.8031, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.2335279399499583, | |
| "grad_norm": 0.5904036648454649, | |
| "learning_rate": 7.77777777777778e-06, | |
| "loss": 1.6945, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.24020016680567138, | |
| "grad_norm": 0.6104299430520795, | |
| "learning_rate": 8.000000000000001e-06, | |
| "loss": 1.6777, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.2468723936613845, | |
| "grad_norm": 0.6047786223470227, | |
| "learning_rate": 8.222222222222222e-06, | |
| "loss": 1.6366, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.2535446205170976, | |
| "grad_norm": 0.6387094238927912, | |
| "learning_rate": 8.444444444444446e-06, | |
| "loss": 1.6636, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.2602168473728107, | |
| "grad_norm": 0.6120207071048721, | |
| "learning_rate": 8.666666666666668e-06, | |
| "loss": 1.6292, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.26688907422852376, | |
| "grad_norm": 0.6088068565354802, | |
| "learning_rate": 8.888888888888888e-06, | |
| "loss": 1.6047, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.2735613010842369, | |
| "grad_norm": 0.6091668716984593, | |
| "learning_rate": 9.111111111111112e-06, | |
| "loss": 1.6258, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.28023352793994993, | |
| "grad_norm": 0.5682506824419054, | |
| "learning_rate": 9.333333333333334e-06, | |
| "loss": 1.5699, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.28690575479566305, | |
| "grad_norm": 0.5635810130587849, | |
| "learning_rate": 9.555555555555556e-06, | |
| "loss": 1.6314, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.29357798165137616, | |
| "grad_norm": 0.5841358241959886, | |
| "learning_rate": 9.777777777777779e-06, | |
| "loss": 1.5456, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.3002502085070892, | |
| "grad_norm": 0.5617293142392159, | |
| "learning_rate": 1e-05, | |
| "loss": 1.5272, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.30692243536280234, | |
| "grad_norm": 0.47772319074173303, | |
| "learning_rate": 9.999847318844664e-06, | |
| "loss": 1.5373, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.31359466221851545, | |
| "grad_norm": 0.510823048338545, | |
| "learning_rate": 9.999389284703265e-06, | |
| "loss": 1.6097, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.3202668890742285, | |
| "grad_norm": 0.5010538752769951, | |
| "learning_rate": 9.99862592554908e-06, | |
| "loss": 1.5344, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.3269391159299416, | |
| "grad_norm": 0.47226467172138603, | |
| "learning_rate": 9.99755728800233e-06, | |
| "loss": 1.5023, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.3336113427856547, | |
| "grad_norm": 0.4849966025454849, | |
| "learning_rate": 9.996183437327342e-06, | |
| "loss": 1.4975, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.3402835696413678, | |
| "grad_norm": 0.47126668835079394, | |
| "learning_rate": 9.994504457428557e-06, | |
| "loss": 1.5127, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.3469557964970809, | |
| "grad_norm": 0.4301425546280333, | |
| "learning_rate": 9.992520450845415e-06, | |
| "loss": 1.4456, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.353628023352794, | |
| "grad_norm": 0.4378182305752446, | |
| "learning_rate": 9.99023153874608e-06, | |
| "loss": 1.4695, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.3603002502085071, | |
| "grad_norm": 0.4297960171566437, | |
| "learning_rate": 9.987637860920053e-06, | |
| "loss": 1.5113, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.3669724770642202, | |
| "grad_norm": 0.4154036526185639, | |
| "learning_rate": 9.984739575769619e-06, | |
| "loss": 1.3419, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.37364470391993326, | |
| "grad_norm": 0.42590048753833515, | |
| "learning_rate": 9.981536860300191e-06, | |
| "loss": 1.4191, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.3803169307756464, | |
| "grad_norm": 0.4067719643823426, | |
| "learning_rate": 9.978029910109491e-06, | |
| "loss": 1.3983, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.3869891576313595, | |
| "grad_norm": 0.3969278459704296, | |
| "learning_rate": 9.9742189393756e-06, | |
| "loss": 1.3794, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.39366138448707255, | |
| "grad_norm": 0.3854670128621615, | |
| "learning_rate": 9.970104180843881e-06, | |
| "loss": 1.4305, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.40033361134278567, | |
| "grad_norm": 0.37743755476343405, | |
| "learning_rate": 9.965685885812773e-06, | |
| "loss": 1.4283, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.4070058381984987, | |
| "grad_norm": 0.3610253450226918, | |
| "learning_rate": 9.960964324118428e-06, | |
| "loss": 1.4135, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.41367806505421184, | |
| "grad_norm": 0.38314151063998836, | |
| "learning_rate": 9.955939784118246e-06, | |
| "loss": 1.4194, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.42035029190992496, | |
| "grad_norm": 0.391502563403777, | |
| "learning_rate": 9.950612572673255e-06, | |
| "loss": 1.3669, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.427022518765638, | |
| "grad_norm": 0.3606335039104152, | |
| "learning_rate": 9.944983015129376e-06, | |
| "loss": 1.3249, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.43369474562135113, | |
| "grad_norm": 0.3836886872726812, | |
| "learning_rate": 9.939051455297548e-06, | |
| "loss": 1.3323, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.44036697247706424, | |
| "grad_norm": 0.39513066746369385, | |
| "learning_rate": 9.932818255432733e-06, | |
| "loss": 1.3228, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.4470391993327773, | |
| "grad_norm": 0.3776395809498074, | |
| "learning_rate": 9.926283796211796e-06, | |
| "loss": 1.4202, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.4537114261884904, | |
| "grad_norm": 0.3536441979654702, | |
| "learning_rate": 9.919448476710248e-06, | |
| "loss": 1.3169, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.4603836530442035, | |
| "grad_norm": 0.35401704664244255, | |
| "learning_rate": 9.91231271437788e-06, | |
| "loss": 1.2869, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.4670558798999166, | |
| "grad_norm": 0.364086698791685, | |
| "learning_rate": 9.904876945013272e-06, | |
| "loss": 1.3558, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.4737281067556297, | |
| "grad_norm": 0.36643453456415837, | |
| "learning_rate": 9.89714162273716e-06, | |
| "loss": 1.2972, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.48040033361134277, | |
| "grad_norm": 0.35341632594055883, | |
| "learning_rate": 9.889107219964726e-06, | |
| "loss": 1.288, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.4870725604670559, | |
| "grad_norm": 0.35727226233776593, | |
| "learning_rate": 9.880774227376727e-06, | |
| "loss": 1.2919, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.493744787322769, | |
| "grad_norm": 0.36655136111497844, | |
| "learning_rate": 9.872143153889538e-06, | |
| "loss": 1.2864, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.5004170141784821, | |
| "grad_norm": 0.3580519341051248, | |
| "learning_rate": 9.863214526624065e-06, | |
| "loss": 1.3545, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.5070892410341952, | |
| "grad_norm": 0.3251286992451364, | |
| "learning_rate": 9.853988890873563e-06, | |
| "loss": 1.2463, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.5137614678899083, | |
| "grad_norm": 0.35154208731758474, | |
| "learning_rate": 9.844466810070319e-06, | |
| "loss": 1.2842, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.5204336947456214, | |
| "grad_norm": 0.35360327701201805, | |
| "learning_rate": 9.834648865751254e-06, | |
| "loss": 1.2074, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.5271059216013344, | |
| "grad_norm": 0.33827984275163614, | |
| "learning_rate": 9.8245356575224e-06, | |
| "loss": 1.2719, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.5337781484570475, | |
| "grad_norm": 0.34688277661237193, | |
| "learning_rate": 9.814127803022281e-06, | |
| "loss": 1.2809, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.5404503753127606, | |
| "grad_norm": 0.30446458893187955, | |
| "learning_rate": 9.803425937884202e-06, | |
| "loss": 1.2302, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.5471226021684737, | |
| "grad_norm": 0.31831189291520345, | |
| "learning_rate": 9.792430715697412e-06, | |
| "loss": 1.2329, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.5537948290241869, | |
| "grad_norm": 0.29801313071352337, | |
| "learning_rate": 9.781142807967205e-06, | |
| "loss": 1.2464, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.5604670558798999, | |
| "grad_norm": 0.3118984625981495, | |
| "learning_rate": 9.769562904073896e-06, | |
| "loss": 1.2388, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.567139282735613, | |
| "grad_norm": 0.28346870486129144, | |
| "learning_rate": 9.757691711230728e-06, | |
| "loss": 1.222, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.5738115095913261, | |
| "grad_norm": 0.3235682827850219, | |
| "learning_rate": 9.745529954440675e-06, | |
| "loss": 1.2167, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.5804837364470392, | |
| "grad_norm": 0.3025602896103518, | |
| "learning_rate": 9.733078376452172e-06, | |
| "loss": 1.2439, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.5871559633027523, | |
| "grad_norm": 0.2943206117574144, | |
| "learning_rate": 9.72033773771374e-06, | |
| "loss": 1.1714, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.5938281901584654, | |
| "grad_norm": 0.3043570350280811, | |
| "learning_rate": 9.707308816327557e-06, | |
| "loss": 1.2466, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.6005004170141784, | |
| "grad_norm": 0.3034637832523981, | |
| "learning_rate": 9.693992408001934e-06, | |
| "loss": 1.1737, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.6071726438698916, | |
| "grad_norm": 0.29378370873758436, | |
| "learning_rate": 9.680389326002708e-06, | |
| "loss": 1.2823, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.6138448707256047, | |
| "grad_norm": 0.25233208089640863, | |
| "learning_rate": 9.666500401103595e-06, | |
| "loss": 1.191, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.6205170975813178, | |
| "grad_norm": 0.264841896316846, | |
| "learning_rate": 9.652326481535434e-06, | |
| "loss": 1.1842, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.6271893244370309, | |
| "grad_norm": 0.276234903477221, | |
| "learning_rate": 9.63786843293439e-06, | |
| "loss": 1.1062, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.6338615512927439, | |
| "grad_norm": 0.25250932151404143, | |
| "learning_rate": 9.623127138289087e-06, | |
| "loss": 1.1036, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.640533778148457, | |
| "grad_norm": 0.27277526043956535, | |
| "learning_rate": 9.608103497886687e-06, | |
| "loss": 1.1332, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.6472060050041701, | |
| "grad_norm": 0.27812831319172243, | |
| "learning_rate": 9.592798429257899e-06, | |
| "loss": 1.1704, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.6538782318598833, | |
| "grad_norm": 0.24189134780796812, | |
| "learning_rate": 9.577212867120947e-06, | |
| "loss": 1.1644, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.6605504587155964, | |
| "grad_norm": 0.25645020937055246, | |
| "learning_rate": 9.561347763324484e-06, | |
| "loss": 1.147, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.6672226855713094, | |
| "grad_norm": 0.2595686790235878, | |
| "learning_rate": 9.545204086789461e-06, | |
| "loss": 1.2334, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.6738949124270225, | |
| "grad_norm": 0.24949960135623445, | |
| "learning_rate": 9.528782823449954e-06, | |
| "loss": 1.0824, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.6805671392827356, | |
| "grad_norm": 0.2457329952973373, | |
| "learning_rate": 9.512084976192944e-06, | |
| "loss": 1.1335, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.6872393661384487, | |
| "grad_norm": 0.26799579166131554, | |
| "learning_rate": 9.495111564797073e-06, | |
| "loss": 1.1403, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.6939115929941618, | |
| "grad_norm": 0.24843110380217398, | |
| "learning_rate": 9.477863625870371e-06, | |
| "loss": 1.1659, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.700583819849875, | |
| "grad_norm": 0.26596297978049915, | |
| "learning_rate": 9.460342212786933e-06, | |
| "loss": 1.1191, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.707256046705588, | |
| "grad_norm": 0.24609569923205743, | |
| "learning_rate": 9.442548395622596e-06, | |
| "loss": 1.1388, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.7139282735613011, | |
| "grad_norm": 0.30159400597802477, | |
| "learning_rate": 9.424483261089584e-06, | |
| "loss": 1.2705, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.7206005004170142, | |
| "grad_norm": 0.24911328963823176, | |
| "learning_rate": 9.406147912470142e-06, | |
| "loss": 1.1393, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.7272727272727273, | |
| "grad_norm": 0.2461854661834755, | |
| "learning_rate": 9.387543469549156e-06, | |
| "loss": 1.0897, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.7339449541284404, | |
| "grad_norm": 0.24427724214296034, | |
| "learning_rate": 9.368671068545761e-06, | |
| "loss": 1.0527, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.7406171809841534, | |
| "grad_norm": 0.24775041629357836, | |
| "learning_rate": 9.349531862043952e-06, | |
| "loss": 1.0695, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.7472894078398665, | |
| "grad_norm": 0.24100289437186362, | |
| "learning_rate": 9.330127018922195e-06, | |
| "loss": 1.0972, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.7539616346955796, | |
| "grad_norm": 0.22817385479165772, | |
| "learning_rate": 9.310457724282034e-06, | |
| "loss": 1.1354, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.7606338615512928, | |
| "grad_norm": 0.24351101122361835, | |
| "learning_rate": 9.290525179375722e-06, | |
| "loss": 1.1867, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.7673060884070059, | |
| "grad_norm": 0.23252679169251178, | |
| "learning_rate": 9.270330601532855e-06, | |
| "loss": 0.9749, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.773978315262719, | |
| "grad_norm": 0.24891852642577655, | |
| "learning_rate": 9.249875224086023e-06, | |
| "loss": 1.0428, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.780650542118432, | |
| "grad_norm": 0.2519892667083049, | |
| "learning_rate": 9.229160296295488e-06, | |
| "loss": 1.1525, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.7873227689741451, | |
| "grad_norm": 0.22842656407732698, | |
| "learning_rate": 9.208187083272895e-06, | |
| "loss": 1.0893, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.7939949958298582, | |
| "grad_norm": 0.23530353826931574, | |
| "learning_rate": 9.186956865904004e-06, | |
| "loss": 1.1461, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.8006672226855713, | |
| "grad_norm": 0.25930781535544006, | |
| "learning_rate": 9.165470940770458e-06, | |
| "loss": 1.0863, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.8073394495412844, | |
| "grad_norm": 0.24794978516182076, | |
| "learning_rate": 9.143730620070609e-06, | |
| "loss": 1.1216, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.8140116763969975, | |
| "grad_norm": 0.23948920252711808, | |
| "learning_rate": 9.121737231539369e-06, | |
| "loss": 1.1121, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.8206839032527106, | |
| "grad_norm": 0.2271047573553413, | |
| "learning_rate": 9.099492118367123e-06, | |
| "loss": 1.1543, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.8273561301084237, | |
| "grad_norm": 0.25315676717796304, | |
| "learning_rate": 9.076996639117708e-06, | |
| "loss": 1.1309, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.8340283569641368, | |
| "grad_norm": 0.24992513383867146, | |
| "learning_rate": 9.054252167645426e-06, | |
| "loss": 1.064, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.8407005838198499, | |
| "grad_norm": 0.24372502494677808, | |
| "learning_rate": 9.03126009301115e-06, | |
| "loss": 1.1069, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.8473728106755629, | |
| "grad_norm": 0.23106765610762486, | |
| "learning_rate": 9.008021819397488e-06, | |
| "loss": 1.1344, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.854045037531276, | |
| "grad_norm": 0.2638490777582001, | |
| "learning_rate": 8.984538766023024e-06, | |
| "loss": 1.1099, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.8607172643869891, | |
| "grad_norm": 0.2436361177633829, | |
| "learning_rate": 8.960812367055646e-06, | |
| "loss": 1.0879, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.8673894912427023, | |
| "grad_norm": 0.2435286627690007, | |
| "learning_rate": 8.93684407152496e-06, | |
| "loss": 1.1281, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.8740617180984154, | |
| "grad_norm": 0.2292808626156267, | |
| "learning_rate": 8.912635343233784e-06, | |
| "loss": 1.0638, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.8807339449541285, | |
| "grad_norm": 0.22525031331854853, | |
| "learning_rate": 8.888187660668762e-06, | |
| "loss": 1.1003, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.8874061718098415, | |
| "grad_norm": 0.22555143759216476, | |
| "learning_rate": 8.863502516910058e-06, | |
| "loss": 0.9963, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.8940783986655546, | |
| "grad_norm": 0.23206135282498955, | |
| "learning_rate": 8.838581419540183e-06, | |
| "loss": 1.0354, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.9007506255212677, | |
| "grad_norm": 0.2332933329851816, | |
| "learning_rate": 8.81342589055191e-06, | |
| "loss": 1.0629, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.9074228523769808, | |
| "grad_norm": 0.21598906842259502, | |
| "learning_rate": 8.788037466255334e-06, | |
| "loss": 1.0368, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.914095079232694, | |
| "grad_norm": 0.2510622429683598, | |
| "learning_rate": 8.762417697184034e-06, | |
| "loss": 1.018, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.920767306088407, | |
| "grad_norm": 0.25761970215631363, | |
| "learning_rate": 8.736568148000386e-06, | |
| "loss": 1.0496, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.9274395329441201, | |
| "grad_norm": 0.267375239616011, | |
| "learning_rate": 8.710490397400007e-06, | |
| "loss": 1.1176, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.9341117597998332, | |
| "grad_norm": 0.23104284046945558, | |
| "learning_rate": 8.684186038015327e-06, | |
| "loss": 1.0221, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.9407839866555463, | |
| "grad_norm": 0.2560542899746131, | |
| "learning_rate": 8.657656676318346e-06, | |
| "loss": 1.0761, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.9474562135112594, | |
| "grad_norm": 0.25675258631662545, | |
| "learning_rate": 8.630903932522496e-06, | |
| "loss": 1.0371, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.9541284403669725, | |
| "grad_norm": 0.2609140179674428, | |
| "learning_rate": 8.603929440483714e-06, | |
| "loss": 1.0895, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.9608006672226855, | |
| "grad_norm": 0.23996253406165996, | |
| "learning_rate": 8.576734847600639e-06, | |
| "loss": 1.0905, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.9674728940783986, | |
| "grad_norm": 0.2545575065577051, | |
| "learning_rate": 8.549321814714018e-06, | |
| "loss": 1.1054, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.9741451209341118, | |
| "grad_norm": 0.2514055738350408, | |
| "learning_rate": 8.521692016005262e-06, | |
| "loss": 1.1183, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.9808173477898249, | |
| "grad_norm": 0.2792023557673277, | |
| "learning_rate": 8.49384713889421e-06, | |
| "loss": 1.106, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.987489574645538, | |
| "grad_norm": 0.22340105561162074, | |
| "learning_rate": 8.46578888393606e-06, | |
| "loss": 1.0232, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.994161801501251, | |
| "grad_norm": 0.25568989349658056, | |
| "learning_rate": 8.43751896471753e-06, | |
| "loss": 1.0899, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 1.0066722268557131, | |
| "grad_norm": 0.7802055347834813, | |
| "learning_rate": 8.40903910775219e-06, | |
| "loss": 1.9774, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 1.0133444537114262, | |
| "grad_norm": 0.2595541681581563, | |
| "learning_rate": 8.380351052375023e-06, | |
| "loss": 1.045, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 1.0200166805671393, | |
| "grad_norm": 0.25723642565217814, | |
| "learning_rate": 8.35145655063621e-06, | |
| "loss": 1.0245, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 1.0266889074228525, | |
| "grad_norm": 0.27362291141864464, | |
| "learning_rate": 8.32235736719411e-06, | |
| "loss": 1.0868, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 1.0333611342785656, | |
| "grad_norm": 0.2458791941756008, | |
| "learning_rate": 8.293055279207503e-06, | |
| "loss": 1.0889, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 1.0400333611342785, | |
| "grad_norm": 0.31147103624378575, | |
| "learning_rate": 8.263552076227048e-06, | |
| "loss": 0.9933, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 1.0467055879899916, | |
| "grad_norm": 0.2672690864798242, | |
| "learning_rate": 8.233849560085994e-06, | |
| "loss": 1.0815, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 1.0533778148457047, | |
| "grad_norm": 0.2422074236035313, | |
| "learning_rate": 8.203949544790131e-06, | |
| "loss": 1.1015, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 1.0600500417014178, | |
| "grad_norm": 0.27579765876295836, | |
| "learning_rate": 8.173853856407011e-06, | |
| "loss": 1.0386, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 1.066722268557131, | |
| "grad_norm": 0.25869695071730053, | |
| "learning_rate": 8.143564332954426e-06, | |
| "loss": 1.0408, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 1.073394495412844, | |
| "grad_norm": 0.2902505945569461, | |
| "learning_rate": 8.113082824288145e-06, | |
| "loss": 1.0253, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 1.0800667222685572, | |
| "grad_norm": 0.23657511088258748, | |
| "learning_rate": 8.082411191988956e-06, | |
| "loss": 1.0282, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 1.0867389491242703, | |
| "grad_norm": 0.2841268408399986, | |
| "learning_rate": 8.051551309248961e-06, | |
| "loss": 1.1055, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 1.0934111759799834, | |
| "grad_norm": 0.26372063674010376, | |
| "learning_rate": 8.02050506075718e-06, | |
| "loss": 1.0589, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 1.1000834028356965, | |
| "grad_norm": 0.27690358093295814, | |
| "learning_rate": 7.989274342584446e-06, | |
| "loss": 0.949, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 1.1067556296914094, | |
| "grad_norm": 0.2712773483518729, | |
| "learning_rate": 7.957861062067614e-06, | |
| "loss": 1.0729, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 1.1134278565471225, | |
| "grad_norm": 0.30516214609809805, | |
| "learning_rate": 7.926267137693066e-06, | |
| "loss": 1.037, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 1.1201000834028356, | |
| "grad_norm": 0.2576349226719825, | |
| "learning_rate": 7.894494498979558e-06, | |
| "loss": 1.0497, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 1.1267723102585487, | |
| "grad_norm": 0.24801992242605783, | |
| "learning_rate": 7.86254508636036e-06, | |
| "loss": 1.0313, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 1.1334445371142619, | |
| "grad_norm": 0.3166344694109662, | |
| "learning_rate": 7.830420851064767e-06, | |
| "loss": 1.0967, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 1.140116763969975, | |
| "grad_norm": 0.24751199793145093, | |
| "learning_rate": 7.798123754998922e-06, | |
| "loss": 1.0031, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 1.146788990825688, | |
| "grad_norm": 0.2586884189404244, | |
| "learning_rate": 7.765655770625997e-06, | |
| "loss": 1.0219, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 1.1534612176814012, | |
| "grad_norm": 0.235632902703973, | |
| "learning_rate": 7.733018880845747e-06, | |
| "loss": 1.0207, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 1.1601334445371143, | |
| "grad_norm": 0.25234095064849044, | |
| "learning_rate": 7.70021507887338e-06, | |
| "loss": 1.0061, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 1.1668056713928274, | |
| "grad_norm": 0.2326635498094972, | |
| "learning_rate": 7.667246368117852e-06, | |
| "loss": 1.0587, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 1.1734778982485405, | |
| "grad_norm": 0.2649027893834019, | |
| "learning_rate": 7.634114762059504e-06, | |
| "loss": 0.9703, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 1.1801501251042534, | |
| "grad_norm": 0.2728476550775381, | |
| "learning_rate": 7.600822284127091e-06, | |
| "loss": 1.0392, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 1.1868223519599665, | |
| "grad_norm": 0.24549914761915478, | |
| "learning_rate": 7.56737096757421e-06, | |
| "loss": 0.9954, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 1.1934945788156797, | |
| "grad_norm": 0.2815251829331306, | |
| "learning_rate": 7.533762855355126e-06, | |
| "loss": 1.0291, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 1.2001668056713928, | |
| "grad_norm": 0.26632117503517927, | |
| "learning_rate": 7.500000000000001e-06, | |
| "loss": 1.1419, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 1.206839032527106, | |
| "grad_norm": 0.2966134642339934, | |
| "learning_rate": 7.466084463489537e-06, | |
| "loss": 1.0036, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 1.213511259382819, | |
| "grad_norm": 0.26223635549866403, | |
| "learning_rate": 7.432018317129056e-06, | |
| "loss": 1.0285, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 1.2201834862385321, | |
| "grad_norm": 0.30911626688190436, | |
| "learning_rate": 7.39780364142199e-06, | |
| "loss": 1.0041, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 1.2268557130942452, | |
| "grad_norm": 0.283762152043488, | |
| "learning_rate": 7.363442525942827e-06, | |
| "loss": 1.0517, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 1.2335279399499584, | |
| "grad_norm": 0.3425818937559889, | |
| "learning_rate": 7.32893706920949e-06, | |
| "loss": 1.0177, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 1.2402001668056715, | |
| "grad_norm": 0.35518044492578266, | |
| "learning_rate": 7.294289378555179e-06, | |
| "loss": 1.039, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 1.2468723936613846, | |
| "grad_norm": 0.2953658322519591, | |
| "learning_rate": 7.25950156999967e-06, | |
| "loss": 1.0239, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 1.2535446205170975, | |
| "grad_norm": 0.30127312537141965, | |
| "learning_rate": 7.2245757681200835e-06, | |
| "loss": 0.9763, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 1.2602168473728108, | |
| "grad_norm": 0.282352211179054, | |
| "learning_rate": 7.189514105921132e-06, | |
| "loss": 0.991, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 1.2668890742285237, | |
| "grad_norm": 0.27942599700546883, | |
| "learning_rate": 7.1543187247048525e-06, | |
| "loss": 0.9997, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 1.2735613010842368, | |
| "grad_norm": 0.34032643799945067, | |
| "learning_rate": 7.118991773939832e-06, | |
| "loss": 1.0323, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 1.28023352793995, | |
| "grad_norm": 0.2833978681025291, | |
| "learning_rate": 7.083535411129934e-06, | |
| "loss": 0.965, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 1.286905754795663, | |
| "grad_norm": 0.28066028962170614, | |
| "learning_rate": 7.047951801682533e-06, | |
| "loss": 0.9827, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 1.2935779816513762, | |
| "grad_norm": 0.3451242223474228, | |
| "learning_rate": 7.01224311877627e-06, | |
| "loss": 1.013, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 1.3002502085070893, | |
| "grad_norm": 0.28576160899229214, | |
| "learning_rate": 6.976411543228328e-06, | |
| "loss": 1.0147, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 1.3069224353628024, | |
| "grad_norm": 0.26405586969472095, | |
| "learning_rate": 6.9404592633612486e-06, | |
| "loss": 1.0222, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 1.3135946622185155, | |
| "grad_norm": 0.23979743466387593, | |
| "learning_rate": 6.904388474869284e-06, | |
| "loss": 0.9763, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 1.3202668890742286, | |
| "grad_norm": 0.26278250196725494, | |
| "learning_rate": 6.8682013806842985e-06, | |
| "loss": 1.0177, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 1.3269391159299415, | |
| "grad_norm": 0.3616613201338565, | |
| "learning_rate": 6.831900190841232e-06, | |
| "loss": 1.0189, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 1.3336113427856546, | |
| "grad_norm": 0.317805954218629, | |
| "learning_rate": 6.795487122343124e-06, | |
| "loss": 1.0761, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 1.3402835696413677, | |
| "grad_norm": 0.3267310563691638, | |
| "learning_rate": 6.758964399025721e-06, | |
| "loss": 1.0103, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 1.3469557964970809, | |
| "grad_norm": 0.3367394677015302, | |
| "learning_rate": 6.722334251421665e-06, | |
| "loss": 0.9504, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 1.353628023352794, | |
| "grad_norm": 0.323985054612315, | |
| "learning_rate": 6.685598916624254e-06, | |
| "loss": 1.0425, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 1.360300250208507, | |
| "grad_norm": 0.31744623672148714, | |
| "learning_rate": 6.648760638150833e-06, | |
| "loss": 1.0284, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 1.3669724770642202, | |
| "grad_norm": 0.32506007928496355, | |
| "learning_rate": 6.611821665805769e-06, | |
| "loss": 1.0494, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 1.3736447039199333, | |
| "grad_norm": 0.31683692701232863, | |
| "learning_rate": 6.574784255543052e-06, | |
| "loss": 0.9341, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 1.3803169307756464, | |
| "grad_norm": 0.339014481901513, | |
| "learning_rate": 6.537650669328518e-06, | |
| "loss": 0.9351, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 1.3869891576313595, | |
| "grad_norm": 0.2908501088905977, | |
| "learning_rate": 6.500423175001705e-06, | |
| "loss": 0.9636, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 1.3936613844870727, | |
| "grad_norm": 0.29622441402682287, | |
| "learning_rate": 6.4631040461373494e-06, | |
| "loss": 0.991, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 1.4003336113427856, | |
| "grad_norm": 0.259706989545213, | |
| "learning_rate": 6.4256955619065375e-06, | |
| "loss": 1.0265, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 1.4070058381984987, | |
| "grad_norm": 0.27351346614577177, | |
| "learning_rate": 6.388200006937503e-06, | |
| "loss": 1.1127, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 1.4136780650542118, | |
| "grad_norm": 0.2569362561775431, | |
| "learning_rate": 6.350619671176111e-06, | |
| "loss": 0.9848, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 1.420350291909925, | |
| "grad_norm": 0.3031398363213745, | |
| "learning_rate": 6.312956849745993e-06, | |
| "loss": 1.0439, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 1.427022518765638, | |
| "grad_norm": 0.29741800552671255, | |
| "learning_rate": 6.275213842808383e-06, | |
| "loss": 1.0295, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 1.4336947456213511, | |
| "grad_norm": 0.2858593332853522, | |
| "learning_rate": 6.237392955421644e-06, | |
| "loss": 0.9632, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 1.4403669724770642, | |
| "grad_norm": 0.2717253920058818, | |
| "learning_rate": 6.19949649740049e-06, | |
| "loss": 0.9708, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 1.4470391993327774, | |
| "grad_norm": 0.2755282096882548, | |
| "learning_rate": 6.161526783174917e-06, | |
| "loss": 1.0894, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 1.4537114261884905, | |
| "grad_norm": 0.277037632676125, | |
| "learning_rate": 6.123486131648859e-06, | |
| "loss": 1.0682, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 1.4603836530442034, | |
| "grad_norm": 0.3020180443959382, | |
| "learning_rate": 6.085376866058569e-06, | |
| "loss": 1.0672, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 1.4670558798999167, | |
| "grad_norm": 0.2819163434265071, | |
| "learning_rate": 6.047201313830724e-06, | |
| "loss": 0.9994, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 1.4737281067556296, | |
| "grad_norm": 0.2800183935957482, | |
| "learning_rate": 6.0089618064402896e-06, | |
| "loss": 1.022, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 1.4804003336113427, | |
| "grad_norm": 0.2719182825453672, | |
| "learning_rate": 5.970660679268139e-06, | |
| "loss": 1.0183, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 1.4870725604670558, | |
| "grad_norm": 0.2922309116254843, | |
| "learning_rate": 5.932300271458406e-06, | |
| "loss": 0.9777, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 1.493744787322769, | |
| "grad_norm": 0.2744738064356238, | |
| "learning_rate": 5.893882925775648e-06, | |
| "loss": 0.9841, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 1.500417014178482, | |
| "grad_norm": 0.2618241413789887, | |
| "learning_rate": 5.85541098846175e-06, | |
| "loss": 0.9789, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 1.5070892410341952, | |
| "grad_norm": 0.29331697175604643, | |
| "learning_rate": 5.816886809092651e-06, | |
| "loss": 0.9889, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 1.5137614678899083, | |
| "grad_norm": 0.293938259486948, | |
| "learning_rate": 5.778312740434835e-06, | |
| "loss": 1.0398, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 1.5204336947456214, | |
| "grad_norm": 0.3149805373750932, | |
| "learning_rate": 5.73969113830165e-06, | |
| "loss": 1.0356, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 1.5271059216013345, | |
| "grad_norm": 0.2882296792753476, | |
| "learning_rate": 5.701024361409431e-06, | |
| "loss": 1.066, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 1.5337781484570474, | |
| "grad_norm": 0.3264587076590878, | |
| "learning_rate": 5.66231477123344e-06, | |
| "loss": 0.9717, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 1.5404503753127607, | |
| "grad_norm": 0.28816187489358347, | |
| "learning_rate": 5.623564731863664e-06, | |
| "loss": 1.0055, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 1.5471226021684736, | |
| "grad_norm": 0.26876813399349714, | |
| "learning_rate": 5.584776609860414e-06, | |
| "loss": 1.0118, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 1.553794829024187, | |
| "grad_norm": 0.25656136721296807, | |
| "learning_rate": 5.545952774109798e-06, | |
| "loss": 1.0441, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 1.5604670558798999, | |
| "grad_norm": 0.2927283912839115, | |
| "learning_rate": 5.507095595679059e-06, | |
| "loss": 0.9875, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 1.567139282735613, | |
| "grad_norm": 0.26847113941733736, | |
| "learning_rate": 5.468207447671755e-06, | |
| "loss": 0.942, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 1.573811509591326, | |
| "grad_norm": 0.35285592222755574, | |
| "learning_rate": 5.42929070508283e-06, | |
| "loss": 1.0382, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 1.5804837364470392, | |
| "grad_norm": 0.28225631872241425, | |
| "learning_rate": 5.390347744653576e-06, | |
| "loss": 1.0041, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 1.5871559633027523, | |
| "grad_norm": 0.2894670959272741, | |
| "learning_rate": 5.351380944726465e-06, | |
| "loss": 1.0017, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 1.5938281901584654, | |
| "grad_norm": 0.28265496307817667, | |
| "learning_rate": 5.312392685099915e-06, | |
| "loss": 1.0073, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 1.6005004170141786, | |
| "grad_norm": 0.3412461315872378, | |
| "learning_rate": 5.2733853468829295e-06, | |
| "loss": 0.9839, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 1.6071726438698914, | |
| "grad_norm": 0.2954038595818766, | |
| "learning_rate": 5.234361312349701e-06, | |
| "loss": 1.0132, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 1.6138448707256048, | |
| "grad_norm": 0.2840364660523181, | |
| "learning_rate": 5.195322964794098e-06, | |
| "loss": 0.9473, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 1.6205170975813177, | |
| "grad_norm": 0.2986907397693315, | |
| "learning_rate": 5.156272688384123e-06, | |
| "loss": 1.0077, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 1.627189324437031, | |
| "grad_norm": 0.292176824846797, | |
| "learning_rate": 5.117212868016303e-06, | |
| "loss": 0.9315, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 1.633861551292744, | |
| "grad_norm": 0.3135774476957044, | |
| "learning_rate": 5.07814588917004e-06, | |
| "loss": 1.0088, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 1.640533778148457, | |
| "grad_norm": 0.31002794668270833, | |
| "learning_rate": 5.03907413776192e-06, | |
| "loss": 1.007, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 1.6472060050041701, | |
| "grad_norm": 0.3045456413966571, | |
| "learning_rate": 5e-06, | |
| "loss": 0.9872, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 1.6538782318598833, | |
| "grad_norm": 0.344623415610902, | |
| "learning_rate": 4.96092586223808e-06, | |
| "loss": 0.9343, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 1.6605504587155964, | |
| "grad_norm": 0.2888473591105003, | |
| "learning_rate": 4.921854110829962e-06, | |
| "loss": 1.0071, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 1.6672226855713093, | |
| "grad_norm": 0.3424869438582918, | |
| "learning_rate": 4.882787131983698e-06, | |
| "loss": 1.0513, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 1.6738949124270226, | |
| "grad_norm": 0.3044802192543269, | |
| "learning_rate": 4.84372731161588e-06, | |
| "loss": 1.0034, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 1.6805671392827355, | |
| "grad_norm": 0.3676369886348423, | |
| "learning_rate": 4.804677035205903e-06, | |
| "loss": 1.0679, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 1.6872393661384488, | |
| "grad_norm": 0.2821236895464674, | |
| "learning_rate": 4.765638687650299e-06, | |
| "loss": 0.9711, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 1.6939115929941617, | |
| "grad_norm": 0.32097634201487724, | |
| "learning_rate": 4.726614653117071e-06, | |
| "loss": 1.0368, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 1.700583819849875, | |
| "grad_norm": 0.30833620328617434, | |
| "learning_rate": 4.687607314900087e-06, | |
| "loss": 0.9419, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 1.707256046705588, | |
| "grad_norm": 0.3161695903169762, | |
| "learning_rate": 4.6486190552735375e-06, | |
| "loss": 0.947, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 1.713928273561301, | |
| "grad_norm": 0.3222992290291398, | |
| "learning_rate": 4.6096522553464265e-06, | |
| "loss": 0.9391, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 1.7206005004170142, | |
| "grad_norm": 0.27889539086250337, | |
| "learning_rate": 4.57070929491717e-06, | |
| "loss": 1.0359, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 1.7272727272727273, | |
| "grad_norm": 0.2916579428323436, | |
| "learning_rate": 4.531792552328247e-06, | |
| "loss": 1.0094, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 1.7339449541284404, | |
| "grad_norm": 0.24997244028612386, | |
| "learning_rate": 4.492904404320942e-06, | |
| "loss": 0.7959, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 1.7406171809841533, | |
| "grad_norm": 0.29868404836632323, | |
| "learning_rate": 4.454047225890204e-06, | |
| "loss": 1.0109, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 1.7472894078398666, | |
| "grad_norm": 0.2969807612243251, | |
| "learning_rate": 4.415223390139588e-06, | |
| "loss": 0.9875, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 1.7539616346955795, | |
| "grad_norm": 0.3604232218198932, | |
| "learning_rate": 4.3764352681363365e-06, | |
| "loss": 0.989, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 1.7606338615512929, | |
| "grad_norm": 0.36824762919465986, | |
| "learning_rate": 4.337685228766561e-06, | |
| "loss": 1.0049, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 1.7673060884070058, | |
| "grad_norm": 0.3500725820945927, | |
| "learning_rate": 4.2989756385905715e-06, | |
| "loss": 0.9718, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 1.773978315262719, | |
| "grad_norm": 0.2834505359886034, | |
| "learning_rate": 4.260308861698351e-06, | |
| "loss": 1.0465, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 1.780650542118432, | |
| "grad_norm": 0.3576818688058849, | |
| "learning_rate": 4.221687259565166e-06, | |
| "loss": 1.0372, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 1.787322768974145, | |
| "grad_norm": 0.31757170580776606, | |
| "learning_rate": 4.183113190907349e-06, | |
| "loss": 0.9565, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 1.7939949958298582, | |
| "grad_norm": 0.3231504978728471, | |
| "learning_rate": 4.144589011538251e-06, | |
| "loss": 0.9622, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 1.8006672226855713, | |
| "grad_norm": 0.30845575748951204, | |
| "learning_rate": 4.106117074224354e-06, | |
| "loss": 1.0258, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 1.8073394495412844, | |
| "grad_norm": 0.30421622601970244, | |
| "learning_rate": 4.067699728541595e-06, | |
| "loss": 1.0356, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 1.8140116763969973, | |
| "grad_norm": 0.2920513251222264, | |
| "learning_rate": 4.029339320731862e-06, | |
| "loss": 0.9939, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 1.8206839032527107, | |
| "grad_norm": 0.27807604400588937, | |
| "learning_rate": 3.99103819355971e-06, | |
| "loss": 0.9878, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 1.8273561301084236, | |
| "grad_norm": 0.3266996176440113, | |
| "learning_rate": 3.952798686169279e-06, | |
| "loss": 1.0041, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 1.834028356964137, | |
| "grad_norm": 0.3053490567275808, | |
| "learning_rate": 3.914623133941432e-06, | |
| "loss": 0.9872, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 1.8407005838198498, | |
| "grad_norm": 0.31477790189099164, | |
| "learning_rate": 3.876513868351142e-06, | |
| "loss": 0.9359, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 1.847372810675563, | |
| "grad_norm": 0.27149053344889523, | |
| "learning_rate": 3.838473216825085e-06, | |
| "loss": 1.0067, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 1.854045037531276, | |
| "grad_norm": 0.3204843702460214, | |
| "learning_rate": 3.800503502599511e-06, | |
| "loss": 0.9585, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 1.8607172643869891, | |
| "grad_norm": 0.3498778905670181, | |
| "learning_rate": 3.762607044578357e-06, | |
| "loss": 0.9697, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 1.8673894912427023, | |
| "grad_norm": 0.3180400555344581, | |
| "learning_rate": 3.7247861571916183e-06, | |
| "loss": 1.0226, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 1.8740617180984154, | |
| "grad_norm": 0.3480360514374549, | |
| "learning_rate": 3.6870431502540096e-06, | |
| "loss": 1.0358, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 1.8807339449541285, | |
| "grad_norm": 0.3187430568425493, | |
| "learning_rate": 3.6493803288238894e-06, | |
| "loss": 1.0073, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 1.8874061718098414, | |
| "grad_norm": 0.3464506380317717, | |
| "learning_rate": 3.611799993062497e-06, | |
| "loss": 0.943, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 1.8940783986655547, | |
| "grad_norm": 0.2880834965055193, | |
| "learning_rate": 3.5743044380934655e-06, | |
| "loss": 0.9721, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 1.9007506255212676, | |
| "grad_norm": 0.3210835495717625, | |
| "learning_rate": 3.536895953862652e-06, | |
| "loss": 1.0209, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 1.907422852376981, | |
| "grad_norm": 0.3047811453432249, | |
| "learning_rate": 3.4995768249982975e-06, | |
| "loss": 1.0284, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 1.9140950792326938, | |
| "grad_norm": 0.35570769296041677, | |
| "learning_rate": 3.462349330671484e-06, | |
| "loss": 0.9161, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 1.920767306088407, | |
| "grad_norm": 0.35749623557347504, | |
| "learning_rate": 3.4252157444569478e-06, | |
| "loss": 0.9152, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 1.92743953294412, | |
| "grad_norm": 0.3377788624601496, | |
| "learning_rate": 3.388178334194232e-06, | |
| "loss": 0.925, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 1.9341117597998332, | |
| "grad_norm": 0.3798661249068725, | |
| "learning_rate": 3.351239361849168e-06, | |
| "loss": 0.9846, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 1.9407839866555463, | |
| "grad_norm": 0.29999286355310334, | |
| "learning_rate": 3.314401083375748e-06, | |
| "loss": 0.9753, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 1.9474562135112594, | |
| "grad_norm": 0.35781167804812974, | |
| "learning_rate": 3.2776657485783357e-06, | |
| "loss": 0.9487, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 1.9541284403669725, | |
| "grad_norm": 0.3084397268778666, | |
| "learning_rate": 3.2410356009742784e-06, | |
| "loss": 1.0787, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 1.9608006672226854, | |
| "grad_norm": 0.317448351646333, | |
| "learning_rate": 3.2045128776568783e-06, | |
| "loss": 0.9257, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 1.9674728940783988, | |
| "grad_norm": 0.3686880962008835, | |
| "learning_rate": 3.168099809158769e-06, | |
| "loss": 0.9529, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 1.9741451209341117, | |
| "grad_norm": 0.30324127420271724, | |
| "learning_rate": 3.1317986193157023e-06, | |
| "loss": 0.9185, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 1.980817347789825, | |
| "grad_norm": 0.3439386421881959, | |
| "learning_rate": 3.095611525130716e-06, | |
| "loss": 1.1113, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 1.9874895746455379, | |
| "grad_norm": 0.31828525154875525, | |
| "learning_rate": 3.059540736638751e-06, | |
| "loss": 0.9624, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 1.994161801501251, | |
| "grad_norm": 0.3528601662749728, | |
| "learning_rate": 3.023588456771674e-06, | |
| "loss": 1.025, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 2.006672226855713, | |
| "grad_norm": 1.0381811083452561, | |
| "learning_rate": 2.9877568812237325e-06, | |
| "loss": 1.8823, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 2.0133444537114262, | |
| "grad_norm": 0.28745904791303756, | |
| "learning_rate": 2.9520481983174675e-06, | |
| "loss": 1.0051, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 2.020016680567139, | |
| "grad_norm": 0.3037470770313497, | |
| "learning_rate": 2.916464588870067e-06, | |
| "loss": 1.0253, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 2.0266889074228525, | |
| "grad_norm": 0.3264286058702715, | |
| "learning_rate": 2.881008226060168e-06, | |
| "loss": 0.9722, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 2.0333611342785654, | |
| "grad_norm": 0.3168792873407282, | |
| "learning_rate": 2.8456812752951483e-06, | |
| "loss": 0.9199, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 2.0400333611342787, | |
| "grad_norm": 0.3658230155927171, | |
| "learning_rate": 2.8104858940788705e-06, | |
| "loss": 0.9747, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 2.0467055879899916, | |
| "grad_norm": 0.3196066484806665, | |
| "learning_rate": 2.7754242318799174e-06, | |
| "loss": 0.9597, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 2.053377814845705, | |
| "grad_norm": 0.36214258330419746, | |
| "learning_rate": 2.740498430000332e-06, | |
| "loss": 0.9637, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 2.060050041701418, | |
| "grad_norm": 0.32144023431445257, | |
| "learning_rate": 2.7057106214448216e-06, | |
| "loss": 1.0071, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 2.066722268557131, | |
| "grad_norm": 0.31788167947698615, | |
| "learning_rate": 2.671062930790511e-06, | |
| "loss": 0.9324, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 2.073394495412844, | |
| "grad_norm": 0.32411973935094873, | |
| "learning_rate": 2.636557474057173e-06, | |
| "loss": 0.9213, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 2.080066722268557, | |
| "grad_norm": 0.3061895381140455, | |
| "learning_rate": 2.6021963585780106e-06, | |
| "loss": 0.9829, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 2.0867389491242703, | |
| "grad_norm": 0.2829970996541261, | |
| "learning_rate": 2.567981682870946e-06, | |
| "loss": 1.0112, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 2.093411175979983, | |
| "grad_norm": 0.3251773046313095, | |
| "learning_rate": 2.533915536510464e-06, | |
| "loss": 0.962, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 2.1000834028356965, | |
| "grad_norm": 0.3117638913905796, | |
| "learning_rate": 2.5000000000000015e-06, | |
| "loss": 1.0286, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 2.1067556296914094, | |
| "grad_norm": 0.3055554951200164, | |
| "learning_rate": 2.466237144644874e-06, | |
| "loss": 0.9782, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 2.1134278565471227, | |
| "grad_norm": 0.3093110501459989, | |
| "learning_rate": 2.4326290324257896e-06, | |
| "loss": 1.0203, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 2.1201000834028356, | |
| "grad_norm": 0.348167844578145, | |
| "learning_rate": 2.3991777158729102e-06, | |
| "loss": 0.9729, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 2.126772310258549, | |
| "grad_norm": 0.35896036414914057, | |
| "learning_rate": 2.3658852379404973e-06, | |
| "loss": 0.9687, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 2.133444537114262, | |
| "grad_norm": 0.33836897224758, | |
| "learning_rate": 2.3327536318821496e-06, | |
| "loss": 0.979, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 2.140116763969975, | |
| "grad_norm": 0.37686877855977563, | |
| "learning_rate": 2.299784921126622e-06, | |
| "loss": 0.9513, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 2.146788990825688, | |
| "grad_norm": 0.355801020401147, | |
| "learning_rate": 2.2669811191542547e-06, | |
| "loss": 0.9859, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 2.153461217681401, | |
| "grad_norm": 0.3010215082303293, | |
| "learning_rate": 2.234344229374003e-06, | |
| "loss": 0.9951, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 2.1601334445371143, | |
| "grad_norm": 0.32911989661603686, | |
| "learning_rate": 2.20187624500108e-06, | |
| "loss": 0.9708, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 2.166805671392827, | |
| "grad_norm": 0.3623670709775593, | |
| "learning_rate": 2.1695791489352346e-06, | |
| "loss": 0.9734, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 2.1734778982485405, | |
| "grad_norm": 0.31909897986159624, | |
| "learning_rate": 2.1374549136396417e-06, | |
| "loss": 0.9517, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 2.1801501251042534, | |
| "grad_norm": 0.31749439512516914, | |
| "learning_rate": 2.1055055010204427e-06, | |
| "loss": 0.9848, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 2.1868223519599668, | |
| "grad_norm": 0.29264160902990144, | |
| "learning_rate": 2.073732862306935e-06, | |
| "loss": 1.0053, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 2.1934945788156797, | |
| "grad_norm": 0.3668210964835384, | |
| "learning_rate": 2.042138937932388e-06, | |
| "loss": 1.0157, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 2.200166805671393, | |
| "grad_norm": 0.3133785285507638, | |
| "learning_rate": 2.0107256574155564e-06, | |
| "loss": 0.9808, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 2.206839032527106, | |
| "grad_norm": 0.3124366924714227, | |
| "learning_rate": 1.979494939242822e-06, | |
| "loss": 0.979, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 2.213511259382819, | |
| "grad_norm": 0.29545310894796867, | |
| "learning_rate": 1.9484486907510405e-06, | |
| "loss": 0.997, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 2.220183486238532, | |
| "grad_norm": 0.32367677669293343, | |
| "learning_rate": 1.917588808011045e-06, | |
| "loss": 0.8975, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 2.226855713094245, | |
| "grad_norm": 0.3299013625375276, | |
| "learning_rate": 1.8869171757118554e-06, | |
| "loss": 1.0025, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 2.2335279399499584, | |
| "grad_norm": 0.395626306344832, | |
| "learning_rate": 1.856435667045577e-06, | |
| "loss": 1.0349, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 2.2402001668056712, | |
| "grad_norm": 0.3255623940901644, | |
| "learning_rate": 1.8261461435929895e-06, | |
| "loss": 0.9994, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 2.2468723936613846, | |
| "grad_norm": 0.28569255358940876, | |
| "learning_rate": 1.796050455209869e-06, | |
| "loss": 0.9491, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 2.2535446205170975, | |
| "grad_norm": 0.29893680579074805, | |
| "learning_rate": 1.7661504399140066e-06, | |
| "loss": 0.979, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 2.260216847372811, | |
| "grad_norm": 0.3459559430460262, | |
| "learning_rate": 1.7364479237729526e-06, | |
| "loss": 0.9042, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 2.2668890742285237, | |
| "grad_norm": 0.3113640341471141, | |
| "learning_rate": 1.7069447207924994e-06, | |
| "loss": 0.9476, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 2.273561301084237, | |
| "grad_norm": 0.29758501306847707, | |
| "learning_rate": 1.677642632805892e-06, | |
| "loss": 0.9388, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 2.28023352793995, | |
| "grad_norm": 0.31704153540027663, | |
| "learning_rate": 1.6485434493637915e-06, | |
| "loss": 0.9918, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 2.2869057547956633, | |
| "grad_norm": 0.3475571370836834, | |
| "learning_rate": 1.6196489476249777e-06, | |
| "loss": 1.0017, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 2.293577981651376, | |
| "grad_norm": 0.3114674457835494, | |
| "learning_rate": 1.5909608922478108e-06, | |
| "loss": 1.0053, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 2.300250208507089, | |
| "grad_norm": 0.30298813780086953, | |
| "learning_rate": 1.5624810352824709e-06, | |
| "loss": 0.9641, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 2.3069224353628024, | |
| "grad_norm": 0.35762529414587724, | |
| "learning_rate": 1.5342111160639412e-06, | |
| "loss": 0.9616, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 2.3135946622185153, | |
| "grad_norm": 0.37005906694872853, | |
| "learning_rate": 1.5061528611057917e-06, | |
| "loss": 1.0249, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 2.3202668890742286, | |
| "grad_norm": 0.30759447735274864, | |
| "learning_rate": 1.4783079839947396e-06, | |
| "loss": 0.9851, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 2.3269391159299415, | |
| "grad_norm": 0.3014932344261013, | |
| "learning_rate": 1.4506781852859836e-06, | |
| "loss": 1.0052, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 2.333611342785655, | |
| "grad_norm": 0.31167411900209185, | |
| "learning_rate": 1.4232651523993635e-06, | |
| "loss": 0.956, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 2.3402835696413677, | |
| "grad_norm": 0.32652379262644254, | |
| "learning_rate": 1.3960705595162876e-06, | |
| "loss": 0.9825, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 2.346955796497081, | |
| "grad_norm": 0.3209030562572718, | |
| "learning_rate": 1.369096067477505e-06, | |
| "loss": 1.0177, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 2.353628023352794, | |
| "grad_norm": 0.3562229549612574, | |
| "learning_rate": 1.3423433236816563e-06, | |
| "loss": 0.9955, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 2.360300250208507, | |
| "grad_norm": 0.31080353838712754, | |
| "learning_rate": 1.3158139619846734e-06, | |
| "loss": 0.993, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 2.36697247706422, | |
| "grad_norm": 0.32276338892191286, | |
| "learning_rate": 1.289509602599996e-06, | |
| "loss": 0.9847, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 2.373644703919933, | |
| "grad_norm": 0.3162201171120057, | |
| "learning_rate": 1.2634318519996148e-06, | |
| "loss": 0.9645, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 2.3803169307756464, | |
| "grad_norm": 0.31907787951020794, | |
| "learning_rate": 1.2375823028159667e-06, | |
| "loss": 0.9949, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 2.3869891576313593, | |
| "grad_norm": 0.330999449587722, | |
| "learning_rate": 1.2119625337446673e-06, | |
| "loss": 0.9617, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 2.3936613844870727, | |
| "grad_norm": 0.37560901242187644, | |
| "learning_rate": 1.186574109448091e-06, | |
| "loss": 0.965, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 2.4003336113427856, | |
| "grad_norm": 0.3032851792093919, | |
| "learning_rate": 1.16141858045982e-06, | |
| "loss": 0.969, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 2.407005838198499, | |
| "grad_norm": 0.3120802157192034, | |
| "learning_rate": 1.1364974830899438e-06, | |
| "loss": 0.9821, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 2.413678065054212, | |
| "grad_norm": 0.34540220887991835, | |
| "learning_rate": 1.1118123393312397e-06, | |
| "loss": 0.9535, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 2.420350291909925, | |
| "grad_norm": 0.29306217474195717, | |
| "learning_rate": 1.0873646567662165e-06, | |
| "loss": 0.9548, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 2.427022518765638, | |
| "grad_norm": 0.3275471084256797, | |
| "learning_rate": 1.0631559284750398e-06, | |
| "loss": 0.9208, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 2.4336947456213514, | |
| "grad_norm": 0.3237098002556768, | |
| "learning_rate": 1.0391876329443534e-06, | |
| "loss": 0.9896, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 2.4403669724770642, | |
| "grad_norm": 0.31683152915795526, | |
| "learning_rate": 1.0154612339769777e-06, | |
| "loss": 0.9515, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 2.447039199332777, | |
| "grad_norm": 0.31260220662008054, | |
| "learning_rate": 9.919781806025136e-07, | |
| "loss": 0.9284, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 2.4537114261884905, | |
| "grad_norm": 0.3196141247915384, | |
| "learning_rate": 9.687399069888515e-07, | |
| "loss": 1.0529, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 2.4603836530442034, | |
| "grad_norm": 0.3390686830414771, | |
| "learning_rate": 9.457478323545749e-07, | |
| "loss": 0.906, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 2.4670558798999167, | |
| "grad_norm": 0.2976435803727035, | |
| "learning_rate": 9.23003360882293e-07, | |
| "loss": 0.9914, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 2.4737281067556296, | |
| "grad_norm": 0.3256745610569351, | |
| "learning_rate": 9.005078816328772e-07, | |
| "loss": 0.9923, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 2.480400333611343, | |
| "grad_norm": 0.3105480823035307, | |
| "learning_rate": 8.782627684606332e-07, | |
| "loss": 0.9996, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 2.487072560467056, | |
| "grad_norm": 0.33299836342381517, | |
| "learning_rate": 8.562693799293931e-07, | |
| "loss": 0.96, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 2.493744787322769, | |
| "grad_norm": 0.3114815813330825, | |
| "learning_rate": 8.345290592295429e-07, | |
| "loss": 0.9143, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 2.500417014178482, | |
| "grad_norm": 0.28937151523319904, | |
| "learning_rate": 8.130431340959982e-07, | |
| "loss": 0.9746, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 2.507089241034195, | |
| "grad_norm": 0.32619060975827346, | |
| "learning_rate": 7.918129167271055e-07, | |
| "loss": 0.9561, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 2.5137614678899083, | |
| "grad_norm": 0.31360364241751837, | |
| "learning_rate": 7.708397037045129e-07, | |
| "loss": 0.8092, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 2.5204336947456216, | |
| "grad_norm": 0.30558210448525414, | |
| "learning_rate": 7.50124775913979e-07, | |
| "loss": 0.9529, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 2.5271059216013345, | |
| "grad_norm": 0.32199398622868297, | |
| "learning_rate": 7.296693984671465e-07, | |
| "loss": 0.9798, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 2.5337781484570474, | |
| "grad_norm": 0.32592296410779964, | |
| "learning_rate": 7.094748206242797e-07, | |
| "loss": 1.0314, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 2.5404503753127607, | |
| "grad_norm": 0.3464187438731346, | |
| "learning_rate": 6.895422757179682e-07, | |
| "loss": 0.9811, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 2.5471226021684736, | |
| "grad_norm": 0.36090219930856193, | |
| "learning_rate": 6.698729810778065e-07, | |
| "loss": 0.9489, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 2.553794829024187, | |
| "grad_norm": 0.31856176752128185, | |
| "learning_rate": 6.50468137956049e-07, | |
| "loss": 0.9542, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 2.5604670558799, | |
| "grad_norm": 0.31996625435359055, | |
| "learning_rate": 6.313289314542392e-07, | |
| "loss": 0.9717, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 2.5671392827356128, | |
| "grad_norm": 0.3486660778765644, | |
| "learning_rate": 6.12456530450844e-07, | |
| "loss": 1.047, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 2.573811509591326, | |
| "grad_norm": 0.40268147548442224, | |
| "learning_rate": 5.938520875298587e-07, | |
| "loss": 1.0224, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 2.5804837364470394, | |
| "grad_norm": 0.3676526483372923, | |
| "learning_rate": 5.755167389104166e-07, | |
| "loss": 0.9871, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 2.5871559633027523, | |
| "grad_norm": 0.29607309034948626, | |
| "learning_rate": 5.574516043774059e-07, | |
| "loss": 0.9608, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 2.593828190158465, | |
| "grad_norm": 0.3351255483218255, | |
| "learning_rate": 5.396577872130676e-07, | |
| "loss": 0.9818, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 2.6005004170141786, | |
| "grad_norm": 0.3153076169997922, | |
| "learning_rate": 5.221363741296298e-07, | |
| "loss": 0.9962, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 2.6071726438698914, | |
| "grad_norm": 0.3816251930803852, | |
| "learning_rate": 5.048884352029271e-07, | |
| "loss": 0.9757, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 2.613844870725605, | |
| "grad_norm": 0.31190657046064496, | |
| "learning_rate": 4.879150238070585e-07, | |
| "loss": 0.9948, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 2.6205170975813177, | |
| "grad_norm": 0.38637780955296236, | |
| "learning_rate": 4.712171765500484e-07, | |
| "loss": 0.9479, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 2.627189324437031, | |
| "grad_norm": 0.3428247774477413, | |
| "learning_rate": 4.5479591321053895e-07, | |
| "loss": 0.9882, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 2.633861551292744, | |
| "grad_norm": 0.3418954616257825, | |
| "learning_rate": 4.386522366755169e-07, | |
| "loss": 0.9817, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 2.6405337781484572, | |
| "grad_norm": 0.30360358878330007, | |
| "learning_rate": 4.2278713287905335e-07, | |
| "loss": 0.9702, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 2.64720600500417, | |
| "grad_norm": 0.3262516329197568, | |
| "learning_rate": 4.072015707421006e-07, | |
| "loss": 0.9566, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 2.653878231859883, | |
| "grad_norm": 0.29298754809339406, | |
| "learning_rate": 3.918965021133131e-07, | |
| "loss": 0.9991, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 2.6605504587155964, | |
| "grad_norm": 0.3523770371797201, | |
| "learning_rate": 3.7687286171091355e-07, | |
| "loss": 0.9052, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 2.6672226855713093, | |
| "grad_norm": 0.3286065151191008, | |
| "learning_rate": 3.621315670656117e-07, | |
| "loss": 1.0313, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 2.6738949124270226, | |
| "grad_norm": 0.3327745593564552, | |
| "learning_rate": 3.4767351846456744e-07, | |
| "loss": 0.9391, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 2.6805671392827355, | |
| "grad_norm": 0.33904862886378573, | |
| "learning_rate": 3.3349959889640516e-07, | |
| "loss": 0.9358, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 2.687239366138449, | |
| "grad_norm": 0.327976050954257, | |
| "learning_rate": 3.196106739972926e-07, | |
| "loss": 1.0326, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 2.6939115929941617, | |
| "grad_norm": 0.3193292187888942, | |
| "learning_rate": 3.0600759199806815e-07, | |
| "loss": 0.9541, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 2.700583819849875, | |
| "grad_norm": 0.3774921796009617, | |
| "learning_rate": 2.9269118367244385e-07, | |
| "loss": 0.9841, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 2.707256046705588, | |
| "grad_norm": 0.31482273028522967, | |
| "learning_rate": 2.7966226228626156e-07, | |
| "loss": 0.9702, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 2.713928273561301, | |
| "grad_norm": 0.2902681756785998, | |
| "learning_rate": 2.669216235478295e-07, | |
| "loss": 0.8949, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 2.720600500417014, | |
| "grad_norm": 0.3358268040426097, | |
| "learning_rate": 2.544700455593252e-07, | |
| "loss": 0.9286, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 2.7272727272727275, | |
| "grad_norm": 0.3223051627272287, | |
| "learning_rate": 2.4230828876927293e-07, | |
| "loss": 0.9739, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 2.7339449541284404, | |
| "grad_norm": 0.31524135263431996, | |
| "learning_rate": 2.3043709592610486e-07, | |
| "loss": 0.9046, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 2.7406171809841533, | |
| "grad_norm": 0.30097063718894235, | |
| "learning_rate": 2.1885719203279587e-07, | |
| "loss": 0.9438, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 2.7472894078398666, | |
| "grad_norm": 0.35005277308808397, | |
| "learning_rate": 2.0756928430258838e-07, | |
| "loss": 0.986, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 2.7539616346955795, | |
| "grad_norm": 0.28128342646548465, | |
| "learning_rate": 1.9657406211579966e-07, | |
| "loss": 0.9488, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 2.760633861551293, | |
| "grad_norm": 0.30197051093171356, | |
| "learning_rate": 1.8587219697771942e-07, | |
| "loss": 0.9738, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 2.7673060884070058, | |
| "grad_norm": 0.31320873231724333, | |
| "learning_rate": 1.7546434247760147e-07, | |
| "loss": 0.9077, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 2.773978315262719, | |
| "grad_norm": 0.3281071984763327, | |
| "learning_rate": 1.6535113424874683e-07, | |
| "loss": 0.9133, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 2.780650542118432, | |
| "grad_norm": 0.34179101949267504, | |
| "learning_rate": 1.555331899296808e-07, | |
| "loss": 1.0089, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 2.7873227689741453, | |
| "grad_norm": 0.31629478659692434, | |
| "learning_rate": 1.460111091264377e-07, | |
| "loss": 0.9674, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 2.793994995829858, | |
| "grad_norm": 0.3248237469597133, | |
| "learning_rate": 1.3678547337593494e-07, | |
| "loss": 1.0218, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 2.800667222685571, | |
| "grad_norm": 0.31720738943716026, | |
| "learning_rate": 1.2785684611046345e-07, | |
| "loss": 0.9407, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 2.8073394495412844, | |
| "grad_norm": 0.2767770769664715, | |
| "learning_rate": 1.1922577262327374e-07, | |
| "loss": 0.9387, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 2.8140116763969973, | |
| "grad_norm": 0.34714214699914475, | |
| "learning_rate": 1.1089278003527438e-07, | |
| "loss": 1.0192, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 2.8206839032527107, | |
| "grad_norm": 0.29122863365999424, | |
| "learning_rate": 1.0285837726283999e-07, | |
| "loss": 0.9915, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 2.8273561301084236, | |
| "grad_norm": 0.35153123474849546, | |
| "learning_rate": 9.512305498672936e-08, | |
| "loss": 1.0139, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 2.834028356964137, | |
| "grad_norm": 0.3686890788414913, | |
| "learning_rate": 8.768728562211948e-08, | |
| "loss": 0.9661, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 2.84070058381985, | |
| "grad_norm": 0.3017566191381606, | |
| "learning_rate": 8.055152328975357e-08, | |
| "loss": 0.9603, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 2.847372810675563, | |
| "grad_norm": 0.37821440247787674, | |
| "learning_rate": 7.371620378820555e-08, | |
| "loss": 1.0185, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 2.854045037531276, | |
| "grad_norm": 0.31177981740624416, | |
| "learning_rate": 6.718174456726789e-08, | |
| "loss": 0.9297, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 2.860717264386989, | |
| "grad_norm": 0.31802350259904644, | |
| "learning_rate": 6.094854470245326e-08, | |
| "loss": 0.9828, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 2.8673894912427023, | |
| "grad_norm": 0.3344812561095061, | |
| "learning_rate": 5.501698487062446e-08, | |
| "loss": 0.9445, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 2.8740617180984156, | |
| "grad_norm": 0.31576512427963244, | |
| "learning_rate": 4.9387427326745287e-08, | |
| "loss": 0.9475, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 2.8807339449541285, | |
| "grad_norm": 0.3157169459636376, | |
| "learning_rate": 4.4060215881755466e-08, | |
| "loss": 0.9258, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 2.8874061718098414, | |
| "grad_norm": 0.3121956356964619, | |
| "learning_rate": 3.903567588157353e-08, | |
| "loss": 0.9958, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 2.8940783986655547, | |
| "grad_norm": 0.34863887849526165, | |
| "learning_rate": 3.431411418722941e-08, | |
| "loss": 0.9645, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 2.9007506255212676, | |
| "grad_norm": 0.35248694493395094, | |
| "learning_rate": 2.989581915611994e-08, | |
| "loss": 0.9588, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 2.907422852376981, | |
| "grad_norm": 0.3201864823159724, | |
| "learning_rate": 2.5781060624401888e-08, | |
| "loss": 0.9367, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 2.914095079232694, | |
| "grad_norm": 0.30304305615722416, | |
| "learning_rate": 2.1970089890509527e-08, | |
| "loss": 0.9279, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 2.9207673060884067, | |
| "grad_norm": 0.32097833547711685, | |
| "learning_rate": 1.8463139699808618e-08, | |
| "loss": 0.9111, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 2.92743953294412, | |
| "grad_norm": 0.3304128602909116, | |
| "learning_rate": 1.5260424230382763e-08, | |
| "loss": 0.9618, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 2.9341117597998334, | |
| "grad_norm": 0.3149926167057802, | |
| "learning_rate": 1.2362139079949431e-08, | |
| "loss": 0.9075, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 2.9407839866555463, | |
| "grad_norm": 0.34302106973635493, | |
| "learning_rate": 9.768461253920614e-09, | |
| "loss": 0.9448, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 2.947456213511259, | |
| "grad_norm": 0.3856805298465123, | |
| "learning_rate": 7.479549154585376e-09, | |
| "loss": 0.9606, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 2.9541284403669725, | |
| "grad_norm": 0.31010665321437775, | |
| "learning_rate": 5.495542571443135e-09, | |
| "loss": 0.9413, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 2.9608006672226854, | |
| "grad_norm": 0.3193071342884102, | |
| "learning_rate": 3.816562672658841e-09, | |
| "loss": 0.9947, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 2.9674728940783988, | |
| "grad_norm": 0.36599714891183205, | |
| "learning_rate": 2.442711997670544e-09, | |
| "loss": 1.0042, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 2.9741451209341117, | |
| "grad_norm": 0.33943389099897936, | |
| "learning_rate": 1.3740744509205263e-09, | |
| "loss": 1.0029, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 2.980817347789825, | |
| "grad_norm": 0.3000657499209369, | |
| "learning_rate": 6.107152967349539e-10, | |
| "loss": 0.9445, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 2.987489574645538, | |
| "grad_norm": 0.3424593362346738, | |
| "learning_rate": 1.526811553370644e-10, | |
| "loss": 0.9586, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 2.994161801501251, | |
| "grad_norm": 0.325670824774226, | |
| "learning_rate": 0.0, | |
| "loss": 0.9889, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 2.994161801501251, | |
| "step": 447, | |
| "total_flos": 1153541636358144.0, | |
| "train_loss": 1.1157796034876932, | |
| "train_runtime": 7535.4115, | |
| "train_samples_per_second": 2.864, | |
| "train_steps_per_second": 0.059 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 447, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 30, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1153541636358144.0, | |
| "train_batch_size": 2, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |