| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.9994810586403736, | |
| "eval_steps": 500, | |
| "global_step": 963, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0, | |
| "grad_norm": 0.1219773218035698, | |
| "learning_rate": 1.0000000000000002e-06, | |
| "loss": 1.3759, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "grad_norm": 0.1272401511669159, | |
| "learning_rate": 2.0000000000000003e-06, | |
| "loss": 1.3538, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "grad_norm": 0.13286344707012177, | |
| "learning_rate": 3e-06, | |
| "loss": 1.3886, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.0, | |
| "grad_norm": 0.1291491985321045, | |
| "learning_rate": 4.000000000000001e-06, | |
| "loss": 1.3143, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 0.12544086575508118, | |
| "learning_rate": 5e-06, | |
| "loss": 1.317, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 0.15034854412078857, | |
| "learning_rate": 6e-06, | |
| "loss": 1.3537, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 0.14342203736305237, | |
| "learning_rate": 7.000000000000001e-06, | |
| "loss": 1.3083, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 0.1392456442117691, | |
| "learning_rate": 8.000000000000001e-06, | |
| "loss": 1.3501, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 0.13647185266017914, | |
| "learning_rate": 9e-06, | |
| "loss": 1.1859, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 0.12714740633964539, | |
| "learning_rate": 1e-05, | |
| "loss": 1.3746, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 0.12918396294116974, | |
| "learning_rate": 1.1000000000000001e-05, | |
| "loss": 1.3672, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 0.2055857628583908, | |
| "learning_rate": 1.2e-05, | |
| "loss": 1.3692, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 0.1577647179365158, | |
| "learning_rate": 1.3000000000000001e-05, | |
| "loss": 1.2316, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 0.1306232064962387, | |
| "learning_rate": 1.4000000000000001e-05, | |
| "loss": 1.2453, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 0.1246161013841629, | |
| "learning_rate": 1.5e-05, | |
| "loss": 1.3613, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 0.1408628672361374, | |
| "learning_rate": 1.6000000000000003e-05, | |
| "loss": 1.3905, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 0.1561778485774994, | |
| "learning_rate": 1.7000000000000003e-05, | |
| "loss": 1.2617, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 0.13680391013622284, | |
| "learning_rate": 1.8e-05, | |
| "loss": 1.3604, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 0.12442106008529663, | |
| "learning_rate": 1.9e-05, | |
| "loss": 1.3812, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 0.13730080425739288, | |
| "learning_rate": 2e-05, | |
| "loss": 1.2252, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 0.11723917722702026, | |
| "learning_rate": 2.1e-05, | |
| "loss": 1.1512, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 0.13034303486347198, | |
| "learning_rate": 2.2000000000000003e-05, | |
| "loss": 1.2698, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 0.15671290457248688, | |
| "learning_rate": 2.3000000000000003e-05, | |
| "loss": 1.2823, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 0.11263088881969452, | |
| "learning_rate": 2.4e-05, | |
| "loss": 1.375, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 0.13272984325885773, | |
| "learning_rate": 2.5e-05, | |
| "loss": 1.2721, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 0.13560321927070618, | |
| "learning_rate": 2.6000000000000002e-05, | |
| "loss": 1.3223, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 0.13235026597976685, | |
| "learning_rate": 2.7000000000000002e-05, | |
| "loss": 1.2181, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 0.12842555344104767, | |
| "learning_rate": 2.8000000000000003e-05, | |
| "loss": 1.3783, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 0.13284699618816376, | |
| "learning_rate": 2.9e-05, | |
| "loss": 1.3668, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 0.12673886120319366, | |
| "learning_rate": 3e-05, | |
| "loss": 1.3657, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 0.1453860104084015, | |
| "learning_rate": 3.1e-05, | |
| "loss": 1.2646, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 0.17890802025794983, | |
| "learning_rate": 3.2000000000000005e-05, | |
| "loss": 1.1986, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 0.1185929998755455, | |
| "learning_rate": 3.3e-05, | |
| "loss": 1.1962, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 0.12994253635406494, | |
| "learning_rate": 3.4000000000000007e-05, | |
| "loss": 1.2893, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 0.1479964405298233, | |
| "learning_rate": 3.5e-05, | |
| "loss": 1.1221, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 0.13566361367702484, | |
| "learning_rate": 3.6e-05, | |
| "loss": 1.2425, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 0.1559300720691681, | |
| "learning_rate": 3.7e-05, | |
| "loss": 1.2995, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 0.14187201857566833, | |
| "learning_rate": 3.8e-05, | |
| "loss": 1.3666, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 0.13634175062179565, | |
| "learning_rate": 3.9000000000000006e-05, | |
| "loss": 1.3111, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 0.1550116240978241, | |
| "learning_rate": 4e-05, | |
| "loss": 1.3729, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 0.14702185988426208, | |
| "learning_rate": 4.1e-05, | |
| "loss": 1.2447, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 0.1325961798429489, | |
| "learning_rate": 4.2e-05, | |
| "loss": 1.174, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 0.1384323686361313, | |
| "learning_rate": 4.3e-05, | |
| "loss": 1.2333, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 0.13662421703338623, | |
| "learning_rate": 4.4000000000000006e-05, | |
| "loss": 1.1815, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 0.13229218125343323, | |
| "learning_rate": 4.5e-05, | |
| "loss": 1.2125, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 0.14347675442695618, | |
| "learning_rate": 4.600000000000001e-05, | |
| "loss": 1.1498, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 0.18536897003650665, | |
| "learning_rate": 4.7e-05, | |
| "loss": 1.2404, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 0.13089048862457275, | |
| "learning_rate": 4.8e-05, | |
| "loss": 1.2904, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 0.12752604484558105, | |
| "learning_rate": 4.9e-05, | |
| "loss": 1.2702, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 0.15005581080913544, | |
| "learning_rate": 5e-05, | |
| "loss": 1.2866, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 0.16374410688877106, | |
| "learning_rate": 5.1000000000000006e-05, | |
| "loss": 1.333, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 0.1684991866350174, | |
| "learning_rate": 5.2000000000000004e-05, | |
| "loss": 1.4278, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 0.1903846263885498, | |
| "learning_rate": 5.300000000000001e-05, | |
| "loss": 1.3035, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 0.16056384146213531, | |
| "learning_rate": 5.4000000000000005e-05, | |
| "loss": 1.2323, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 0.14987166225910187, | |
| "learning_rate": 5.500000000000001e-05, | |
| "loss": 1.0263, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 0.13468973338603973, | |
| "learning_rate": 5.6000000000000006e-05, | |
| "loss": 1.2534, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 0.1467704474925995, | |
| "learning_rate": 5.6999999999999996e-05, | |
| "loss": 1.3025, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 0.17560411989688873, | |
| "learning_rate": 5.8e-05, | |
| "loss": 1.2307, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 0.1423458606004715, | |
| "learning_rate": 5.9e-05, | |
| "loss": 1.1628, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 0.12863807380199432, | |
| "learning_rate": 6e-05, | |
| "loss": 1.1174, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 0.13459883630275726, | |
| "learning_rate": 6.1e-05, | |
| "loss": 1.2223, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 0.13762715458869934, | |
| "learning_rate": 6.2e-05, | |
| "loss": 1.2393, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 0.1310354322195053, | |
| "learning_rate": 6.3e-05, | |
| "loss": 1.2487, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 0.12477163225412369, | |
| "learning_rate": 6.400000000000001e-05, | |
| "loss": 1.2632, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 0.15380260348320007, | |
| "learning_rate": 6.500000000000001e-05, | |
| "loss": 1.3762, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 0.1248747706413269, | |
| "learning_rate": 6.6e-05, | |
| "loss": 1.1336, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 0.12641173601150513, | |
| "learning_rate": 6.7e-05, | |
| "loss": 1.2674, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 0.13872478902339935, | |
| "learning_rate": 6.800000000000001e-05, | |
| "loss": 1.3014, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 0.1377810686826706, | |
| "learning_rate": 6.9e-05, | |
| "loss": 1.2816, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 0.13459351658821106, | |
| "learning_rate": 7e-05, | |
| "loss": 1.2872, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 0.13774238526821136, | |
| "learning_rate": 7.1e-05, | |
| "loss": 1.1481, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 0.1276445835828781, | |
| "learning_rate": 7.2e-05, | |
| "loss": 1.1924, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 0.12415210157632828, | |
| "learning_rate": 7.3e-05, | |
| "loss": 1.3093, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 0.11668894439935684, | |
| "learning_rate": 7.4e-05, | |
| "loss": 1.1941, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 0.16712161898612976, | |
| "learning_rate": 7.500000000000001e-05, | |
| "loss": 1.2581, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 0.1783616989850998, | |
| "learning_rate": 7.6e-05, | |
| "loss": 1.0885, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 0.1388709843158722, | |
| "learning_rate": 7.7e-05, | |
| "loss": 1.1999, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 0.13248670101165771, | |
| "learning_rate": 7.800000000000001e-05, | |
| "loss": 1.1312, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 0.8402003645896912, | |
| "learning_rate": 7.900000000000001e-05, | |
| "loss": 1.281, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 0.1560419201850891, | |
| "learning_rate": 8e-05, | |
| "loss": 1.3314, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 0.13449999690055847, | |
| "learning_rate": 8.1e-05, | |
| "loss": 1.1742, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 0.14383117854595184, | |
| "learning_rate": 8.2e-05, | |
| "loss": 1.1958, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 0.15058472752571106, | |
| "learning_rate": 8.3e-05, | |
| "loss": 1.2992, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 0.14776362478733063, | |
| "learning_rate": 8.4e-05, | |
| "loss": 1.3611, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 0.14592403173446655, | |
| "learning_rate": 8.5e-05, | |
| "loss": 1.1165, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 0.1357845813035965, | |
| "learning_rate": 8.6e-05, | |
| "loss": 1.2509, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 0.15681371092796326, | |
| "learning_rate": 8.7e-05, | |
| "loss": 1.0857, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 0.14496199786663055, | |
| "learning_rate": 8.800000000000001e-05, | |
| "loss": 1.1111, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 0.1501457244157791, | |
| "learning_rate": 8.900000000000001e-05, | |
| "loss": 1.2556, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 0.14578907191753387, | |
| "learning_rate": 9e-05, | |
| "loss": 1.1509, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 0.13890138268470764, | |
| "learning_rate": 9.1e-05, | |
| "loss": 1.1687, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 0.14948885142803192, | |
| "learning_rate": 9.200000000000001e-05, | |
| "loss": 1.0968, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 0.13117392361164093, | |
| "learning_rate": 9.300000000000001e-05, | |
| "loss": 1.3073, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 0.13922806084156036, | |
| "learning_rate": 9.4e-05, | |
| "loss": 1.2502, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 0.16392332315444946, | |
| "learning_rate": 9.5e-05, | |
| "loss": 1.3148, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 0.11756494641304016, | |
| "learning_rate": 9.6e-05, | |
| "loss": 1.3215, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 0.11909925937652588, | |
| "learning_rate": 9.7e-05, | |
| "loss": 1.2593, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 0.1371874213218689, | |
| "learning_rate": 9.8e-05, | |
| "loss": 1.1537, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 0.14141380786895752, | |
| "learning_rate": 9.900000000000001e-05, | |
| "loss": 1.221, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 0.13502179086208344, | |
| "learning_rate": 0.0001, | |
| "loss": 1.3492, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 0.15468691289424896, | |
| "learning_rate": 0.000101, | |
| "loss": 1.2592, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 0.13041648268699646, | |
| "learning_rate": 0.00010200000000000001, | |
| "loss": 1.3792, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 0.11658702045679092, | |
| "learning_rate": 0.00010300000000000001, | |
| "loss": 1.1845, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 0.13084186613559723, | |
| "learning_rate": 0.00010400000000000001, | |
| "loss": 1.1808, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 0.13180844485759735, | |
| "learning_rate": 0.000105, | |
| "loss": 1.1721, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 0.1375647485256195, | |
| "learning_rate": 0.00010600000000000002, | |
| "loss": 1.3622, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 0.11864122748374939, | |
| "learning_rate": 0.00010700000000000001, | |
| "loss": 1.238, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 0.11780542880296707, | |
| "learning_rate": 0.00010800000000000001, | |
| "loss": 1.1581, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 0.11449900269508362, | |
| "learning_rate": 0.000109, | |
| "loss": 1.3076, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 0.11585865169763565, | |
| "learning_rate": 0.00011000000000000002, | |
| "loss": 1.3553, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 0.12288644164800644, | |
| "learning_rate": 0.00011100000000000001, | |
| "loss": 1.1448, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 0.12131894379854202, | |
| "learning_rate": 0.00011200000000000001, | |
| "loss": 1.3456, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 0.13827624917030334, | |
| "learning_rate": 0.000113, | |
| "loss": 1.3183, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 0.13708753883838654, | |
| "learning_rate": 0.00011399999999999999, | |
| "loss": 1.3293, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 0.11893680691719055, | |
| "learning_rate": 0.00011499999999999999, | |
| "loss": 1.1954, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 0.12022116035223007, | |
| "learning_rate": 0.000116, | |
| "loss": 1.1822, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 0.11215274035930634, | |
| "learning_rate": 0.000117, | |
| "loss": 1.2453, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 0.11778578907251358, | |
| "learning_rate": 0.000118, | |
| "loss": 1.1939, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 0.12714551389217377, | |
| "learning_rate": 0.000119, | |
| "loss": 1.1917, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 0.11972431093454361, | |
| "learning_rate": 0.00012, | |
| "loss": 1.2769, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 0.12945646047592163, | |
| "learning_rate": 0.000121, | |
| "loss": 1.2929, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 0.11721435189247131, | |
| "learning_rate": 0.000122, | |
| "loss": 1.3606, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 0.117381751537323, | |
| "learning_rate": 0.000123, | |
| "loss": 1.3187, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 0.13110613822937012, | |
| "learning_rate": 0.000124, | |
| "loss": 1.3786, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 0.12571324408054352, | |
| "learning_rate": 0.000125, | |
| "loss": 1.0165, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 0.12401648610830307, | |
| "learning_rate": 0.000126, | |
| "loss": 1.1858, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 0.11960897594690323, | |
| "learning_rate": 0.000127, | |
| "loss": 1.2638, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 0.11195061355829239, | |
| "learning_rate": 0.00012800000000000002, | |
| "loss": 1.2469, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 0.1292635053396225, | |
| "learning_rate": 0.00012900000000000002, | |
| "loss": 1.3642, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 0.11140226572751999, | |
| "learning_rate": 0.00013000000000000002, | |
| "loss": 1.2238, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 0.10454066097736359, | |
| "learning_rate": 0.000131, | |
| "loss": 1.2418, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 0.11138935387134552, | |
| "learning_rate": 0.000132, | |
| "loss": 1.1964, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 0.12519143521785736, | |
| "learning_rate": 0.000133, | |
| "loss": 1.1889, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 0.1129814088344574, | |
| "learning_rate": 0.000134, | |
| "loss": 1.3425, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 0.11608937382698059, | |
| "learning_rate": 0.00013500000000000003, | |
| "loss": 1.0762, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 0.1410205215215683, | |
| "learning_rate": 0.00013600000000000003, | |
| "loss": 1.2502, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 0.11246364563703537, | |
| "learning_rate": 0.00013700000000000002, | |
| "loss": 1.3605, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 0.1213245838880539, | |
| "learning_rate": 0.000138, | |
| "loss": 1.2632, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 0.11144573241472244, | |
| "learning_rate": 0.000139, | |
| "loss": 1.1771, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 0.1700059324502945, | |
| "learning_rate": 0.00014, | |
| "loss": 1.3396, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 0.11950355768203735, | |
| "learning_rate": 0.000141, | |
| "loss": 1.1329, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 0.11367864161729813, | |
| "learning_rate": 0.000142, | |
| "loss": 1.3323, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 0.22367842495441437, | |
| "learning_rate": 0.000143, | |
| "loss": 1.201, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 0.11639752984046936, | |
| "learning_rate": 0.000144, | |
| "loss": 1.2507, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 0.12356352061033249, | |
| "learning_rate": 0.000145, | |
| "loss": 1.3437, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 0.10859037935733795, | |
| "learning_rate": 0.000146, | |
| "loss": 1.1677, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 0.11901260167360306, | |
| "learning_rate": 0.000147, | |
| "loss": 1.1125, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 0.1092861220240593, | |
| "learning_rate": 0.000148, | |
| "loss": 1.1939, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 0.10728005319833755, | |
| "learning_rate": 0.00014900000000000002, | |
| "loss": 1.1974, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 0.11113261431455612, | |
| "learning_rate": 0.00015000000000000001, | |
| "loss": 1.1107, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 0.14674580097198486, | |
| "learning_rate": 0.000151, | |
| "loss": 1.2339, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 0.10521555691957474, | |
| "learning_rate": 0.000152, | |
| "loss": 1.2917, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 0.12526080012321472, | |
| "learning_rate": 0.000153, | |
| "loss": 1.3247, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 0.2366744726896286, | |
| "learning_rate": 0.000154, | |
| "loss": 1.3637, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 0.11178407818078995, | |
| "learning_rate": 0.000155, | |
| "loss": 1.2649, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 0.10156022757291794, | |
| "learning_rate": 0.00015600000000000002, | |
| "loss": 1.1637, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 0.09755547344684601, | |
| "learning_rate": 0.00015700000000000002, | |
| "loss": 1.241, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 0.11344686150550842, | |
| "learning_rate": 0.00015800000000000002, | |
| "loss": 1.2208, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 0.1021842285990715, | |
| "learning_rate": 0.00015900000000000002, | |
| "loss": 1.1936, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 0.11096473783254623, | |
| "learning_rate": 0.00016, | |
| "loss": 1.2908, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 0.11365038901567459, | |
| "learning_rate": 0.000161, | |
| "loss": 1.1754, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 0.1159452423453331, | |
| "learning_rate": 0.000162, | |
| "loss": 1.2411, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 0.1240570917725563, | |
| "learning_rate": 0.000163, | |
| "loss": 1.2655, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 0.11343025416135788, | |
| "learning_rate": 0.000164, | |
| "loss": 1.3486, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 0.10576006770133972, | |
| "learning_rate": 0.000165, | |
| "loss": 1.2787, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 0.182003453373909, | |
| "learning_rate": 0.000166, | |
| "loss": 1.2045, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 0.10536637157201767, | |
| "learning_rate": 0.000167, | |
| "loss": 1.1388, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 0.10617152601480484, | |
| "learning_rate": 0.000168, | |
| "loss": 1.2686, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 0.11899720877408981, | |
| "learning_rate": 0.00016900000000000002, | |
| "loss": 1.1732, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 0.11736088991165161, | |
| "learning_rate": 0.00017, | |
| "loss": 1.2772, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 0.10500625520944595, | |
| "learning_rate": 0.000171, | |
| "loss": 1.2193, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 0.10299772769212723, | |
| "learning_rate": 0.000172, | |
| "loss": 1.2021, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 0.10550544410943985, | |
| "learning_rate": 0.000173, | |
| "loss": 1.2393, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 0.1132335364818573, | |
| "learning_rate": 0.000174, | |
| "loss": 1.2713, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 0.11581739038228989, | |
| "learning_rate": 0.000175, | |
| "loss": 1.3282, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 0.11211756616830826, | |
| "learning_rate": 0.00017600000000000002, | |
| "loss": 1.2374, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 0.10848426073789597, | |
| "learning_rate": 0.00017700000000000002, | |
| "loss": 1.2916, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 0.11354674398899078, | |
| "learning_rate": 0.00017800000000000002, | |
| "loss": 1.1914, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 0.12108771502971649, | |
| "learning_rate": 0.00017900000000000001, | |
| "loss": 1.3693, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 0.11068333685398102, | |
| "learning_rate": 0.00018, | |
| "loss": 1.086, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 0.11752796918153763, | |
| "learning_rate": 0.000181, | |
| "loss": 1.1508, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 0.10650038719177246, | |
| "learning_rate": 0.000182, | |
| "loss": 1.204, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 0.12248033285140991, | |
| "learning_rate": 0.000183, | |
| "loss": 1.2288, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 0.10320232063531876, | |
| "learning_rate": 0.00018400000000000003, | |
| "loss": 1.1309, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 0.1450732797384262, | |
| "learning_rate": 0.00018500000000000002, | |
| "loss": 1.2696, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 0.10568010061979294, | |
| "learning_rate": 0.00018600000000000002, | |
| "loss": 1.2643, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 0.12027282267808914, | |
| "learning_rate": 0.00018700000000000002, | |
| "loss": 1.2455, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 0.09627476334571838, | |
| "learning_rate": 0.000188, | |
| "loss": 1.1999, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 0.1297677904367447, | |
| "learning_rate": 0.00018899999999999999, | |
| "loss": 1.0843, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 0.1251845359802246, | |
| "learning_rate": 0.00019, | |
| "loss": 1.2957, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 0.11355076730251312, | |
| "learning_rate": 0.000191, | |
| "loss": 1.2705, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 0.11948814988136292, | |
| "learning_rate": 0.000192, | |
| "loss": 1.2352, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 0.12006771564483643, | |
| "learning_rate": 0.000193, | |
| "loss": 1.2489, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 0.11943700909614563, | |
| "learning_rate": 0.000194, | |
| "loss": 1.0768, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 0.12115556001663208, | |
| "learning_rate": 0.000195, | |
| "loss": 1.2614, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 0.1361999660730362, | |
| "learning_rate": 0.000196, | |
| "loss": 1.1418, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 0.10292702168226242, | |
| "learning_rate": 0.00019700000000000002, | |
| "loss": 1.1649, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 0.12569741904735565, | |
| "learning_rate": 0.00019800000000000002, | |
| "loss": 1.2054, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 0.09684847295284271, | |
| "learning_rate": 0.000199, | |
| "loss": 1.0854, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 0.10697442293167114, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1895, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 0.10017522424459457, | |
| "learning_rate": 0.00019999915234318063, | |
| "loss": 1.2775, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 0.0991068184375763, | |
| "learning_rate": 0.0001999966093870929, | |
| "loss": 1.3151, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 0.10074901580810547, | |
| "learning_rate": 0.00019999237117484796, | |
| "loss": 1.2666, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 0.10549337416887283, | |
| "learning_rate": 0.00019998643777829672, | |
| "loss": 1.2483, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 0.10364524275064468, | |
| "learning_rate": 0.00019997880929802894, | |
| "loss": 1.0562, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 0.21587559580802917, | |
| "learning_rate": 0.00019996948586337125, | |
| "loss": 1.3881, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 0.10044362396001816, | |
| "learning_rate": 0.00019995846763238512, | |
| "loss": 1.2114, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 0.1037297397851944, | |
| "learning_rate": 0.0001999457547918641, | |
| "loss": 1.2611, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 0.11380340903997421, | |
| "learning_rate": 0.00019993134755733074, | |
| "loss": 1.3762, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 0.11318037658929825, | |
| "learning_rate": 0.0001999152461730328, | |
| "loss": 1.2906, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 0.1097273975610733, | |
| "learning_rate": 0.0001998974509119393, | |
| "loss": 1.0855, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 0.11358966678380966, | |
| "learning_rate": 0.00019987796207573573, | |
| "loss": 1.2061, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 0.11139972507953644, | |
| "learning_rate": 0.00019985677999481894, | |
| "loss": 1.2453, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 0.1275399923324585, | |
| "learning_rate": 0.00019983390502829166, | |
| "loss": 1.1917, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 0.1333334743976593, | |
| "learning_rate": 0.00019980933756395634, | |
| "loss": 1.2102, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 0.10892786085605621, | |
| "learning_rate": 0.00019978307801830854, | |
| "loss": 1.1177, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 0.10664794594049454, | |
| "learning_rate": 0.00019975512683652986, | |
| "loss": 1.1484, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 0.20084381103515625, | |
| "learning_rate": 0.0001997254844924806, | |
| "loss": 1.0399, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 0.11178454756736755, | |
| "learning_rate": 0.0001996941514886914, | |
| "loss": 1.3401, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 0.10734831541776657, | |
| "learning_rate": 0.00019966112835635493, | |
| "loss": 1.1964, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 0.12020622938871384, | |
| "learning_rate": 0.00019962641565531692, | |
| "loss": 1.1535, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 0.10945051908493042, | |
| "learning_rate": 0.00019959001397406646, | |
| "loss": 1.2086, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 0.11263303458690643, | |
| "learning_rate": 0.00019955192392972628, | |
| "loss": 1.1488, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 0.119475819170475, | |
| "learning_rate": 0.00019951214616804202, | |
| "loss": 1.3071, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 0.10934551060199738, | |
| "learning_rate": 0.00019947068136337158, | |
| "loss": 1.2639, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 0.11675988882780075, | |
| "learning_rate": 0.0001994275302186734, | |
| "loss": 1.142, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 0.28856417536735535, | |
| "learning_rate": 0.00019938269346549475, | |
| "loss": 1.2437, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 0.11922767013311386, | |
| "learning_rate": 0.00019933617186395917, | |
| "loss": 1.1668, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 0.12038363516330719, | |
| "learning_rate": 0.00019928796620275377, | |
| "loss": 1.0465, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 0.10866128653287888, | |
| "learning_rate": 0.00019923807729911567, | |
| "loss": 1.1417, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 0.134980708360672, | |
| "learning_rate": 0.00019918650599881827, | |
| "loss": 1.3574, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 0.11622115224599838, | |
| "learning_rate": 0.00019913325317615684, | |
| "loss": 1.2457, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 0.11134631186723709, | |
| "learning_rate": 0.0001990783197339338, | |
| "loss": 1.2194, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 0.11312945932149887, | |
| "learning_rate": 0.0001990217066034432, | |
| "loss": 1.179, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 0.10070719569921494, | |
| "learning_rate": 0.00019896341474445525, | |
| "loss": 1.299, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 0.11468905210494995, | |
| "learning_rate": 0.00019890344514519973, | |
| "loss": 1.2141, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 0.12275680154561996, | |
| "learning_rate": 0.00019884179882234944, | |
| "loss": 1.2581, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 0.1063394546508789, | |
| "learning_rate": 0.00019877847682100294, | |
| "loss": 1.3027, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 0.10849407315254211, | |
| "learning_rate": 0.00019871348021466672, | |
| "loss": 1.1914, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 0.1102900579571724, | |
| "learning_rate": 0.0001986468101052371, | |
| "loss": 1.2589, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 0.11001666635274887, | |
| "learning_rate": 0.00019857846762298155, | |
| "loss": 1.141, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 0.10907820612192154, | |
| "learning_rate": 0.0001985084539265195, | |
| "loss": 1.1139, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 0.1067519262433052, | |
| "learning_rate": 0.0001984367702028027, | |
| "loss": 1.2389, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 0.11710456013679504, | |
| "learning_rate": 0.0001983634176670951, | |
| "loss": 1.2421, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 0.10886801034212112, | |
| "learning_rate": 0.00019828839756295223, | |
| "loss": 1.1696, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 0.1404796689748764, | |
| "learning_rate": 0.00019821171116220012, | |
| "loss": 1.29, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 0.1096196323633194, | |
| "learning_rate": 0.00019813335976491385, | |
| "loss": 1.3536, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 0.11487864702939987, | |
| "learning_rate": 0.00019805334469939528, | |
| "loss": 1.1371, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 0.1555541455745697, | |
| "learning_rate": 0.00019797166732215076, | |
| "loss": 1.1812, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 0.10709488391876221, | |
| "learning_rate": 0.000197888329017868, | |
| "loss": 1.101, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 0.10238959640264511, | |
| "learning_rate": 0.00019780333119939262, | |
| "loss": 1.2363, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 0.12807561457157135, | |
| "learning_rate": 0.00019771667530770425, | |
| "loss": 1.217, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 0.10418405383825302, | |
| "learning_rate": 0.00019762836281189206, | |
| "loss": 1.36, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 0.10400766879320145, | |
| "learning_rate": 0.00019753839520912983, | |
| "loss": 1.2368, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 0.1629195362329483, | |
| "learning_rate": 0.00019744677402465052, | |
| "loss": 1.2236, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 0.11335594952106476, | |
| "learning_rate": 0.00019735350081172067, | |
| "loss": 1.2062, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 0.18272048234939575, | |
| "learning_rate": 0.00019725857715161373, | |
| "loss": 1.3138, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 0.10922593623399734, | |
| "learning_rate": 0.0001971620046535835, | |
| "loss": 1.1697, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 0.10762567073106766, | |
| "learning_rate": 0.00019706378495483665, | |
| "loss": 1.1303, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 0.11802787333726883, | |
| "learning_rate": 0.00019696391972050514, | |
| "loss": 1.2909, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 0.11715908348560333, | |
| "learning_rate": 0.00019686241064361792, | |
| "loss": 1.1456, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 0.11491627991199493, | |
| "learning_rate": 0.00019675925944507224, | |
| "loss": 1.2083, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 0.11131855100393295, | |
| "learning_rate": 0.0001966544678736044, | |
| "loss": 1.2579, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 0.12063883244991302, | |
| "learning_rate": 0.00019654803770576026, | |
| "loss": 1.3887, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 0.14915111660957336, | |
| "learning_rate": 0.0001964399707458649, | |
| "loss": 1.2351, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 0.11454611271619797, | |
| "learning_rate": 0.00019633026882599227, | |
| "loss": 1.3325, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 0.10960078984498978, | |
| "learning_rate": 0.00019621893380593396, | |
| "loss": 1.3, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 0.11405951529741287, | |
| "learning_rate": 0.0001961059675731678, | |
| "loss": 1.0641, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 0.10570972412824631, | |
| "learning_rate": 0.00019599137204282565, | |
| "loss": 1.1543, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 0.11385346204042435, | |
| "learning_rate": 0.00019587514915766124, | |
| "loss": 1.1546, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 0.10122133791446686, | |
| "learning_rate": 0.00019575730088801697, | |
| "loss": 1.1946, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 0.14278005063533783, | |
| "learning_rate": 0.00019563782923179063, | |
| "loss": 1.1853, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 0.11730815470218658, | |
| "learning_rate": 0.00019551673621440147, | |
| "loss": 1.4478, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 0.12385374307632446, | |
| "learning_rate": 0.00019539402388875598, | |
| "loss": 1.4392, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 0.10802686959505081, | |
| "learning_rate": 0.00019526969433521297, | |
| "loss": 1.3126, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 0.12256834656000137, | |
| "learning_rate": 0.00019514374966154825, | |
| "loss": 1.1855, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 0.10675772279500961, | |
| "learning_rate": 0.00019501619200291908, | |
| "loss": 1.2578, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 0.1070014238357544, | |
| "learning_rate": 0.00019488702352182786, | |
| "loss": 1.1426, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 0.11141553521156311, | |
| "learning_rate": 0.00019475624640808543, | |
| "loss": 1.1046, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 0.10741408169269562, | |
| "learning_rate": 0.00019462386287877407, | |
| "loss": 1.1991, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 0.11526710540056229, | |
| "learning_rate": 0.00019448987517820982, | |
| "loss": 1.3507, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 0.1025758609175682, | |
| "learning_rate": 0.0001943542855779044, | |
| "loss": 1.3, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 0.11231940984725952, | |
| "learning_rate": 0.0001942170963765268, | |
| "loss": 1.2408, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 0.1058710590004921, | |
| "learning_rate": 0.0001940783098998643, | |
| "loss": 1.1668, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 0.1024676039814949, | |
| "learning_rate": 0.00019393792850078292, | |
| "loss": 1.2192, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 0.10104867815971375, | |
| "learning_rate": 0.00019379595455918774, | |
| "loss": 1.2881, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 0.1211661621928215, | |
| "learning_rate": 0.00019365239048198225, | |
| "loss": 1.3614, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 0.1058136522769928, | |
| "learning_rate": 0.0001935072387030279, | |
| "loss": 1.2043, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 0.3414977490901947, | |
| "learning_rate": 0.0001933605016831026, | |
| "loss": 1.1638, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 0.09908904135227203, | |
| "learning_rate": 0.00019321218190985904, | |
| "loss": 1.0565, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 0.11039788275957108, | |
| "learning_rate": 0.00019306228189778254, | |
| "loss": 1.1324, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 0.1092277467250824, | |
| "learning_rate": 0.0001929108041881485, | |
| "loss": 1.0857, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 0.11304416507482529, | |
| "learning_rate": 0.00019275775134897918, | |
| "loss": 1.3365, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 0.1239667758345604, | |
| "learning_rate": 0.00019260312597500022, | |
| "loss": 1.2251, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 0.1069445088505745, | |
| "learning_rate": 0.00019244693068759666, | |
| "loss": 1.1466, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 0.11692642420530319, | |
| "learning_rate": 0.00019228916813476853, | |
| "loss": 0.947, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 0.11571913212537766, | |
| "learning_rate": 0.00019212984099108594, | |
| "loss": 1.1865, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 0.14679546654224396, | |
| "learning_rate": 0.00019196895195764362, | |
| "loss": 1.2747, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 0.11377564072608948, | |
| "learning_rate": 0.00019180650376201535, | |
| "loss": 1.2149, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 0.14246807992458344, | |
| "learning_rate": 0.0001916424991582075, | |
| "loss": 1.3234, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 0.11343812942504883, | |
| "learning_rate": 0.00019147694092661255, | |
| "loss": 1.2125, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 0.49394461512565613, | |
| "learning_rate": 0.00019130983187396172, | |
| "loss": 1.2545, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 0.12038710713386536, | |
| "learning_rate": 0.0001911411748332776, | |
| "loss": 1.2441, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 0.112415611743927, | |
| "learning_rate": 0.00019097097266382597, | |
| "loss": 1.1986, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 0.1328456848859787, | |
| "learning_rate": 0.0001907992282510675, | |
| "loss": 1.0427, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 0.12180749326944351, | |
| "learning_rate": 0.00019062594450660857, | |
| "loss": 1.2805, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 0.16928140819072723, | |
| "learning_rate": 0.00019045112436815215, | |
| "loss": 1.1842, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 0.15169909596443176, | |
| "learning_rate": 0.00019027477079944788, | |
| "loss": 1.1867, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 0.11884856224060059, | |
| "learning_rate": 0.0001900968867902419, | |
| "loss": 1.2883, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 0.10564276576042175, | |
| "learning_rate": 0.0001899174753562261, | |
| "loss": 1.264, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 0.12126502394676208, | |
| "learning_rate": 0.0001897365395389869, | |
| "loss": 1.2866, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 0.11135692894458771, | |
| "learning_rate": 0.00018955408240595395, | |
| "loss": 1.3224, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 0.11942031234502792, | |
| "learning_rate": 0.00018937010705034787, | |
| "loss": 1.2606, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 0.11544723808765411, | |
| "learning_rate": 0.00018918461659112802, | |
| "loss": 1.2025, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 0.12025148421525955, | |
| "learning_rate": 0.00018899761417293941, | |
| "loss": 1.4092, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 0.126939594745636, | |
| "learning_rate": 0.00018880910296605955, | |
| "loss": 1.1811, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 0.29961127042770386, | |
| "learning_rate": 0.00018861908616634465, | |
| "loss": 1.1998, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 0.12669076025485992, | |
| "learning_rate": 0.00018842756699517538, | |
| "loss": 1.1652, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 0.12721942365169525, | |
| "learning_rate": 0.00018823454869940242, | |
| "loss": 1.2145, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 0.1199335977435112, | |
| "learning_rate": 0.0001880400345512913, | |
| "loss": 1.2953, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 0.13686604797840118, | |
| "learning_rate": 0.00018784402784846683, | |
| "loss": 1.2455, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 0.1259993016719818, | |
| "learning_rate": 0.00018764653191385736, | |
| "loss": 0.9901, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 0.13265901803970337, | |
| "learning_rate": 0.00018744755009563848, | |
| "loss": 1.2294, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 0.11366987973451614, | |
| "learning_rate": 0.00018724708576717605, | |
| "loss": 1.3416, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 0.11126185208559036, | |
| "learning_rate": 0.0001870451423269692, | |
| "loss": 1.1254, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 0.11551456153392792, | |
| "learning_rate": 0.0001868417231985926, | |
| "loss": 1.2113, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 0.11155180633068085, | |
| "learning_rate": 0.00018663683183063845, | |
| "loss": 1.1587, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 0.1066342443227768, | |
| "learning_rate": 0.0001864304716966581, | |
| "loss": 1.3596, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 0.11637596786022186, | |
| "learning_rate": 0.00018622264629510299, | |
| "loss": 1.2117, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 0.10947410017251968, | |
| "learning_rate": 0.00018601335914926557, | |
| "loss": 0.9963, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 0.1826743334531784, | |
| "learning_rate": 0.0001858026138072193, | |
| "loss": 1.1548, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 0.12238079309463501, | |
| "learning_rate": 0.00018559041384175876, | |
| "loss": 1.0507, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 0.1196766272187233, | |
| "learning_rate": 0.00018537676285033887, | |
| "loss": 1.1443, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 0.1195516511797905, | |
| "learning_rate": 0.00018516166445501405, | |
| "loss": 1.083, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 0.111576147377491, | |
| "learning_rate": 0.0001849451223023767, | |
| "loss": 1.192, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 0.12735474109649658, | |
| "learning_rate": 0.0001847271400634955, | |
| "loss": 1.2692, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 0.10551782697439194, | |
| "learning_rate": 0.00018450772143385307, | |
| "loss": 1.2476, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 0.11405996978282928, | |
| "learning_rate": 0.00018428687013328336, | |
| "loss": 1.1888, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 0.13341975212097168, | |
| "learning_rate": 0.00018406458990590858, | |
| "loss": 1.2129, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 0.13987426459789276, | |
| "learning_rate": 0.00018384088452007578, | |
| "loss": 1.2114, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 0.1929798275232315, | |
| "learning_rate": 0.0001836157577682928, | |
| "loss": 1.1411, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 0.11936288326978683, | |
| "learning_rate": 0.00018338921346716425, | |
| "loss": 1.2696, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 0.10896290838718414, | |
| "learning_rate": 0.0001831612554573265, | |
| "loss": 1.2438, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 0.11335362493991852, | |
| "learning_rate": 0.00018293188760338286, | |
| "loss": 1.0288, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 0.11729809641838074, | |
| "learning_rate": 0.00018270111379383775, | |
| "loss": 1.1118, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 0.11622367799282074, | |
| "learning_rate": 0.0001824689379410311, | |
| "loss": 1.1515, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 0.11727786809206009, | |
| "learning_rate": 0.00018223536398107176, | |
| "loss": 1.2468, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 0.11702138185501099, | |
| "learning_rate": 0.00018200039587377097, | |
| "loss": 1.2647, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 0.12109760195016861, | |
| "learning_rate": 0.0001817640376025751, | |
| "loss": 1.2282, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 0.11537209153175354, | |
| "learning_rate": 0.00018152629317449813, | |
| "loss": 1.178, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 0.1228351965546608, | |
| "learning_rate": 0.00018128716662005383, | |
| "loss": 1.2099, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 0.15511314570903778, | |
| "learning_rate": 0.0001810466619931872, | |
| "loss": 1.2881, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 0.1076040267944336, | |
| "learning_rate": 0.00018080478337120605, | |
| "loss": 1.264, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 0.14106139540672302, | |
| "learning_rate": 0.00018056153485471166, | |
| "loss": 1.2821, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 0.11988777667284012, | |
| "learning_rate": 0.00018031692056752924, | |
| "loss": 1.2771, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 0.10703491419553757, | |
| "learning_rate": 0.0001800709446566382, | |
| "loss": 1.2525, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 0.11741764843463898, | |
| "learning_rate": 0.00017982361129210172, | |
| "loss": 1.2151, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 0.19867832958698273, | |
| "learning_rate": 0.00017957492466699605, | |
| "loss": 1.2559, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 0.12436340004205704, | |
| "learning_rate": 0.00017932488899733947, | |
| "loss": 1.3145, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 0.12323563545942307, | |
| "learning_rate": 0.00017907350852202077, | |
| "loss": 1.2145, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 0.126682311296463, | |
| "learning_rate": 0.00017882078750272742, | |
| "loss": 1.2805, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 0.10707955062389374, | |
| "learning_rate": 0.00017856673022387338, | |
| "loss": 1.1415, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 0.1281125843524933, | |
| "learning_rate": 0.0001783113409925263, | |
| "loss": 1.3117, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 0.11295267939567566, | |
| "learning_rate": 0.00017805462413833468, | |
| "loss": 1.1453, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 0.12421952188014984, | |
| "learning_rate": 0.00017779658401345438, | |
| "loss": 1.2293, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 0.11468549817800522, | |
| "learning_rate": 0.00017753722499247478, | |
| "loss": 1.1391, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 0.12964022159576416, | |
| "learning_rate": 0.00017727655147234476, | |
| "loss": 1.3975, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 0.11601359397172928, | |
| "learning_rate": 0.00017701456787229804, | |
| "loss": 1.2622, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 0.12835174798965454, | |
| "learning_rate": 0.00017675127863377838, | |
| "loss": 1.1703, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 0.11938933283090591, | |
| "learning_rate": 0.00017648668822036408, | |
| "loss": 1.2702, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 0.12092038989067078, | |
| "learning_rate": 0.00017622080111769257, | |
| "loss": 1.3653, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 0.1410731077194214, | |
| "learning_rate": 0.00017595362183338412, | |
| "loss": 1.2465, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 0.11769446730613708, | |
| "learning_rate": 0.00017568515489696558, | |
| "loss": 1.235, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 0.10854797065258026, | |
| "learning_rate": 0.00017541540485979359, | |
| "loss": 1.1758, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 0.10094323009252548, | |
| "learning_rate": 0.0001751443762949772, | |
| "loss": 1.0939, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 0.10836099833250046, | |
| "learning_rate": 0.00017487207379730078, | |
| "loss": 1.1618, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 0.1138240173459053, | |
| "learning_rate": 0.00017459850198314563, | |
| "loss": 1.1423, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 0.11237498372793198, | |
| "learning_rate": 0.000174323665490412, | |
| "loss": 1.0702, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 0.1078115925192833, | |
| "learning_rate": 0.00017404756897844052, | |
| "loss": 1.2523, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 0.11401957273483276, | |
| "learning_rate": 0.00017377021712793297, | |
| "loss": 1.2443, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 0.1109926849603653, | |
| "learning_rate": 0.00017349161464087312, | |
| "loss": 1.264, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 0.10931116342544556, | |
| "learning_rate": 0.00017321176624044687, | |
| "loss": 1.294, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 0.10655161738395691, | |
| "learning_rate": 0.00017293067667096239, | |
| "loss": 1.2629, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 0.11112368851900101, | |
| "learning_rate": 0.00017264835069776945, | |
| "loss": 1.25, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 0.1312362551689148, | |
| "learning_rate": 0.0001723647931071788, | |
| "loss": 1.1807, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 0.1174694374203682, | |
| "learning_rate": 0.00017208000870638093, | |
| "loss": 1.1719, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 0.11656120419502258, | |
| "learning_rate": 0.00017179400232336463, | |
| "loss": 1.4266, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 0.11490406095981598, | |
| "learning_rate": 0.00017150677880683514, | |
| "loss": 1.3041, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 0.1138778105378151, | |
| "learning_rate": 0.00017121834302613186, | |
| "loss": 1.3364, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 0.11219751089811325, | |
| "learning_rate": 0.000170928699871146, | |
| "loss": 1.3409, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 0.10369167476892471, | |
| "learning_rate": 0.0001706378542522374, | |
| "loss": 1.1775, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 0.11245618760585785, | |
| "learning_rate": 0.00017034581110015156, | |
| "loss": 1.1841, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 0.10895467549562454, | |
| "learning_rate": 0.00017005257536593577, | |
| "loss": 1.1855, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 0.10433463007211685, | |
| "learning_rate": 0.00016975815202085554, | |
| "loss": 1.2853, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 0.10486618429422379, | |
| "learning_rate": 0.00016946254605630992, | |
| "loss": 1.2029, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 0.12566353380680084, | |
| "learning_rate": 0.00016916576248374719, | |
| "loss": 1.187, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 0.11196866631507874, | |
| "learning_rate": 0.00016886780633457973, | |
| "loss": 1.1956, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 0.12952035665512085, | |
| "learning_rate": 0.00016856868266009874, | |
| "loss": 1.24, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 0.1032567098736763, | |
| "learning_rate": 0.0001682683965313887, | |
| "loss": 1.207, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 0.10673535615205765, | |
| "learning_rate": 0.00016796695303924127, | |
| "loss": 1.23, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 0.1103532612323761, | |
| "learning_rate": 0.00016766435729406913, | |
| "loss": 1.2994, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 0.11205995827913284, | |
| "learning_rate": 0.0001673606144258192, | |
| "loss": 1.1168, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 0.11950290203094482, | |
| "learning_rate": 0.00016705572958388576, | |
| "loss": 1.2152, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 0.17785818874835968, | |
| "learning_rate": 0.0001667497079370231, | |
| "loss": 1.2523, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 0.10789604485034943, | |
| "learning_rate": 0.00016644255467325794, | |
| "loss": 1.289, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 0.12129693478345871, | |
| "learning_rate": 0.00016613427499980143, | |
| "loss": 1.257, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 0.11083897203207016, | |
| "learning_rate": 0.00016582487414296097, | |
| "loss": 1.3349, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 0.09888826310634613, | |
| "learning_rate": 0.0001655143573480515, | |
| "loss": 1.1737, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 0.11880184710025787, | |
| "learning_rate": 0.00016520272987930652, | |
| "loss": 1.0338, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 0.11628952622413635, | |
| "learning_rate": 0.00016488999701978903, | |
| "loss": 1.2898, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 0.113735631108284, | |
| "learning_rate": 0.0001645761640713019, | |
| "loss": 1.1826, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 0.10678713023662567, | |
| "learning_rate": 0.00016426123635429786, | |
| "loss": 1.2231, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 0.11631446331739426, | |
| "learning_rate": 0.00016394521920778948, | |
| "loss": 1.0241, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 0.11213452368974686, | |
| "learning_rate": 0.00016362811798925853, | |
| "loss": 1.4935, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 0.1304715871810913, | |
| "learning_rate": 0.0001633099380745652, | |
| "loss": 1.1686, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 0.10429395735263824, | |
| "learning_rate": 0.000162990684857857, | |
| "loss": 1.0118, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 0.10091353952884674, | |
| "learning_rate": 0.00016267036375147725, | |
| "loss": 1.1401, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 0.10777878761291504, | |
| "learning_rate": 0.00016234898018587337, | |
| "loss": 1.33, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 0.11149658262729645, | |
| "learning_rate": 0.00016202653960950474, | |
| "loss": 1.2201, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 0.10549131035804749, | |
| "learning_rate": 0.0001617030474887505, | |
| "loss": 1.0758, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 0.12649282813072205, | |
| "learning_rate": 0.0001613785093078166, | |
| "loss": 1.2241, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 0.11720848083496094, | |
| "learning_rate": 0.00016105293056864315, | |
| "loss": 1.3001, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 0.12630730867385864, | |
| "learning_rate": 0.0001607263167908109, | |
| "loss": 0.986, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 0.11931440979242325, | |
| "learning_rate": 0.00016039867351144778, | |
| "loss": 1.2037, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 0.1198689192533493, | |
| "learning_rate": 0.000160070006285135, | |
| "loss": 1.2556, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 0.11019983142614365, | |
| "learning_rate": 0.0001597403206838128, | |
| "loss": 1.004, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 0.11280027031898499, | |
| "learning_rate": 0.00015940962229668625, | |
| "loss": 1.3164, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 0.12924431264400482, | |
| "learning_rate": 0.00015907791673013016, | |
| "loss": 1.1612, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 0.1206149235367775, | |
| "learning_rate": 0.00015874520960759423, | |
| "loss": 1.2171, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 0.12173935025930405, | |
| "learning_rate": 0.0001584115065695077, | |
| "loss": 1.2215, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 0.126984640955925, | |
| "learning_rate": 0.0001580768132731837, | |
| "loss": 1.2379, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 0.12142551690340042, | |
| "learning_rate": 0.0001577411353927233, | |
| "loss": 1.1171, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 0.11669143289327621, | |
| "learning_rate": 0.00015740447861891944, | |
| "loss": 1.1148, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 0.12235064059495926, | |
| "learning_rate": 0.00015706684865916026, | |
| "loss": 1.2705, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 0.11695750802755356, | |
| "learning_rate": 0.00015672825123733258, | |
| "loss": 1.1602, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 0.10883978754281998, | |
| "learning_rate": 0.00015638869209372457, | |
| "loss": 1.3629, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 0.09961367398500443, | |
| "learning_rate": 0.00015604817698492885, | |
| "loss": 1.1844, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 0.10849742591381073, | |
| "learning_rate": 0.00015570671168374438, | |
| "loss": 1.2453, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 0.13158155977725983, | |
| "learning_rate": 0.00015536430197907904, | |
| "loss": 1.0024, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 0.11819962412118912, | |
| "learning_rate": 0.00015502095367585122, | |
| "loss": 1.2432, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 0.10652721673250198, | |
| "learning_rate": 0.0001546766725948916, | |
| "loss": 1.2042, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 0.10861877351999283, | |
| "learning_rate": 0.0001543314645728442, | |
| "loss": 1.4347, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 0.11992961913347244, | |
| "learning_rate": 0.00015398533546206778, | |
| "loss": 1.2897, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 0.11120472103357315, | |
| "learning_rate": 0.00015363829113053634, | |
| "loss": 1.1729, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 0.11962945014238358, | |
| "learning_rate": 0.00015329033746173975, | |
| "loss": 1.0863, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 0.10950727015733719, | |
| "learning_rate": 0.00015294148035458404, | |
| "loss": 1.2127, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 0.11483786255121231, | |
| "learning_rate": 0.0001525917257232913, | |
| "loss": 1.1465, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 0.11853177845478058, | |
| "learning_rate": 0.0001522410794972995, | |
| "loss": 1.2527, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 0.10667990148067474, | |
| "learning_rate": 0.00015188954762116195, | |
| "loss": 1.2347, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 0.1277850717306137, | |
| "learning_rate": 0.0001515371360544465, | |
| "loss": 1.2875, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 0.10839621722698212, | |
| "learning_rate": 0.00015118385077163445, | |
| "loss": 1.3415, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 0.11405637860298157, | |
| "learning_rate": 0.00015082969776201947, | |
| "loss": 1.1112, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 0.1116739809513092, | |
| "learning_rate": 0.00015047468302960577, | |
| "loss": 1.2876, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 0.10504814237356186, | |
| "learning_rate": 0.00015011881259300653, | |
| "loss": 1.2027, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 0.10286784917116165, | |
| "learning_rate": 0.00014976209248534182, | |
| "loss": 1.1372, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 0.1141451969742775, | |
| "learning_rate": 0.00014940452875413627, | |
| "loss": 1.0566, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 0.12613342702388763, | |
| "learning_rate": 0.00014904612746121659, | |
| "loss": 0.9762, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 0.11521808058023453, | |
| "learning_rate": 0.00014868689468260874, | |
| "loss": 1.144, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 0.10090262442827225, | |
| "learning_rate": 0.0001483268365084351, | |
| "loss": 1.0197, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 0.11367885768413544, | |
| "learning_rate": 0.0001479659590428109, | |
| "loss": 1.3636, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 0.11813089996576309, | |
| "learning_rate": 0.0001476042684037411, | |
| "loss": 1.314, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 0.10709206759929657, | |
| "learning_rate": 0.0001472417707230164, | |
| "loss": 1.2899, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 0.11429055780172348, | |
| "learning_rate": 0.00014687847214610943, | |
| "loss": 1.2237, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 0.10949011892080307, | |
| "learning_rate": 0.00014651437883207054, | |
| "loss": 1.3596, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 0.1138731986284256, | |
| "learning_rate": 0.00014614949695342336, | |
| "loss": 1.0538, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 0.10718726366758347, | |
| "learning_rate": 0.00014578383269606005, | |
| "loss": 1.2639, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 0.10555477440357208, | |
| "learning_rate": 0.0001454173922591367, | |
| "loss": 1.2917, | |
| "step": 467 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 0.10175728797912598, | |
| "learning_rate": 0.00014505018185496802, | |
| "loss": 1.111, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 0.113253653049469, | |
| "learning_rate": 0.00014468220770892206, | |
| "loss": 1.0426, | |
| "step": 469 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 0.11308235675096512, | |
| "learning_rate": 0.0001443134760593147, | |
| "loss": 1.2584, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 0.10383269935846329, | |
| "learning_rate": 0.0001439439931573039, | |
| "loss": 1.0795, | |
| "step": 471 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 0.10998781770467758, | |
| "learning_rate": 0.00014357376526678367, | |
| "loss": 1.2224, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 0.12233158946037292, | |
| "learning_rate": 0.00014320279866427796, | |
| "loss": 1.3125, | |
| "step": 473 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 0.11626295000314713, | |
| "learning_rate": 0.00014283109963883419, | |
| "loss": 1.3173, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 0.12676700949668884, | |
| "learning_rate": 0.0001424586744919166, | |
| "loss": 1.3502, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 0.11780429631471634, | |
| "learning_rate": 0.00014208552953729947, | |
| "loss": 1.3161, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 0.11618795245885849, | |
| "learning_rate": 0.00014171167110096016, | |
| "loss": 1.2574, | |
| "step": 477 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 0.10930506885051727, | |
| "learning_rate": 0.00014133710552097174, | |
| "loss": 1.243, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 0.11596094816923141, | |
| "learning_rate": 0.00014096183914739553, | |
| "loss": 1.2908, | |
| "step": 479 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 0.13881249725818634, | |
| "learning_rate": 0.00014058587834217355, | |
| "loss": 1.1379, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 0.09807305783033371, | |
| "learning_rate": 0.00014020922947902066, | |
| "loss": 1.1799, | |
| "step": 481 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 0.10549982637166977, | |
| "learning_rate": 0.00013983189894331636, | |
| "loss": 1.2467, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 0.7992545366287231, | |
| "learning_rate": 0.00013945389313199669, | |
| "loss": 1.2711, | |
| "step": 483 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 0.18441331386566162, | |
| "learning_rate": 0.00013907521845344571, | |
| "loss": 1.0915, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 0.12779419124126434, | |
| "learning_rate": 0.00013869588132738695, | |
| "loss": 1.2633, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 0.13042038679122925, | |
| "learning_rate": 0.00013831588818477436, | |
| "loss": 1.2233, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 0.14113464951515198, | |
| "learning_rate": 0.00013793524546768356, | |
| "loss": 1.0653, | |
| "step": 487 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 0.12155605107545853, | |
| "learning_rate": 0.0001375539596292025, | |
| "loss": 1.2181, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 0.12651774287223816, | |
| "learning_rate": 0.00013717203713332188, | |
| "loss": 1.3338, | |
| "step": 489 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 0.2244076430797577, | |
| "learning_rate": 0.00013678948445482598, | |
| "loss": 1.3459, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 0.1257290095090866, | |
| "learning_rate": 0.00013640630807918247, | |
| "loss": 1.0061, | |
| "step": 491 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 0.11858661472797394, | |
| "learning_rate": 0.00013602251450243273, | |
| "loss": 1.1775, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 0.1279936581850052, | |
| "learning_rate": 0.00013563811023108158, | |
| "loss": 1.3702, | |
| "step": 493 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 0.13254624605178833, | |
| "learning_rate": 0.00013525310178198705, | |
| "loss": 1.3612, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 0.1243981122970581, | |
| "learning_rate": 0.00013486749568225001, | |
| "loss": 1.1652, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 0.1259343922138214, | |
| "learning_rate": 0.0001344812984691031, | |
| "loss": 1.2998, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 0.12421772629022598, | |
| "learning_rate": 0.00013409451668980048, | |
| "loss": 1.2688, | |
| "step": 497 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 0.11533673852682114, | |
| "learning_rate": 0.0001337071569015063, | |
| "loss": 1.2123, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 0.2135774940252304, | |
| "learning_rate": 0.00013331922567118393, | |
| "loss": 1.173, | |
| "step": 499 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 0.1305667608976364, | |
| "learning_rate": 0.00013293072957548443, | |
| "loss": 1.2795, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 0.11869475990533829, | |
| "learning_rate": 0.00013254167520063509, | |
| "loss": 1.1984, | |
| "step": 501 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 0.13584646582603455, | |
| "learning_rate": 0.0001321520691423278, | |
| "loss": 1.2018, | |
| "step": 502 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 0.12778714299201965, | |
| "learning_rate": 0.0001317619180056072, | |
| "loss": 1.2817, | |
| "step": 503 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 0.12326094508171082, | |
| "learning_rate": 0.0001313712284047587, | |
| "loss": 1.3434, | |
| "step": 504 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 0.11255846172571182, | |
| "learning_rate": 0.0001309800069631964, | |
| "loss": 1.1986, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 0.11870656907558441, | |
| "learning_rate": 0.0001305882603133508, | |
| "loss": 1.2409, | |
| "step": 506 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 0.11277451366186142, | |
| "learning_rate": 0.0001301959950965562, | |
| "loss": 1.4011, | |
| "step": 507 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 0.11900844424962997, | |
| "learning_rate": 0.00012980321796293836, | |
| "loss": 1.2379, | |
| "step": 508 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 0.14404135942459106, | |
| "learning_rate": 0.00012940993557130164, | |
| "loss": 1.3902, | |
| "step": 509 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 0.12229623645544052, | |
| "learning_rate": 0.000129016154589016, | |
| "loss": 1.2842, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 0.1197381243109703, | |
| "learning_rate": 0.00012862188169190418, | |
| "loss": 1.3836, | |
| "step": 511 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 0.12395986169576645, | |
| "learning_rate": 0.00012822712356412838, | |
| "loss": 1.3508, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 0.12418900430202484, | |
| "learning_rate": 0.00012783188689807697, | |
| "loss": 1.1969, | |
| "step": 513 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 0.1138986125588417, | |
| "learning_rate": 0.00012743617839425098, | |
| "loss": 1.3654, | |
| "step": 514 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 0.11219038814306259, | |
| "learning_rate": 0.0001270400047611508, | |
| "loss": 1.2621, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 0.11220891773700714, | |
| "learning_rate": 0.00012664337271516194, | |
| "loss": 1.2384, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 0.11203695833683014, | |
| "learning_rate": 0.0001262462889804416, | |
| "loss": 1.1551, | |
| "step": 517 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 0.10459273308515549, | |
| "learning_rate": 0.00012584876028880453, | |
| "loss": 1.2276, | |
| "step": 518 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 0.10843029618263245, | |
| "learning_rate": 0.00012545079337960882, | |
| "loss": 1.2414, | |
| "step": 519 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 0.10928460955619812, | |
| "learning_rate": 0.00012505239499964178, | |
| "loss": 1.0391, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 0.1170789897441864, | |
| "learning_rate": 0.0001246535719030055, | |
| "loss": 1.2558, | |
| "step": 521 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 0.118638776242733, | |
| "learning_rate": 0.00012425433085100224, | |
| "loss": 1.2906, | |
| "step": 522 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 0.12077479064464569, | |
| "learning_rate": 0.0001238546786120201, | |
| "loss": 1.3222, | |
| "step": 523 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 0.12042129039764404, | |
| "learning_rate": 0.00012345462196141797, | |
| "loss": 1.2399, | |
| "step": 524 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 0.11980915814638138, | |
| "learning_rate": 0.00012305416768141082, | |
| "loss": 1.2426, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 0.10789747536182404, | |
| "learning_rate": 0.00012265332256095464, | |
| "loss": 1.3437, | |
| "step": 526 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 0.10967132449150085, | |
| "learning_rate": 0.00012225209339563145, | |
| "loss": 1.1329, | |
| "step": 527 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 0.11903247982263565, | |
| "learning_rate": 0.00012185048698753403, | |
| "loss": 1.2107, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 0.10917498171329498, | |
| "learning_rate": 0.00012144851014515055, | |
| "loss": 1.1689, | |
| "step": 529 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 0.11308782547712326, | |
| "learning_rate": 0.00012104616968324927, | |
| "loss": 1.2487, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 0.10030585527420044, | |
| "learning_rate": 0.00012064347242276293, | |
| "loss": 1.1451, | |
| "step": 531 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 0.1195712685585022, | |
| "learning_rate": 0.00012024042519067309, | |
| "loss": 1.2525, | |
| "step": 532 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 0.12115343660116196, | |
| "learning_rate": 0.00011983703481989443, | |
| "loss": 1.2348, | |
| "step": 533 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 0.1060243546962738, | |
| "learning_rate": 0.00011943330814915896, | |
| "loss": 1.0223, | |
| "step": 534 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 0.2178022414445877, | |
| "learning_rate": 0.00011902925202289996, | |
| "loss": 1.1158, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 0.13022154569625854, | |
| "learning_rate": 0.00011862487329113606, | |
| "loss": 1.2742, | |
| "step": 536 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 0.11495082825422287, | |
| "learning_rate": 0.00011822017880935505, | |
| "loss": 1.2562, | |
| "step": 537 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 0.11310116201639175, | |
| "learning_rate": 0.00011781517543839768, | |
| "loss": 1.2719, | |
| "step": 538 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 0.13280071318149567, | |
| "learning_rate": 0.00011740987004434137, | |
| "loss": 1.1481, | |
| "step": 539 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 0.12752439081668854, | |
| "learning_rate": 0.00011700426949838363, | |
| "loss": 1.3024, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 0.11650800704956055, | |
| "learning_rate": 0.00011659838067672599, | |
| "loss": 1.2154, | |
| "step": 541 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 0.11279530823230743, | |
| "learning_rate": 0.00011619221046045687, | |
| "loss": 1.1627, | |
| "step": 542 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 0.12641020119190216, | |
| "learning_rate": 0.0001157857657354354, | |
| "loss": 1.0148, | |
| "step": 543 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 0.10685007274150848, | |
| "learning_rate": 0.00011537905339217448, | |
| "loss": 1.288, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 0.10929278284311295, | |
| "learning_rate": 0.00011497208032572384, | |
| "loss": 1.123, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 0.11334774643182755, | |
| "learning_rate": 0.00011456485343555344, | |
| "loss": 1.2894, | |
| "step": 546 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 0.10330837219953537, | |
| "learning_rate": 0.0001141573796254363, | |
| "loss": 1.1484, | |
| "step": 547 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 0.09870249778032303, | |
| "learning_rate": 0.00011374966580333147, | |
| "loss": 1.0223, | |
| "step": 548 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 0.12955228984355927, | |
| "learning_rate": 0.00011334171888126697, | |
| "loss": 1.2058, | |
| "step": 549 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 0.11018647998571396, | |
| "learning_rate": 0.00011293354577522263, | |
| "loss": 1.1589, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 0.12055356800556183, | |
| "learning_rate": 0.00011252515340501281, | |
| "loss": 1.091, | |
| "step": 551 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 0.10831710696220398, | |
| "learning_rate": 0.00011211654869416901, | |
| "loss": 1.2076, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 0.11912648379802704, | |
| "learning_rate": 0.00011170773856982268, | |
| "loss": 1.2953, | |
| "step": 553 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 0.11210989207029343, | |
| "learning_rate": 0.00011129872996258757, | |
| "loss": 1.0549, | |
| "step": 554 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 0.10992615669965744, | |
| "learning_rate": 0.0001108895298064424, | |
| "loss": 1.1818, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 0.1124982088804245, | |
| "learning_rate": 0.00011048014503861321, | |
| "loss": 1.3218, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 0.12512408196926117, | |
| "learning_rate": 0.00011007058259945584, | |
| "loss": 1.2262, | |
| "step": 557 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 0.1207667887210846, | |
| "learning_rate": 0.00010966084943233817, | |
| "loss": 1.1636, | |
| "step": 558 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 0.10861886292695999, | |
| "learning_rate": 0.00010925095248352239, | |
| "loss": 1.1363, | |
| "step": 559 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 0.11505457758903503, | |
| "learning_rate": 0.0001088408987020475, | |
| "loss": 1.1823, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 0.10509267449378967, | |
| "learning_rate": 0.00010843069503961111, | |
| "loss": 1.2269, | |
| "step": 561 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 0.10376826673746109, | |
| "learning_rate": 0.00010802034845045188, | |
| "loss": 1.0244, | |
| "step": 562 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 0.10977223515510559, | |
| "learning_rate": 0.00010760986589123146, | |
| "loss": 1.2713, | |
| "step": 563 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 0.11492714285850525, | |
| "learning_rate": 0.0001071992543209167, | |
| "loss": 1.2242, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 0.10544589161872864, | |
| "learning_rate": 0.00010678852070066159, | |
| "loss": 1.2959, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 0.11260811984539032, | |
| "learning_rate": 0.00010637767199368912, | |
| "loss": 1.3398, | |
| "step": 566 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 0.10535457730293274, | |
| "learning_rate": 0.00010596671516517356, | |
| "loss": 1.3178, | |
| "step": 567 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 0.10959797352552414, | |
| "learning_rate": 0.00010555565718212197, | |
| "loss": 1.0378, | |
| "step": 568 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 0.10014626383781433, | |
| "learning_rate": 0.00010514450501325645, | |
| "loss": 1.2384, | |
| "step": 569 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 0.1097472757101059, | |
| "learning_rate": 0.00010473326562889582, | |
| "loss": 1.1366, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 0.10998505353927612, | |
| "learning_rate": 0.0001043219460008374, | |
| "loss": 1.1083, | |
| "step": 571 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 0.11129113286733627, | |
| "learning_rate": 0.00010391055310223899, | |
| "loss": 1.1951, | |
| "step": 572 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 0.10348080098628998, | |
| "learning_rate": 0.00010349909390750046, | |
| "loss": 1.1871, | |
| "step": 573 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 0.10865020006895065, | |
| "learning_rate": 0.00010308757539214572, | |
| "loss": 1.2269, | |
| "step": 574 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 0.1128375232219696, | |
| "learning_rate": 0.00010267600453270421, | |
| "loss": 1.2358, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 0.13256721198558807, | |
| "learning_rate": 0.0001022643883065929, | |
| "loss": 1.3445, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 0.10738494992256165, | |
| "learning_rate": 0.00010185273369199781, | |
| "loss": 1.1042, | |
| "step": 577 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 0.10427848994731903, | |
| "learning_rate": 0.00010144104766775572, | |
| "loss": 1.0901, | |
| "step": 578 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 0.11618473380804062, | |
| "learning_rate": 0.000101029337213236, | |
| "loss": 0.781, | |
| "step": 579 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 0.20147624611854553, | |
| "learning_rate": 0.00010061760930822211, | |
| "loss": 1.2707, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 0.11436042189598083, | |
| "learning_rate": 0.0001002058709327934, | |
| "loss": 1.193, | |
| "step": 581 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 0.12396758794784546, | |
| "learning_rate": 9.979412906720663e-05, | |
| "loss": 1.0342, | |
| "step": 582 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 0.11723118275403976, | |
| "learning_rate": 9.938239069177791e-05, | |
| "loss": 1.1791, | |
| "step": 583 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 0.12710683047771454, | |
| "learning_rate": 9.897066278676404e-05, | |
| "loss": 1.023, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 0.10067608207464218, | |
| "learning_rate": 9.85589523322443e-05, | |
| "loss": 1.1753, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 0.11074398458003998, | |
| "learning_rate": 9.81472663080022e-05, | |
| "loss": 1.2567, | |
| "step": 586 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 0.16948086023330688, | |
| "learning_rate": 9.773561169340712e-05, | |
| "loss": 1.3789, | |
| "step": 587 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 0.10852295160293579, | |
| "learning_rate": 9.732399546729578e-05, | |
| "loss": 1.3642, | |
| "step": 588 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 0.10131103545427322, | |
| "learning_rate": 9.691242460785433e-05, | |
| "loss": 1.129, | |
| "step": 589 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 0.11037011444568634, | |
| "learning_rate": 9.650090609249956e-05, | |
| "loss": 1.1974, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 0.12829488515853882, | |
| "learning_rate": 9.608944689776103e-05, | |
| "loss": 1.2126, | |
| "step": 591 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 0.16550996899604797, | |
| "learning_rate": 9.567805399916259e-05, | |
| "loss": 1.228, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 0.1353956162929535, | |
| "learning_rate": 9.526673437110419e-05, | |
| "loss": 1.2877, | |
| "step": 593 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 0.17490524053573608, | |
| "learning_rate": 9.485549498674358e-05, | |
| "loss": 1.212, | |
| "step": 594 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 0.10544750839471817, | |
| "learning_rate": 9.444434281787807e-05, | |
| "loss": 1.3147, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 0.10780143737792969, | |
| "learning_rate": 9.403328483482646e-05, | |
| "loss": 1.1851, | |
| "step": 596 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 0.12775209546089172, | |
| "learning_rate": 9.362232800631088e-05, | |
| "loss": 1.1219, | |
| "step": 597 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 0.11476773023605347, | |
| "learning_rate": 9.321147929933847e-05, | |
| "loss": 1.2289, | |
| "step": 598 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 0.11137551069259644, | |
| "learning_rate": 9.28007456790833e-05, | |
| "loss": 1.2016, | |
| "step": 599 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 0.13261277973651886, | |
| "learning_rate": 9.239013410876857e-05, | |
| "loss": 1.389, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 0.10396253317594528, | |
| "learning_rate": 9.197965154954815e-05, | |
| "loss": 1.3316, | |
| "step": 601 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 0.11388034373521805, | |
| "learning_rate": 9.15693049603889e-05, | |
| "loss": 1.2501, | |
| "step": 602 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 0.1430211067199707, | |
| "learning_rate": 9.115910129795251e-05, | |
| "loss": 1.2259, | |
| "step": 603 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 0.12369144707918167, | |
| "learning_rate": 9.074904751647762e-05, | |
| "loss": 1.138, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 0.11555325239896774, | |
| "learning_rate": 9.033915056766187e-05, | |
| "loss": 1.1845, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 0.15159161388874054, | |
| "learning_rate": 8.992941740054418e-05, | |
| "loss": 1.3823, | |
| "step": 606 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 0.12028995901346207, | |
| "learning_rate": 8.951985496138678e-05, | |
| "loss": 1.0985, | |
| "step": 607 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 0.12002217024564743, | |
| "learning_rate": 8.911047019355764e-05, | |
| "loss": 1.2431, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 0.11859162151813507, | |
| "learning_rate": 8.870127003741244e-05, | |
| "loss": 1.1669, | |
| "step": 609 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 0.10495835542678833, | |
| "learning_rate": 8.829226143017735e-05, | |
| "loss": 1.1424, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 0.11981993168592453, | |
| "learning_rate": 8.788345130583099e-05, | |
| "loss": 1.0444, | |
| "step": 611 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 0.11924871802330017, | |
| "learning_rate": 8.747484659498724e-05, | |
| "loss": 1.1574, | |
| "step": 612 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 0.15531030297279358, | |
| "learning_rate": 8.706645422477739e-05, | |
| "loss": 1.1786, | |
| "step": 613 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 0.11026681959629059, | |
| "learning_rate": 8.665828111873307e-05, | |
| "loss": 1.2112, | |
| "step": 614 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 0.12397878617048264, | |
| "learning_rate": 8.625033419666857e-05, | |
| "loss": 1.2931, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 0.10594267398118973, | |
| "learning_rate": 8.584262037456373e-05, | |
| "loss": 1.1422, | |
| "step": 616 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 0.11834708601236343, | |
| "learning_rate": 8.54351465644466e-05, | |
| "loss": 1.16, | |
| "step": 617 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 0.12605226039886475, | |
| "learning_rate": 8.502791967427618e-05, | |
| "loss": 1.3575, | |
| "step": 618 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 0.12767426669597626, | |
| "learning_rate": 8.462094660782556e-05, | |
| "loss": 1.3343, | |
| "step": 619 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 0.12377568334341049, | |
| "learning_rate": 8.42142342645646e-05, | |
| "loss": 1.2027, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 0.11611881852149963, | |
| "learning_rate": 8.380778953954313e-05, | |
| "loss": 1.1748, | |
| "step": 621 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 0.10847889631986618, | |
| "learning_rate": 8.340161932327404e-05, | |
| "loss": 1.138, | |
| "step": 622 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 0.11104707419872284, | |
| "learning_rate": 8.299573050161638e-05, | |
| "loss": 1.2403, | |
| "step": 623 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 0.10397801548242569, | |
| "learning_rate": 8.259012995565867e-05, | |
| "loss": 1.2159, | |
| "step": 624 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 0.10560380667448044, | |
| "learning_rate": 8.21848245616023e-05, | |
| "loss": 1.0932, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 0.10818043351173401, | |
| "learning_rate": 8.177982119064496e-05, | |
| "loss": 1.2667, | |
| "step": 626 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 0.1053985133767128, | |
| "learning_rate": 8.137512670886397e-05, | |
| "loss": 1.3267, | |
| "step": 627 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 0.10604721307754517, | |
| "learning_rate": 8.097074797710006e-05, | |
| "loss": 1.2254, | |
| "step": 628 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 0.1083521693944931, | |
| "learning_rate": 8.056669185084108e-05, | |
| "loss": 1.389, | |
| "step": 629 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 0.12045227736234665, | |
| "learning_rate": 8.016296518010558e-05, | |
| "loss": 1.245, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 0.11613083630800247, | |
| "learning_rate": 7.975957480932695e-05, | |
| "loss": 1.2618, | |
| "step": 631 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 0.11401872336864471, | |
| "learning_rate": 7.935652757723708e-05, | |
| "loss": 1.2073, | |
| "step": 632 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 0.1841176450252533, | |
| "learning_rate": 7.895383031675074e-05, | |
| "loss": 1.2132, | |
| "step": 633 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 0.10668601095676422, | |
| "learning_rate": 7.855148985484946e-05, | |
| "loss": 1.2177, | |
| "step": 634 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 0.10970567911863327, | |
| "learning_rate": 7.814951301246598e-05, | |
| "loss": 1.2306, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 0.10744346678256989, | |
| "learning_rate": 7.774790660436858e-05, | |
| "loss": 1.2155, | |
| "step": 636 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 0.09847363084554672, | |
| "learning_rate": 7.734667743904539e-05, | |
| "loss": 1.1383, | |
| "step": 637 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 0.12585873901844025, | |
| "learning_rate": 7.69458323185892e-05, | |
| "loss": 1.2411, | |
| "step": 638 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 0.13609972596168518, | |
| "learning_rate": 7.654537803858204e-05, | |
| "loss": 1.1972, | |
| "step": 639 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 0.10885506868362427, | |
| "learning_rate": 7.614532138797993e-05, | |
| "loss": 1.2197, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 0.10706409066915512, | |
| "learning_rate": 7.57456691489978e-05, | |
| "loss": 1.0704, | |
| "step": 641 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 0.11232727020978928, | |
| "learning_rate": 7.534642809699455e-05, | |
| "loss": 1.2443, | |
| "step": 642 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 0.1036246195435524, | |
| "learning_rate": 7.494760500035823e-05, | |
| "loss": 1.1623, | |
| "step": 643 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 0.11857923120260239, | |
| "learning_rate": 7.454920662039117e-05, | |
| "loss": 1.3161, | |
| "step": 644 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 0.1063128188252449, | |
| "learning_rate": 7.415123971119549e-05, | |
| "loss": 1.37, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 0.11207219213247299, | |
| "learning_rate": 7.375371101955841e-05, | |
| "loss": 1.1433, | |
| "step": 646 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 0.1198432594537735, | |
| "learning_rate": 7.335662728483808e-05, | |
| "loss": 1.2341, | |
| "step": 647 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 0.14110444486141205, | |
| "learning_rate": 7.295999523884921e-05, | |
| "loss": 1.2201, | |
| "step": 648 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 0.10053621232509613, | |
| "learning_rate": 7.256382160574901e-05, | |
| "loss": 1.1276, | |
| "step": 649 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 0.1087542399764061, | |
| "learning_rate": 7.216811310192308e-05, | |
| "loss": 1.3124, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 0.10470487177371979, | |
| "learning_rate": 7.177287643587163e-05, | |
| "loss": 1.0677, | |
| "step": 651 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 0.12348546087741852, | |
| "learning_rate": 7.137811830809583e-05, | |
| "loss": 1.0808, | |
| "step": 652 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 0.12709330022335052, | |
| "learning_rate": 7.0983845410984e-05, | |
| "loss": 1.1746, | |
| "step": 653 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 0.0996909812092781, | |
| "learning_rate": 7.059006442869839e-05, | |
| "loss": 1.2121, | |
| "step": 654 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 0.12972918152809143, | |
| "learning_rate": 7.019678203706163e-05, | |
| "loss": 1.149, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 0.10252442955970764, | |
| "learning_rate": 6.980400490344383e-05, | |
| "loss": 1.2643, | |
| "step": 656 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 0.11937206983566284, | |
| "learning_rate": 6.941173968664923e-05, | |
| "loss": 1.2484, | |
| "step": 657 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 0.11440838873386383, | |
| "learning_rate": 6.901999303680358e-05, | |
| "loss": 1.3256, | |
| "step": 658 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 0.10103975981473923, | |
| "learning_rate": 6.862877159524132e-05, | |
| "loss": 1.3196, | |
| "step": 659 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 0.10081090033054352, | |
| "learning_rate": 6.823808199439284e-05, | |
| "loss": 1.1371, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 0.16644109785556793, | |
| "learning_rate": 6.784793085767221e-05, | |
| "loss": 1.1239, | |
| "step": 661 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 0.1086033284664154, | |
| "learning_rate": 6.745832479936491e-05, | |
| "loss": 1.3244, | |
| "step": 662 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 0.11194080859422684, | |
| "learning_rate": 6.70692704245156e-05, | |
| "loss": 1.3153, | |
| "step": 663 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 0.1124996691942215, | |
| "learning_rate": 6.668077432881609e-05, | |
| "loss": 1.2942, | |
| "step": 664 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 0.3689382076263428, | |
| "learning_rate": 6.629284309849372e-05, | |
| "loss": 1.2606, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 0.11704660952091217, | |
| "learning_rate": 6.590548331019957e-05, | |
| "loss": 1.3214, | |
| "step": 666 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 0.11636264622211456, | |
| "learning_rate": 6.55187015308969e-05, | |
| "loss": 1.3366, | |
| "step": 667 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 0.10623035579919815, | |
| "learning_rate": 6.513250431775004e-05, | |
| "loss": 1.1485, | |
| "step": 668 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 0.10182057321071625, | |
| "learning_rate": 6.474689821801295e-05, | |
| "loss": 1.1521, | |
| "step": 669 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 0.11016492545604706, | |
| "learning_rate": 6.436188976891846e-05, | |
| "loss": 1.2067, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 0.11019518971443176, | |
| "learning_rate": 6.39774854975673e-05, | |
| "loss": 1.2559, | |
| "step": 671 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 0.11215123534202576, | |
| "learning_rate": 6.359369192081756e-05, | |
| "loss": 1.2557, | |
| "step": 672 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 0.1162608414888382, | |
| "learning_rate": 6.321051554517407e-05, | |
| "loss": 1.1967, | |
| "step": 673 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 0.1348588615655899, | |
| "learning_rate": 6.282796286667814e-05, | |
| "loss": 1.1944, | |
| "step": 674 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 0.12534910440444946, | |
| "learning_rate": 6.244604037079753e-05, | |
| "loss": 1.1354, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 0.11349747329950333, | |
| "learning_rate": 6.206475453231644e-05, | |
| "loss": 1.3001, | |
| "step": 676 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 0.11473750323057175, | |
| "learning_rate": 6.168411181522569e-05, | |
| "loss": 1.3084, | |
| "step": 677 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 0.11263560503721237, | |
| "learning_rate": 6.13041186726131e-05, | |
| "loss": 1.0361, | |
| "step": 678 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 0.11515972018241882, | |
| "learning_rate": 6.092478154655431e-05, | |
| "loss": 1.1342, | |
| "step": 679 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 0.10681987553834915, | |
| "learning_rate": 6.0546106868003324e-05, | |
| "loss": 1.2221, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 0.10119231045246124, | |
| "learning_rate": 6.016810105668365e-05, | |
| "loss": 1.1237, | |
| "step": 681 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 0.10480265319347382, | |
| "learning_rate": 5.979077052097936e-05, | |
| "loss": 1.2506, | |
| "step": 682 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 0.11073455959558487, | |
| "learning_rate": 5.941412165782645e-05, | |
| "loss": 1.0603, | |
| "step": 683 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 0.11075324565172195, | |
| "learning_rate": 5.9038160852604476e-05, | |
| "loss": 1.1351, | |
| "step": 684 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 0.10219030827283859, | |
| "learning_rate": 5.866289447902828e-05, | |
| "loss": 1.2163, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 0.10446413606405258, | |
| "learning_rate": 5.8288328899039834e-05, | |
| "loss": 1.2578, | |
| "step": 686 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 0.10746556520462036, | |
| "learning_rate": 5.791447046270056e-05, | |
| "loss": 1.2526, | |
| "step": 687 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 0.12443575263023376, | |
| "learning_rate": 5.754132550808344e-05, | |
| "loss": 1.3376, | |
| "step": 688 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 0.12434745579957962, | |
| "learning_rate": 5.7168900361165824e-05, | |
| "loss": 1.369, | |
| "step": 689 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 0.11743592470884323, | |
| "learning_rate": 5.679720133572206e-05, | |
| "loss": 1.2044, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 0.12590321898460388, | |
| "learning_rate": 5.6426234733216376e-05, | |
| "loss": 1.0729, | |
| "step": 691 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 0.10635906457901001, | |
| "learning_rate": 5.605600684269614e-05, | |
| "loss": 1.2448, | |
| "step": 692 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 0.10504303872585297, | |
| "learning_rate": 5.568652394068532e-05, | |
| "loss": 1.0663, | |
| "step": 693 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 0.10200861841440201, | |
| "learning_rate": 5.5317792291077965e-05, | |
| "loss": 1.0939, | |
| "step": 694 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 0.11620642244815826, | |
| "learning_rate": 5.4949818145031986e-05, | |
| "loss": 1.1002, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 0.10879629105329514, | |
| "learning_rate": 5.4582607740863324e-05, | |
| "loss": 1.1695, | |
| "step": 696 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 0.11083561182022095, | |
| "learning_rate": 5.421616730394e-05, | |
| "loss": 1.2967, | |
| "step": 697 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 0.11363320797681808, | |
| "learning_rate": 5.385050304657669e-05, | |
| "loss": 1.0099, | |
| "step": 698 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 0.10837159305810928, | |
| "learning_rate": 5.348562116792946e-05, | |
| "loss": 1.1575, | |
| "step": 699 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 0.7101715207099915, | |
| "learning_rate": 5.312152785389055e-05, | |
| "loss": 1.3229, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 0.11040350794792175, | |
| "learning_rate": 5.2758229276983614e-05, | |
| "loss": 1.183, | |
| "step": 701 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 0.10885446518659592, | |
| "learning_rate": 5.2395731596258925e-05, | |
| "loss": 1.139, | |
| "step": 702 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 0.10934137552976608, | |
| "learning_rate": 5.20340409571891e-05, | |
| "loss": 1.291, | |
| "step": 703 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 0.11229011416435242, | |
| "learning_rate": 5.167316349156495e-05, | |
| "loss": 1.1406, | |
| "step": 704 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 0.12293226271867752, | |
| "learning_rate": 5.1313105317391284e-05, | |
| "loss": 1.1102, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 0.10751207172870636, | |
| "learning_rate": 5.095387253878345e-05, | |
| "loss": 1.1296, | |
| "step": 706 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 0.10970565676689148, | |
| "learning_rate": 5.059547124586375e-05, | |
| "loss": 1.3482, | |
| "step": 707 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 0.10699622333049774, | |
| "learning_rate": 5.0237907514658176e-05, | |
| "loss": 1.3642, | |
| "step": 708 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 0.1132645532488823, | |
| "learning_rate": 4.988118740699346e-05, | |
| "loss": 1.1504, | |
| "step": 709 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 0.12392119318246841, | |
| "learning_rate": 4.9525316970394245e-05, | |
| "loss": 1.3906, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 0.10416343808174133, | |
| "learning_rate": 4.917030223798057e-05, | |
| "loss": 1.3419, | |
| "step": 711 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 0.11010722070932388, | |
| "learning_rate": 4.881614922836555e-05, | |
| "loss": 1.1175, | |
| "step": 712 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 0.11078814417123795, | |
| "learning_rate": 4.8462863945553516e-05, | |
| "loss": 1.3513, | |
| "step": 713 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 0.12069610506296158, | |
| "learning_rate": 4.8110452378838035e-05, | |
| "loss": 1.3913, | |
| "step": 714 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 0.11017792671918869, | |
| "learning_rate": 4.775892050270051e-05, | |
| "loss": 1.2217, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 0.10235986858606339, | |
| "learning_rate": 4.740827427670871e-05, | |
| "loss": 1.1375, | |
| "step": 716 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 0.11740509420633316, | |
| "learning_rate": 4.7058519645416e-05, | |
| "loss": 1.1649, | |
| "step": 717 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 0.16506601870059967, | |
| "learning_rate": 4.6709662538260267e-05, | |
| "loss": 1.0392, | |
| "step": 718 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 0.22638647258281708, | |
| "learning_rate": 4.63617088694637e-05, | |
| "loss": 1.21, | |
| "step": 719 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 0.11359802633523941, | |
| "learning_rate": 4.601466453793224e-05, | |
| "loss": 1.2481, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 0.10808828473091125, | |
| "learning_rate": 4.566853542715581e-05, | |
| "loss": 1.2008, | |
| "step": 721 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 0.09867287427186966, | |
| "learning_rate": 4.5323327405108416e-05, | |
| "loss": 1.0815, | |
| "step": 722 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 0.11430750787258148, | |
| "learning_rate": 4.4979046324148785e-05, | |
| "loss": 1.3082, | |
| "step": 723 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 0.1082565113902092, | |
| "learning_rate": 4.4635698020921014e-05, | |
| "loss": 1.0386, | |
| "step": 724 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 0.11467121541500092, | |
| "learning_rate": 4.4293288316255653e-05, | |
| "loss": 1.299, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 0.10791204124689102, | |
| "learning_rate": 4.395182301507118e-05, | |
| "loss": 1.2306, | |
| "step": 726 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 0.11149678379297256, | |
| "learning_rate": 4.3611307906275414e-05, | |
| "loss": 1.3074, | |
| "step": 727 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 0.11212314665317535, | |
| "learning_rate": 4.327174876266743e-05, | |
| "loss": 1.1512, | |
| "step": 728 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 0.1070462167263031, | |
| "learning_rate": 4.293315134083975e-05, | |
| "loss": 1.2378, | |
| "step": 729 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 0.11427627503871918, | |
| "learning_rate": 4.2595521381080604e-05, | |
| "loss": 1.2729, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 0.1135973185300827, | |
| "learning_rate": 4.2258864607276705e-05, | |
| "loss": 1.1399, | |
| "step": 731 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 0.11445244401693344, | |
| "learning_rate": 4.192318672681631e-05, | |
| "loss": 1.1549, | |
| "step": 732 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 0.10163676738739014, | |
| "learning_rate": 4.158849343049232e-05, | |
| "loss": 1.2751, | |
| "step": 733 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 0.11175056546926498, | |
| "learning_rate": 4.1254790392405797e-05, | |
| "loss": 1.2371, | |
| "step": 734 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 0.11040692776441574, | |
| "learning_rate": 4.092208326986986e-05, | |
| "loss": 1.3218, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 0.10620775073766708, | |
| "learning_rate": 4.0590377703313785e-05, | |
| "loss": 1.1667, | |
| "step": 736 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 0.10938204824924469, | |
| "learning_rate": 4.0259679316187214e-05, | |
| "loss": 1.0897, | |
| "step": 737 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 0.10027039796113968, | |
| "learning_rate": 3.992999371486508e-05, | |
| "loss": 1.1646, | |
| "step": 738 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 0.13585035502910614, | |
| "learning_rate": 3.960132648855226e-05, | |
| "loss": 1.1759, | |
| "step": 739 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 0.12253659218549728, | |
| "learning_rate": 3.9273683209189115e-05, | |
| "loss": 1.3585, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 0.14041317999362946, | |
| "learning_rate": 3.8947069431356855e-05, | |
| "loss": 1.2736, | |
| "step": 741 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 0.10640061646699905, | |
| "learning_rate": 3.862149069218343e-05, | |
| "loss": 1.2859, | |
| "step": 742 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 0.10962343961000443, | |
| "learning_rate": 3.829695251124953e-05, | |
| "loss": 1.2581, | |
| "step": 743 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 0.12311571091413498, | |
| "learning_rate": 3.797346039049528e-05, | |
| "loss": 1.3115, | |
| "step": 744 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 0.12078052759170532, | |
| "learning_rate": 3.7651019814126654e-05, | |
| "loss": 1.03, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 0.10591412335634232, | |
| "learning_rate": 3.732963624852275e-05, | |
| "loss": 1.0906, | |
| "step": 746 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 0.1085178554058075, | |
| "learning_rate": 3.700931514214303e-05, | |
| "loss": 1.1478, | |
| "step": 747 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 0.10265811532735825, | |
| "learning_rate": 3.6690061925434814e-05, | |
| "loss": 1.1375, | |
| "step": 748 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 0.10623796284198761, | |
| "learning_rate": 3.637188201074149e-05, | |
| "loss": 1.2655, | |
| "step": 749 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 0.1077788770198822, | |
| "learning_rate": 3.605478079221054e-05, | |
| "loss": 1.2805, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 0.1078350618481636, | |
| "learning_rate": 3.573876364570214e-05, | |
| "loss": 1.356, | |
| "step": 751 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 0.10767360776662827, | |
| "learning_rate": 3.5423835928698125e-05, | |
| "loss": 1.2346, | |
| "step": 752 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 0.10546786338090897, | |
| "learning_rate": 3.5110002980210975e-05, | |
| "loss": 1.2541, | |
| "step": 753 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 0.09910236299037933, | |
| "learning_rate": 3.479727012069349e-05, | |
| "loss": 1.2015, | |
| "step": 754 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 0.10072822123765945, | |
| "learning_rate": 3.4485642651948516e-05, | |
| "loss": 1.3005, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 0.11063572019338608, | |
| "learning_rate": 3.417512585703903e-05, | |
| "loss": 1.0739, | |
| "step": 756 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 0.09937085956335068, | |
| "learning_rate": 3.38657250001986e-05, | |
| "loss": 1.3106, | |
| "step": 757 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 0.10743856430053711, | |
| "learning_rate": 3.355744532674211e-05, | |
| "loss": 1.1492, | |
| "step": 758 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 0.11000777781009674, | |
| "learning_rate": 3.325029206297694e-05, | |
| "loss": 1.1876, | |
| "step": 759 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 0.11820725351572037, | |
| "learning_rate": 3.294427041611425e-05, | |
| "loss": 1.2457, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 0.10308333486318588, | |
| "learning_rate": 3.263938557418082e-05, | |
| "loss": 1.2801, | |
| "step": 761 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 0.1042330265045166, | |
| "learning_rate": 3.233564270593088e-05, | |
| "loss": 1.2237, | |
| "step": 762 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 0.10803329199552536, | |
| "learning_rate": 3.2033046960758764e-05, | |
| "loss": 1.2841, | |
| "step": 763 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 0.11426287144422531, | |
| "learning_rate": 3.1731603468611337e-05, | |
| "loss": 1.355, | |
| "step": 764 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 0.10273044556379318, | |
| "learning_rate": 3.143131733990127e-05, | |
| "loss": 1.2975, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 0.11743076890707016, | |
| "learning_rate": 3.113219366542031e-05, | |
| "loss": 1.2532, | |
| "step": 766 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 0.1157737672328949, | |
| "learning_rate": 3.083423751625282e-05, | |
| "loss": 1.4476, | |
| "step": 767 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 0.10297975689172745, | |
| "learning_rate": 3.0537453943690074e-05, | |
| "loss": 1.1543, | |
| "step": 768 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 0.10176219791173935, | |
| "learning_rate": 3.0241847979144487e-05, | |
| "loss": 1.0901, | |
| "step": 769 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 0.11567193269729614, | |
| "learning_rate": 2.994742463406427e-05, | |
| "loss": 1.25, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 0.12137153744697571, | |
| "learning_rate": 2.96541888998485e-05, | |
| "loss": 1.3956, | |
| "step": 771 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 0.10772846639156342, | |
| "learning_rate": 2.9362145747762625e-05, | |
| "loss": 1.4177, | |
| "step": 772 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 0.10832470655441284, | |
| "learning_rate": 2.9071300128854007e-05, | |
| "loss": 1.1614, | |
| "step": 773 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 0.09898839145898819, | |
| "learning_rate": 2.878165697386812e-05, | |
| "loss": 1.2994, | |
| "step": 774 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 0.1113181784749031, | |
| "learning_rate": 2.849322119316489e-05, | |
| "loss": 1.1914, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 0.10062463581562042, | |
| "learning_rate": 2.82059976766354e-05, | |
| "loss": 1.1165, | |
| "step": 776 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 0.10819394886493683, | |
| "learning_rate": 2.7919991293619098e-05, | |
| "loss": 1.283, | |
| "step": 777 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 0.1365766078233719, | |
| "learning_rate": 2.763520689282122e-05, | |
| "loss": 1.1963, | |
| "step": 778 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 0.11200159043073654, | |
| "learning_rate": 2.7351649302230553e-05, | |
| "loss": 1.1885, | |
| "step": 779 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 0.11218269914388657, | |
| "learning_rate": 2.7069323329037634e-05, | |
| "loss": 1.2944, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 0.12175281345844269, | |
| "learning_rate": 2.678823375955314e-05, | |
| "loss": 1.2379, | |
| "step": 781 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 0.125010147690773, | |
| "learning_rate": 2.650838535912692e-05, | |
| "loss": 1.2309, | |
| "step": 782 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 0.12338108569383621, | |
| "learning_rate": 2.6229782872067045e-05, | |
| "loss": 1.1489, | |
| "step": 783 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 0.11083382368087769, | |
| "learning_rate": 2.595243102155951e-05, | |
| "loss": 1.1595, | |
| "step": 784 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 0.11741047352552414, | |
| "learning_rate": 2.5676334509588008e-05, | |
| "loss": 1.2438, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 0.10495740175247192, | |
| "learning_rate": 2.540149801685441e-05, | |
| "loss": 1.299, | |
| "step": 786 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 0.11125294119119644, | |
| "learning_rate": 2.5127926202699238e-05, | |
| "loss": 1.1244, | |
| "step": 787 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 0.10036681592464447, | |
| "learning_rate": 2.485562370502279e-05, | |
| "loss": 1.2892, | |
| "step": 788 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 0.11835762113332748, | |
| "learning_rate": 2.4584595140206458e-05, | |
| "loss": 1.1257, | |
| "step": 789 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 0.11104938387870789, | |
| "learning_rate": 2.4314845103034456e-05, | |
| "loss": 1.1601, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 0.10659179091453552, | |
| "learning_rate": 2.40463781666159e-05, | |
| "loss": 1.2066, | |
| "step": 791 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 0.10941512882709503, | |
| "learning_rate": 2.377919888230744e-05, | |
| "loss": 1.3887, | |
| "step": 792 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 0.18816524744033813, | |
| "learning_rate": 2.3513311779635905e-05, | |
| "loss": 1.2191, | |
| "step": 793 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 0.1119375005364418, | |
| "learning_rate": 2.3248721366221638e-05, | |
| "loss": 1.2773, | |
| "step": 794 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 0.10966973006725311, | |
| "learning_rate": 2.2985432127701946e-05, | |
| "loss": 1.2283, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 0.10751962661743164, | |
| "learning_rate": 2.2723448527655266e-05, | |
| "loss": 1.375, | |
| "step": 796 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 0.19195051491260529, | |
| "learning_rate": 2.246277500752524e-05, | |
| "loss": 1.3717, | |
| "step": 797 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 0.4861750900745392, | |
| "learning_rate": 2.220341598654565e-05, | |
| "loss": 1.1786, | |
| "step": 798 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 0.10594101250171661, | |
| "learning_rate": 2.194537586166532e-05, | |
| "loss": 1.2528, | |
| "step": 799 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 0.11542479693889618, | |
| "learning_rate": 2.16886590074737e-05, | |
| "loss": 1.2277, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 0.10653986781835556, | |
| "learning_rate": 2.143326977612662e-05, | |
| "loss": 1.1785, | |
| "step": 801 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 0.11760350316762924, | |
| "learning_rate": 2.1179212497272584e-05, | |
| "loss": 1.2573, | |
| "step": 802 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 0.1091405376791954, | |
| "learning_rate": 2.092649147797927e-05, | |
| "loss": 1.0923, | |
| "step": 803 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 0.10867162048816681, | |
| "learning_rate": 2.0675111002660562e-05, | |
| "loss": 1.2358, | |
| "step": 804 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 0.10017253458499908, | |
| "learning_rate": 2.0425075333003953e-05, | |
| "loss": 1.2741, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 0.11259835958480835, | |
| "learning_rate": 2.0176388707898274e-05, | |
| "loss": 1.0486, | |
| "step": 806 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 0.1280423402786255, | |
| "learning_rate": 1.992905534336178e-05, | |
| "loss": 1.3091, | |
| "step": 807 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 0.10603731125593185, | |
| "learning_rate": 1.9683079432470775e-05, | |
| "loss": 1.2541, | |
| "step": 808 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 0.10933735966682434, | |
| "learning_rate": 1.9438465145288374e-05, | |
| "loss": 1.2186, | |
| "step": 809 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 0.11166153848171234, | |
| "learning_rate": 1.9195216628793956e-05, | |
| "loss": 1.2563, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 0.11184462904930115, | |
| "learning_rate": 1.8953338006812805e-05, | |
| "loss": 1.3052, | |
| "step": 811 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 0.11681969463825226, | |
| "learning_rate": 1.8712833379946215e-05, | |
| "loss": 1.1702, | |
| "step": 812 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 0.11982112377882004, | |
| "learning_rate": 1.847370682550187e-05, | |
| "loss": 1.3025, | |
| "step": 813 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 0.11721136420965195, | |
| "learning_rate": 1.82359623974249e-05, | |
| "loss": 1.1182, | |
| "step": 814 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 0.11441758275032043, | |
| "learning_rate": 1.799960412622904e-05, | |
| "loss": 1.1366, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 0.11492711305618286, | |
| "learning_rate": 1.776463601892825e-05, | |
| "loss": 1.1784, | |
| "step": 816 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 0.11135811358690262, | |
| "learning_rate": 1.753106205896895e-05, | |
| "loss": 1.2411, | |
| "step": 817 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 0.0985221266746521, | |
| "learning_rate": 1.729888620616228e-05, | |
| "loss": 1.283, | |
| "step": 818 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 0.10769841074943542, | |
| "learning_rate": 1.7068112396617163e-05, | |
| "loss": 1.2442, | |
| "step": 819 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 0.09943833947181702, | |
| "learning_rate": 1.683874454267349e-05, | |
| "loss": 1.1121, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 0.11267413944005966, | |
| "learning_rate": 1.6610786532835774e-05, | |
| "loss": 1.2912, | |
| "step": 821 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 0.10905832797288895, | |
| "learning_rate": 1.6384242231707204e-05, | |
| "loss": 1.1264, | |
| "step": 822 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 0.11152827739715576, | |
| "learning_rate": 1.6159115479924257e-05, | |
| "loss": 1.0625, | |
| "step": 823 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 0.10507339239120483, | |
| "learning_rate": 1.593541009409143e-05, | |
| "loss": 1.282, | |
| "step": 824 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 0.11833300441503525, | |
| "learning_rate": 1.5713129866716648e-05, | |
| "loss": 1.0683, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 0.10734719038009644, | |
| "learning_rate": 1.5492278566146946e-05, | |
| "loss": 1.2312, | |
| "step": 826 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 0.09920140355825424, | |
| "learning_rate": 1.5272859936504514e-05, | |
| "loss": 1.3253, | |
| "step": 827 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 0.10309410840272903, | |
| "learning_rate": 1.5054877697623304e-05, | |
| "loss": 1.3389, | |
| "step": 828 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 0.09966228157281876, | |
| "learning_rate": 1.4838335544985981e-05, | |
| "loss": 1.1592, | |
| "step": 829 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 0.10500723123550415, | |
| "learning_rate": 1.462323714966114e-05, | |
| "loss": 1.2318, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 0.10912671685218811, | |
| "learning_rate": 1.4409586158241272e-05, | |
| "loss": 1.2195, | |
| "step": 831 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 0.1132998913526535, | |
| "learning_rate": 1.4197386192780715e-05, | |
| "loss": 1.2476, | |
| "step": 832 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 0.11414579302072525, | |
| "learning_rate": 1.3986640850734444e-05, | |
| "loss": 1.229, | |
| "step": 833 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 0.10963235795497894, | |
| "learning_rate": 1.3777353704897e-05, | |
| "loss": 1.1468, | |
| "step": 834 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 0.11412128806114197, | |
| "learning_rate": 1.3569528303341927e-05, | |
| "loss": 1.141, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 0.11492784321308136, | |
| "learning_rate": 1.3363168169361573e-05, | |
| "loss": 1.0967, | |
| "step": 836 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 0.11600396037101746, | |
| "learning_rate": 1.3158276801407431e-05, | |
| "loss": 1.2414, | |
| "step": 837 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 0.11931241303682327, | |
| "learning_rate": 1.2954857673030807e-05, | |
| "loss": 1.2318, | |
| "step": 838 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 0.10490193963050842, | |
| "learning_rate": 1.2752914232823942e-05, | |
| "loss": 1.2481, | |
| "step": 839 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 0.11982220411300659, | |
| "learning_rate": 1.2552449904361529e-05, | |
| "loss": 1.0291, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 0.0949595496058464, | |
| "learning_rate": 1.235346808614264e-05, | |
| "loss": 1.0835, | |
| "step": 841 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 0.11206669360399246, | |
| "learning_rate": 1.2155972151533224e-05, | |
| "loss": 1.2174, | |
| "step": 842 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 0.10925093293190002, | |
| "learning_rate": 1.1959965448708733e-05, | |
| "loss": 1.359, | |
| "step": 843 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 0.10597091913223267, | |
| "learning_rate": 1.1765451300597573e-05, | |
| "loss": 1.2567, | |
| "step": 844 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 0.12670493125915527, | |
| "learning_rate": 1.1572433004824635e-05, | |
| "loss": 1.1798, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 0.10978806763887405, | |
| "learning_rate": 1.1380913833655382e-05, | |
| "loss": 1.2303, | |
| "step": 846 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 0.1855391412973404, | |
| "learning_rate": 1.1190897033940461e-05, | |
| "loss": 1.2306, | |
| "step": 847 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 0.11130934208631516, | |
| "learning_rate": 1.1002385827060602e-05, | |
| "loss": 1.373, | |
| "step": 848 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 0.10244546085596085, | |
| "learning_rate": 1.0815383408871982e-05, | |
| "loss": 1.3044, | |
| "step": 849 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 0.11533805727958679, | |
| "learning_rate": 1.0629892949652132e-05, | |
| "loss": 1.0704, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 0.12121516466140747, | |
| "learning_rate": 1.0445917594046072e-05, | |
| "loss": 1.265, | |
| "step": 851 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 0.1047421246767044, | |
| "learning_rate": 1.026346046101312e-05, | |
| "loss": 1.1294, | |
| "step": 852 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 0.10979347676038742, | |
| "learning_rate": 1.0082524643773916e-05, | |
| "loss": 0.9608, | |
| "step": 853 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 0.12287425249814987, | |
| "learning_rate": 9.903113209758096e-06, | |
| "loss": 0.8795, | |
| "step": 854 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 0.11486369371414185, | |
| "learning_rate": 9.725229200552122e-06, | |
| "loss": 1.3881, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 0.11395483464002609, | |
| "learning_rate": 9.548875631847875e-06, | |
| "loss": 1.1552, | |
| "step": 856 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 0.10814516991376877, | |
| "learning_rate": 9.374055493391455e-06, | |
| "loss": 1.2155, | |
| "step": 857 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 0.10679764300584793, | |
| "learning_rate": 9.200771748932513e-06, | |
| "loss": 1.1449, | |
| "step": 858 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 0.1089312732219696, | |
| "learning_rate": 9.029027336174022e-06, | |
| "loss": 1.2221, | |
| "step": 859 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 0.11471953988075256, | |
| "learning_rate": 8.858825166722417e-06, | |
| "loss": 0.9955, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 0.11357836425304413, | |
| "learning_rate": 8.6901681260383e-06, | |
| "loss": 1.0855, | |
| "step": 861 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 0.11274585127830505, | |
| "learning_rate": 8.523059073387474e-06, | |
| "loss": 1.1812, | |
| "step": 862 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 0.10156439989805222, | |
| "learning_rate": 8.35750084179251e-06, | |
| "loss": 1.1702, | |
| "step": 863 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 0.10595791786909103, | |
| "learning_rate": 8.193496237984677e-06, | |
| "loss": 1.1719, | |
| "step": 864 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 0.10934492200613022, | |
| "learning_rate": 8.031048042356392e-06, | |
| "loss": 1.1735, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 0.10104786604642868, | |
| "learning_rate": 7.870159008914069e-06, | |
| "loss": 1.1967, | |
| "step": 866 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 0.10521814227104187, | |
| "learning_rate": 7.710831865231461e-06, | |
| "loss": 1.3094, | |
| "step": 867 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 0.1166355162858963, | |
| "learning_rate": 7.553069312403349e-06, | |
| "loss": 1.2912, | |
| "step": 868 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 0.09746871888637543, | |
| "learning_rate": 7.3968740249998115e-06, | |
| "loss": 1.1133, | |
| "step": 869 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 0.10336558520793915, | |
| "learning_rate": 7.242248651020844e-06, | |
| "loss": 1.1893, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 0.11591080576181412, | |
| "learning_rate": 7.089195811851501e-06, | |
| "loss": 1.1395, | |
| "step": 871 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 0.22089900076389313, | |
| "learning_rate": 6.937718102217461e-06, | |
| "loss": 1.1993, | |
| "step": 872 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 0.108499675989151, | |
| "learning_rate": 6.787818090140985e-06, | |
| "loss": 1.1195, | |
| "step": 873 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 0.10180932283401489, | |
| "learning_rate": 6.639498316897419e-06, | |
| "loss": 1.3188, | |
| "step": 874 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 0.10950213670730591, | |
| "learning_rate": 6.492761296972116e-06, | |
| "loss": 1.1373, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 0.09740964323282242, | |
| "learning_rate": 6.347609518017761e-06, | |
| "loss": 1.1551, | |
| "step": 876 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 0.10599476099014282, | |
| "learning_rate": 6.2040454408123e-06, | |
| "loss": 1.2161, | |
| "step": 877 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 0.11540497094392776, | |
| "learning_rate": 6.062071499217081e-06, | |
| "loss": 1.256, | |
| "step": 878 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 0.10363679379224777, | |
| "learning_rate": 5.921690100135713e-06, | |
| "loss": 1.2462, | |
| "step": 879 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 0.10643389075994492, | |
| "learning_rate": 5.782903623473201e-06, | |
| "loss": 1.34, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 0.14184726774692535, | |
| "learning_rate": 5.645714422095627e-06, | |
| "loss": 1.0978, | |
| "step": 881 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 0.11375614255666733, | |
| "learning_rate": 5.510124821790208e-06, | |
| "loss": 1.2334, | |
| "step": 882 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 0.11842206120491028, | |
| "learning_rate": 5.376137121225933e-06, | |
| "loss": 1.2043, | |
| "step": 883 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 0.1027233749628067, | |
| "learning_rate": 5.24375359191458e-06, | |
| "loss": 1.2038, | |
| "step": 884 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 0.09779934585094452, | |
| "learning_rate": 5.11297647817216e-06, | |
| "loss": 1.2799, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 0.28459811210632324, | |
| "learning_rate": 4.983807997080925e-06, | |
| "loss": 1.2972, | |
| "step": 886 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 0.12350057065486908, | |
| "learning_rate": 4.856250338451762e-06, | |
| "loss": 1.1445, | |
| "step": 887 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 0.10094188153743744, | |
| "learning_rate": 4.73030566478706e-06, | |
| "loss": 1.0144, | |
| "step": 888 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 0.10270640254020691, | |
| "learning_rate": 4.605976111244015e-06, | |
| "loss": 1.1673, | |
| "step": 889 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 0.12842753529548645, | |
| "learning_rate": 4.483263785598524e-06, | |
| "loss": 1.1681, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 0.10717935115098953, | |
| "learning_rate": 4.362170768209406e-06, | |
| "loss": 1.2661, | |
| "step": 891 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 0.20119836926460266, | |
| "learning_rate": 4.24269911198305e-06, | |
| "loss": 1.1492, | |
| "step": 892 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 0.11531922966241837, | |
| "learning_rate": 4.124850842338779e-06, | |
| "loss": 1.2629, | |
| "step": 893 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 0.10086618363857269, | |
| "learning_rate": 4.008627957174372e-06, | |
| "loss": 1.1223, | |
| "step": 894 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 0.10228842496871948, | |
| "learning_rate": 3.8940324268322285e-06, | |
| "loss": 1.2285, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 0.1152658760547638, | |
| "learning_rate": 3.7810661940660517e-06, | |
| "loss": 1.3102, | |
| "step": 896 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 0.09905364364385605, | |
| "learning_rate": 3.66973117400774e-06, | |
| "loss": 1.2029, | |
| "step": 897 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 0.11741360276937485, | |
| "learning_rate": 3.5600292541351e-06, | |
| "loss": 1.0504, | |
| "step": 898 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 0.10663048177957535, | |
| "learning_rate": 3.451962294239741e-06, | |
| "loss": 1.0434, | |
| "step": 899 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 0.13997888565063477, | |
| "learning_rate": 3.3455321263955786e-06, | |
| "loss": 1.1909, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 0.10722199827432632, | |
| "learning_rate": 3.240740554927768e-06, | |
| "loss": 1.1066, | |
| "step": 901 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 0.10431613773107529, | |
| "learning_rate": 3.1375893563820756e-06, | |
| "loss": 1.1097, | |
| "step": 902 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 0.10055112838745117, | |
| "learning_rate": 3.0360802794948686e-06, | |
| "loss": 1.119, | |
| "step": 903 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 0.14287303388118744, | |
| "learning_rate": 2.936215045163371e-06, | |
| "loss": 1.1977, | |
| "step": 904 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 0.11945986747741699, | |
| "learning_rate": 2.837995346416533e-06, | |
| "loss": 1.0618, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 0.11288944631814957, | |
| "learning_rate": 2.7414228483862657e-06, | |
| "loss": 1.1562, | |
| "step": 906 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 0.11274345964193344, | |
| "learning_rate": 2.6464991882793278e-06, | |
| "loss": 1.2794, | |
| "step": 907 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 0.10257246345281601, | |
| "learning_rate": 2.5532259753494827e-06, | |
| "loss": 1.1981, | |
| "step": 908 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 0.11306274682283401, | |
| "learning_rate": 2.4616047908702087e-06, | |
| "loss": 1.2817, | |
| "step": 909 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 0.1063089668750763, | |
| "learning_rate": 2.3716371881079557e-06, | |
| "loss": 1.2429, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 0.11371003836393356, | |
| "learning_rate": 2.2833246922957407e-06, | |
| "loss": 1.2017, | |
| "step": 911 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 0.10964678227901459, | |
| "learning_rate": 2.1966688006073823e-06, | |
| "loss": 1.0438, | |
| "step": 912 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 0.10942777991294861, | |
| "learning_rate": 2.111670982132008e-06, | |
| "loss": 1.1411, | |
| "step": 913 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 0.20406045019626617, | |
| "learning_rate": 2.028332677849254e-06, | |
| "loss": 1.26, | |
| "step": 914 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 0.1080511286854744, | |
| "learning_rate": 1.9466553006047383e-06, | |
| "loss": 1.2078, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 0.10461273789405823, | |
| "learning_rate": 1.86664023508617e-06, | |
| "loss": 1.0991, | |
| "step": 916 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 0.14060625433921814, | |
| "learning_rate": 1.7882888377998786e-06, | |
| "loss": 1.1999, | |
| "step": 917 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 0.10702642798423767, | |
| "learning_rate": 1.7116024370477878e-06, | |
| "loss": 1.1649, | |
| "step": 918 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 0.12004446983337402, | |
| "learning_rate": 1.6365823329049124e-06, | |
| "loss": 1.1923, | |
| "step": 919 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 0.12491834163665771, | |
| "learning_rate": 1.5632297971972965e-06, | |
| "loss": 1.1329, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 0.11271212995052338, | |
| "learning_rate": 1.4915460734805097e-06, | |
| "loss": 1.1198, | |
| "step": 921 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 0.11066664755344391, | |
| "learning_rate": 1.4215323770184641e-06, | |
| "loss": 1.2727, | |
| "step": 922 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 0.12216173112392426, | |
| "learning_rate": 1.3531898947629296e-06, | |
| "loss": 1.1179, | |
| "step": 923 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 0.10309845954179764, | |
| "learning_rate": 1.2865197853333178e-06, | |
| "loss": 1.328, | |
| "step": 924 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 0.1231243684887886, | |
| "learning_rate": 1.221523178997075e-06, | |
| "loss": 1.1911, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 0.12335013598203659, | |
| "learning_rate": 1.158201177650553e-06, | |
| "loss": 1.2257, | |
| "step": 926 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 0.11128874868154526, | |
| "learning_rate": 1.0965548548002803e-06, | |
| "loss": 1.2306, | |
| "step": 927 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 0.10876341909170151, | |
| "learning_rate": 1.036585255544764e-06, | |
| "loss": 1.2044, | |
| "step": 928 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 0.12024974822998047, | |
| "learning_rate": 9.782933965567954e-07, | |
| "loss": 1.1953, | |
| "step": 929 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 0.11236795783042908, | |
| "learning_rate": 9.21680266066216e-07, | |
| "loss": 1.3033, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 0.11030973494052887, | |
| "learning_rate": 8.667468238431453e-07, | |
| "loss": 1.2, | |
| "step": 931 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 0.10306066274642944, | |
| "learning_rate": 8.13494001181725e-07, | |
| "loss": 1.3141, | |
| "step": 932 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 0.10356069356203079, | |
| "learning_rate": 7.619227008843322e-07, | |
| "loss": 1.2244, | |
| "step": 933 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 0.10682324320077896, | |
| "learning_rate": 7.120337972462365e-07, | |
| "loss": 1.1282, | |
| "step": 934 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 0.11171166598796844, | |
| "learning_rate": 6.638281360408339e-07, | |
| "loss": 1.1534, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 0.09517676383256912, | |
| "learning_rate": 6.173065345052687e-07, | |
| "loss": 1.1292, | |
| "step": 936 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 0.11539721488952637, | |
| "learning_rate": 5.724697813265901e-07, | |
| "loss": 1.1716, | |
| "step": 937 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 0.11124927550554276, | |
| "learning_rate": 5.293186366284176e-07, | |
| "loss": 1.1894, | |
| "step": 938 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 0.261902779340744, | |
| "learning_rate": 4.878538319579629e-07, | |
| "loss": 1.2252, | |
| "step": 939 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 0.11507264524698257, | |
| "learning_rate": 4.4807607027372855e-07, | |
| "loss": 1.2918, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 0.1210726797580719, | |
| "learning_rate": 4.0998602593352863e-07, | |
| "loss": 1.1893, | |
| "step": 941 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 0.11546792089939117, | |
| "learning_rate": 3.7358434468308667e-07, | |
| "loss": 1.1486, | |
| "step": 942 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 0.12196337431669235, | |
| "learning_rate": 3.388716436450667e-07, | |
| "loss": 1.3495, | |
| "step": 943 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 0.10631462186574936, | |
| "learning_rate": 3.0584851130861513e-07, | |
| "loss": 1.1866, | |
| "step": 944 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 0.14002542197704315, | |
| "learning_rate": 2.745155075194128e-07, | |
| "loss": 0.9632, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 0.15344469249248505, | |
| "learning_rate": 2.448731634701384e-07, | |
| "loss": 1.3407, | |
| "step": 946 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 0.11114881932735443, | |
| "learning_rate": 2.169219816914869e-07, | |
| "loss": 1.1723, | |
| "step": 947 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 0.09970024973154068, | |
| "learning_rate": 1.9066243604367594e-07, | |
| "loss": 1.3012, | |
| "step": 948 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 0.10021547228097916, | |
| "learning_rate": 1.6609497170834155e-07, | |
| "loss": 1.1621, | |
| "step": 949 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 0.11028821766376495, | |
| "learning_rate": 1.4322000518106616e-07, | |
| "loss": 1.092, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 0.10497645288705826, | |
| "learning_rate": 1.220379242642844e-07, | |
| "loss": 1.1194, | |
| "step": 951 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 0.10754618048667908, | |
| "learning_rate": 1.025490880606883e-07, | |
| "loss": 1.0586, | |
| "step": 952 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 0.1118793711066246, | |
| "learning_rate": 8.475382696719881e-08, | |
| "loss": 1.1977, | |
| "step": 953 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 0.11205057799816132, | |
| "learning_rate": 6.865244266928139e-08, | |
| "loss": 0.944, | |
| "step": 954 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 0.10550331324338913, | |
| "learning_rate": 5.424520813590572e-08, | |
| "loss": 1.1463, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 0.10324594378471375, | |
| "learning_rate": 4.153236761488266e-08, | |
| "loss": 1.1325, | |
| "step": 956 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 0.10522052645683289, | |
| "learning_rate": 3.051413662874536e-08, | |
| "loss": 1.1922, | |
| "step": 957 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 0.12036368250846863, | |
| "learning_rate": 2.1190701971052218e-08, | |
| "loss": 1.123, | |
| "step": 958 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 0.11026199162006378, | |
| "learning_rate": 1.3562221703267152e-08, | |
| "loss": 1.261, | |
| "step": 959 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 0.28202083706855774, | |
| "learning_rate": 7.628825152050657e-09, | |
| "loss": 1.2547, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 0.10787337273359299, | |
| "learning_rate": 3.390612907094859e-09, | |
| "loss": 1.2832, | |
| "step": 961 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 0.10484518110752106, | |
| "learning_rate": 8.476568193804824e-10, | |
| "loss": 1.2884, | |
| "step": 962 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 0.11853493005037308, | |
| "learning_rate": 0.0, | |
| "loss": 1.2526, | |
| "step": 963 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "step": 963, | |
| "total_flos": 1.3034774175330337e+19, | |
| "train_loss": 1.2195818047285822, | |
| "train_runtime": 50845.1793, | |
| "train_samples_per_second": 0.152, | |
| "train_steps_per_second": 0.019 | |
| } | |
| ], | |
| "logging_steps": 1.0, | |
| "max_steps": 963, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 100, | |
| "total_flos": 1.3034774175330337e+19, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |