| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.26737967914438504, | |
| "eval_steps": 500, | |
| "global_step": 500, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0005347593582887701, | |
| "grad_norm": 10.25, | |
| "learning_rate": 0.0002, | |
| "loss": 5.5244, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.0010695187165775401, | |
| "grad_norm": 5.65625, | |
| "learning_rate": 0.0002, | |
| "loss": 1.5152, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.0016042780748663102, | |
| "grad_norm": 5.0, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9678, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.0021390374331550803, | |
| "grad_norm": 1.09375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2586, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.00267379679144385, | |
| "grad_norm": 2.21875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8375, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.0032085561497326204, | |
| "grad_norm": 2.359375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8124, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.0037433155080213902, | |
| "grad_norm": 1.3515625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4274, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.0042780748663101605, | |
| "grad_norm": 1.9765625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6398, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.004812834224598931, | |
| "grad_norm": 1.1796875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4485, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.0053475935828877, | |
| "grad_norm": 0.48828125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3639, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.0058823529411764705, | |
| "grad_norm": 0.7265625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3139, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.006417112299465241, | |
| "grad_norm": 1.3671875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1684, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.006951871657754011, | |
| "grad_norm": 1.359375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4753, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.0074866310160427805, | |
| "grad_norm": 0.56640625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0907, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.008021390374331552, | |
| "grad_norm": 1.9375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3021, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.008556149732620321, | |
| "grad_norm": 1.421875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.292, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.00909090909090909, | |
| "grad_norm": 0.4453125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2465, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.009625668449197862, | |
| "grad_norm": 1.0703125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2335, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.010160427807486631, | |
| "grad_norm": 4.84375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2976, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.0106951871657754, | |
| "grad_norm": 4.84375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6077, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.011229946524064172, | |
| "grad_norm": 2.25, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3396, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.011764705882352941, | |
| "grad_norm": 1.4453125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.218, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.01229946524064171, | |
| "grad_norm": 1.1640625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0862, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.012834224598930482, | |
| "grad_norm": 2.65625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4492, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.013368983957219251, | |
| "grad_norm": 1.1796875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1986, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.013903743315508022, | |
| "grad_norm": 1.03125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2902, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.014438502673796792, | |
| "grad_norm": 1.5703125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.6038, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.014973262032085561, | |
| "grad_norm": 1.359375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2869, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.015508021390374332, | |
| "grad_norm": 0.455078125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2646, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.016042780748663103, | |
| "grad_norm": 0.45703125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1653, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.016577540106951873, | |
| "grad_norm": 0.84765625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2535, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.017112299465240642, | |
| "grad_norm": 1.1015625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2485, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.01764705882352941, | |
| "grad_norm": 0.74609375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2425, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.01818181818181818, | |
| "grad_norm": 0.353515625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0913, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.01871657754010695, | |
| "grad_norm": 0.546875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1082, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.019251336898395723, | |
| "grad_norm": 1.0, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4729, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.019786096256684493, | |
| "grad_norm": 1.265625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3255, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.020320855614973262, | |
| "grad_norm": 2.734375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4583, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.02085561497326203, | |
| "grad_norm": 0.203125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1099, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.0213903743315508, | |
| "grad_norm": 0.71875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1941, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.021925133689839574, | |
| "grad_norm": 0.64453125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.278, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.022459893048128343, | |
| "grad_norm": 0.69921875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4062, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.022994652406417113, | |
| "grad_norm": 1.2109375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4801, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.023529411764705882, | |
| "grad_norm": 0.3515625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2161, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.02406417112299465, | |
| "grad_norm": 1.109375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2423, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.02459893048128342, | |
| "grad_norm": 0.69921875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.208, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.025133689839572194, | |
| "grad_norm": 0.37890625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2149, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.025668449197860963, | |
| "grad_norm": 0.6796875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2528, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.026203208556149733, | |
| "grad_norm": 0.2890625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2094, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.026737967914438502, | |
| "grad_norm": 0.228515625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1544, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.02727272727272727, | |
| "grad_norm": 0.62109375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2201, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.027807486631016044, | |
| "grad_norm": 2.859375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3119, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.028342245989304814, | |
| "grad_norm": 0.578125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2003, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.028877005347593583, | |
| "grad_norm": 0.80078125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2708, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.029411764705882353, | |
| "grad_norm": 1.515625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2502, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.029946524064171122, | |
| "grad_norm": 0.6796875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2379, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.03048128342245989, | |
| "grad_norm": 0.4140625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.031016042780748664, | |
| "grad_norm": 0.271484375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0705, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.03155080213903743, | |
| "grad_norm": 0.51953125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2252, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.03208556149732621, | |
| "grad_norm": 1.8359375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2479, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.032620320855614976, | |
| "grad_norm": 1.25, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3067, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.033155080213903745, | |
| "grad_norm": 1.390625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2026, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.033689839572192515, | |
| "grad_norm": 0.255859375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1771, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.034224598930481284, | |
| "grad_norm": 1.2421875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1886, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.034759358288770054, | |
| "grad_norm": 0.59765625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1603, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.03529411764705882, | |
| "grad_norm": 0.298828125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1577, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.03582887700534759, | |
| "grad_norm": 0.48046875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2385, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.03636363636363636, | |
| "grad_norm": 0.291015625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2457, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.03689839572192513, | |
| "grad_norm": 1.203125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2403, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.0374331550802139, | |
| "grad_norm": 1.34375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2959, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.03796791443850268, | |
| "grad_norm": 0.47265625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2408, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.038502673796791446, | |
| "grad_norm": 2.0625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5305, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.039037433155080216, | |
| "grad_norm": 0.87109375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1871, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.039572192513368985, | |
| "grad_norm": 0.98046875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2837, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.040106951871657755, | |
| "grad_norm": 0.515625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.133, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.040641711229946524, | |
| "grad_norm": 0.59375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2151, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.041176470588235294, | |
| "grad_norm": 0.4765625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1379, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.04171122994652406, | |
| "grad_norm": 0.62109375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1798, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.04224598930481283, | |
| "grad_norm": 0.396484375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1735, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.0427807486631016, | |
| "grad_norm": 0.546875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2279, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.04331550802139037, | |
| "grad_norm": 0.455078125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2083, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.04385026737967915, | |
| "grad_norm": 0.453125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2038, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.04438502673796792, | |
| "grad_norm": 0.259765625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2185, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.044919786096256686, | |
| "grad_norm": 0.1044921875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1184, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.045454545454545456, | |
| "grad_norm": 0.21484375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1767, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.045989304812834225, | |
| "grad_norm": 0.10107421875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1461, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.046524064171122995, | |
| "grad_norm": 0.283203125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1723, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.047058823529411764, | |
| "grad_norm": 0.314453125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1796, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.04759358288770053, | |
| "grad_norm": 0.98828125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1747, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.0481283422459893, | |
| "grad_norm": 0.419921875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1606, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.04866310160427807, | |
| "grad_norm": 0.1640625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1675, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.04919786096256684, | |
| "grad_norm": 2.203125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3444, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.04973262032085562, | |
| "grad_norm": 0.326171875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2217, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.05026737967914439, | |
| "grad_norm": 0.27734375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2684, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.05080213903743316, | |
| "grad_norm": 1.4921875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2776, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.051336898395721926, | |
| "grad_norm": 0.10107421875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1626, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.051871657754010696, | |
| "grad_norm": 0.6953125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1733, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.052406417112299465, | |
| "grad_norm": 0.46875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2631, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.052941176470588235, | |
| "grad_norm": 5.78125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3128, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.053475935828877004, | |
| "grad_norm": 1.7421875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3108, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.05401069518716577, | |
| "grad_norm": 0.87890625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4231, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.05454545454545454, | |
| "grad_norm": 0.171875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1636, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.05508021390374331, | |
| "grad_norm": 0.8203125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4548, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.05561497326203209, | |
| "grad_norm": 0.208984375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1718, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.05614973262032086, | |
| "grad_norm": 0.28125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1438, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.05668449197860963, | |
| "grad_norm": 0.55078125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2534, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.0572192513368984, | |
| "grad_norm": 0.412109375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2303, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.057754010695187166, | |
| "grad_norm": 0.322265625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2488, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.058288770053475936, | |
| "grad_norm": 0.85546875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2962, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.058823529411764705, | |
| "grad_norm": 0.357421875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2266, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.059358288770053474, | |
| "grad_norm": 0.236328125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2208, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.059893048128342244, | |
| "grad_norm": 0.56640625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2489, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.06042780748663101, | |
| "grad_norm": 0.2490234375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3115, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.06096256684491978, | |
| "grad_norm": 0.498046875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2222, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.06149732620320856, | |
| "grad_norm": 1.2421875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.389, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.06203208556149733, | |
| "grad_norm": 0.58984375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2148, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.06256684491978609, | |
| "grad_norm": 0.3515625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2527, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.06310160427807486, | |
| "grad_norm": 0.341796875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2395, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.06363636363636363, | |
| "grad_norm": 1.1171875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3902, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.06417112299465241, | |
| "grad_norm": 0.46484375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2973, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.06470588235294118, | |
| "grad_norm": 0.671875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2476, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.06524064171122995, | |
| "grad_norm": 0.462890625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.186, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.06577540106951872, | |
| "grad_norm": 0.244140625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.165, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.06631016042780749, | |
| "grad_norm": 1.546875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5423, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.06684491978609626, | |
| "grad_norm": 0.95703125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4157, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.06737967914438503, | |
| "grad_norm": 3.40625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5118, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.0679144385026738, | |
| "grad_norm": 0.5546875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1738, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.06844919786096257, | |
| "grad_norm": 0.1396484375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0812, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.06898395721925134, | |
| "grad_norm": 0.2392578125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1073, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.06951871657754011, | |
| "grad_norm": 0.193359375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1118, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.07005347593582888, | |
| "grad_norm": 1.015625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2042, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.07058823529411765, | |
| "grad_norm": 0.400390625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1226, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.07112299465240642, | |
| "grad_norm": 0.234375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1684, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.07165775401069518, | |
| "grad_norm": 0.1630859375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1795, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.07219251336898395, | |
| "grad_norm": 0.60546875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2594, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.07272727272727272, | |
| "grad_norm": 0.138671875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0331, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.0732620320855615, | |
| "grad_norm": 0.193359375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1815, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.07379679144385026, | |
| "grad_norm": 0.9375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.458, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.07433155080213903, | |
| "grad_norm": 0.5859375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.266, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.0748663101604278, | |
| "grad_norm": 0.5859375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2798, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.07540106951871657, | |
| "grad_norm": 0.494140625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2428, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.07593582887700535, | |
| "grad_norm": 0.2060546875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1931, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.07647058823529412, | |
| "grad_norm": 0.49609375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2107, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.07700534759358289, | |
| "grad_norm": 0.2490234375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1649, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.07754010695187166, | |
| "grad_norm": 0.1396484375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1408, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.07807486631016043, | |
| "grad_norm": 1.5234375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3663, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.0786096256684492, | |
| "grad_norm": 0.70703125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2539, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.07914438502673797, | |
| "grad_norm": 0.51171875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2954, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.07967914438502674, | |
| "grad_norm": 0.404296875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3071, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.08021390374331551, | |
| "grad_norm": 0.3671875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1121, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.08074866310160428, | |
| "grad_norm": 0.251953125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1059, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.08128342245989305, | |
| "grad_norm": 0.490234375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2265, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.08181818181818182, | |
| "grad_norm": 0.36328125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1875, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.08235294117647059, | |
| "grad_norm": 0.51171875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2944, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.08288770053475936, | |
| "grad_norm": 0.6796875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2663, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.08342245989304813, | |
| "grad_norm": 0.58984375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1909, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.0839572192513369, | |
| "grad_norm": 0.24609375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1816, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.08449197860962566, | |
| "grad_norm": 0.267578125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0903, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.08502673796791443, | |
| "grad_norm": 0.8359375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1757, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.0855614973262032, | |
| "grad_norm": 0.283203125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1636, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.08609625668449197, | |
| "grad_norm": 0.2451171875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1363, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.08663101604278074, | |
| "grad_norm": 0.318359375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1262, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.08716577540106951, | |
| "grad_norm": 0.5, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2162, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.0877005347593583, | |
| "grad_norm": 0.1123046875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0586, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.08823529411764706, | |
| "grad_norm": 0.0771484375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0908, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.08877005347593583, | |
| "grad_norm": 0.486328125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.109, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.0893048128342246, | |
| "grad_norm": 0.2333984375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1283, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.08983957219251337, | |
| "grad_norm": 0.0023651123046875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0003, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.09037433155080214, | |
| "grad_norm": 0.87890625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0709, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.09090909090909091, | |
| "grad_norm": 0.1845703125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1616, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.09144385026737968, | |
| "grad_norm": 0.83203125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1496, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.09197860962566845, | |
| "grad_norm": 1.2421875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3648, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.09251336898395722, | |
| "grad_norm": 1.1484375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4267, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.09304812834224599, | |
| "grad_norm": 1.71875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1812, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.09358288770053476, | |
| "grad_norm": 3.34375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3108, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.09411764705882353, | |
| "grad_norm": 1.21875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2502, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.0946524064171123, | |
| "grad_norm": 0.8671875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3469, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.09518716577540107, | |
| "grad_norm": 0.2353515625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1502, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.09572192513368984, | |
| "grad_norm": 0.859375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3546, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.0962566844919786, | |
| "grad_norm": 0.42578125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2236, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.09679144385026738, | |
| "grad_norm": 0.2109375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0742, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.09732620320855614, | |
| "grad_norm": 0.2421875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2046, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.09786096256684491, | |
| "grad_norm": 0.2060546875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1355, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.09839572192513368, | |
| "grad_norm": 0.40234375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2228, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.09893048128342247, | |
| "grad_norm": 0.8046875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2756, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.09946524064171124, | |
| "grad_norm": 0.357421875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.228, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 0.201171875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.122, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.10053475935828877, | |
| "grad_norm": 0.4296875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1864, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.10106951871657754, | |
| "grad_norm": 0.2099609375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1586, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.10160427807486631, | |
| "grad_norm": 0.2734375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1432, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.10213903743315508, | |
| "grad_norm": 0.2265625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1971, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.10267379679144385, | |
| "grad_norm": 0.5625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2811, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.10320855614973262, | |
| "grad_norm": 0.3203125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1839, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.10374331550802139, | |
| "grad_norm": 0.291015625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2092, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.10427807486631016, | |
| "grad_norm": 0.6640625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2743, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.10481283422459893, | |
| "grad_norm": 0.328125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2752, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.1053475935828877, | |
| "grad_norm": 0.8359375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1735, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.10588235294117647, | |
| "grad_norm": 0.369140625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.137, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.10641711229946524, | |
| "grad_norm": 0.53515625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2498, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.10695187165775401, | |
| "grad_norm": 0.376953125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1308, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.10748663101604278, | |
| "grad_norm": 0.88671875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2849, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.10802139037433155, | |
| "grad_norm": 0.291015625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1342, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.10855614973262032, | |
| "grad_norm": 0.26953125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.105, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.10909090909090909, | |
| "grad_norm": 0.478515625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1172, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.10962566844919786, | |
| "grad_norm": 0.326171875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0579, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.11016042780748662, | |
| "grad_norm": 0.2060546875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1235, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.11069518716577541, | |
| "grad_norm": 1.375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.5436, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.11122994652406418, | |
| "grad_norm": 0.21875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2197, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.11176470588235295, | |
| "grad_norm": 0.75, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2319, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.11229946524064172, | |
| "grad_norm": 0.43359375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1091, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.11283422459893049, | |
| "grad_norm": 0.154296875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1932, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.11336898395721925, | |
| "grad_norm": 0.10009765625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1829, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.11390374331550802, | |
| "grad_norm": 0.0986328125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1738, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.1144385026737968, | |
| "grad_norm": 0.53125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1746, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.11497326203208556, | |
| "grad_norm": 0.484375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1831, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.11550802139037433, | |
| "grad_norm": 1.609375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4379, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.1160427807486631, | |
| "grad_norm": 0.265625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1707, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.11657754010695187, | |
| "grad_norm": 1.5625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1901, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.11711229946524064, | |
| "grad_norm": 0.1904296875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2282, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.11764705882352941, | |
| "grad_norm": 0.1982421875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1478, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.11818181818181818, | |
| "grad_norm": 0.408203125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2262, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 0.11871657754010695, | |
| "grad_norm": 0.271484375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1373, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.11925133689839572, | |
| "grad_norm": 0.34765625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.123, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 0.11978609625668449, | |
| "grad_norm": 0.08837890625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1712, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.12032085561497326, | |
| "grad_norm": 0.4140625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1837, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.12085561497326203, | |
| "grad_norm": 1.171875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3184, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.1213903743315508, | |
| "grad_norm": 0.6640625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1646, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 0.12192513368983957, | |
| "grad_norm": 0.765625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1659, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.12245989304812835, | |
| "grad_norm": 0.126953125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.151, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 0.12299465240641712, | |
| "grad_norm": 0.25, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1841, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.12352941176470589, | |
| "grad_norm": 1.1171875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0834, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.12406417112299466, | |
| "grad_norm": 0.1728515625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1417, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.12459893048128343, | |
| "grad_norm": 0.099609375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1227, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 0.12513368983957218, | |
| "grad_norm": 0.8125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1835, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.12566844919786097, | |
| "grad_norm": 1.2109375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3382, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.12620320855614972, | |
| "grad_norm": 1.4296875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4302, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.1267379679144385, | |
| "grad_norm": 0.546875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1895, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.12727272727272726, | |
| "grad_norm": 2.5, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3654, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.12780748663101604, | |
| "grad_norm": 1.0859375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4818, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 0.12834224598930483, | |
| "grad_norm": 0.65625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3575, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.12887700534759358, | |
| "grad_norm": 0.25, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0789, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 0.12941176470588237, | |
| "grad_norm": 0.4765625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2638, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.12994652406417112, | |
| "grad_norm": 0.267578125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1225, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.1304812834224599, | |
| "grad_norm": 0.6015625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0733, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.13101604278074866, | |
| "grad_norm": 0.2158203125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0766, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.13155080213903744, | |
| "grad_norm": 0.7734375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3096, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.1320855614973262, | |
| "grad_norm": 0.94921875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2276, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 0.13262032085561498, | |
| "grad_norm": 0.345703125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2357, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.13315508021390374, | |
| "grad_norm": 0.134765625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1144, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.13368983957219252, | |
| "grad_norm": 0.2373046875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1118, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.13422459893048128, | |
| "grad_norm": 0.306640625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1682, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 0.13475935828877006, | |
| "grad_norm": 0.193359375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0453, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.13529411764705881, | |
| "grad_norm": 0.388671875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1895, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 0.1358288770053476, | |
| "grad_norm": 0.94921875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4819, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 0.13636363636363635, | |
| "grad_norm": 0.185546875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1763, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.13689839572192514, | |
| "grad_norm": 0.34765625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.268, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.1374331550802139, | |
| "grad_norm": 0.2578125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1276, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 0.13796791443850268, | |
| "grad_norm": 0.2421875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1819, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.13850267379679143, | |
| "grad_norm": 0.076171875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0893, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 0.13903743315508021, | |
| "grad_norm": 0.65234375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2154, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.139572192513369, | |
| "grad_norm": 0.443359375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3956, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 0.14010695187165775, | |
| "grad_norm": 0.150390625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1585, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 0.14064171122994654, | |
| "grad_norm": 0.125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0876, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 0.1411764705882353, | |
| "grad_norm": 0.80078125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0673, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.14171122994652408, | |
| "grad_norm": 4.0625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.184, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.14224598930481283, | |
| "grad_norm": 0.0830078125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1651, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 0.14278074866310161, | |
| "grad_norm": 0.2470703125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0828, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 0.14331550802139037, | |
| "grad_norm": 0.5625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1545, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.14385026737967915, | |
| "grad_norm": 0.30859375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1641, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 0.1443850267379679, | |
| "grad_norm": 0.625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3245, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.1449197860962567, | |
| "grad_norm": 0.45703125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1592, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 0.14545454545454545, | |
| "grad_norm": 0.1162109375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1406, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.14598930481283423, | |
| "grad_norm": 0.40625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1832, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 0.146524064171123, | |
| "grad_norm": 0.099609375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.091, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 0.14705882352941177, | |
| "grad_norm": 0.416015625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0943, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.14759358288770053, | |
| "grad_norm": 2.359375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3355, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.1481283422459893, | |
| "grad_norm": 0.2138671875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1074, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 0.14866310160427806, | |
| "grad_norm": 0.3828125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.157, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 0.14919786096256685, | |
| "grad_norm": 0.357421875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.149, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 0.1497326203208556, | |
| "grad_norm": 0.478515625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2129, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.15026737967914439, | |
| "grad_norm": 0.73828125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3051, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 0.15080213903743314, | |
| "grad_norm": 1.1953125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.147, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.15133689839572192, | |
| "grad_norm": 0.4375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1934, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 0.1518716577540107, | |
| "grad_norm": 0.2392578125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0722, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.15240641711229946, | |
| "grad_norm": 0.2353515625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.062, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.15294117647058825, | |
| "grad_norm": 0.71484375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2352, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.153475935828877, | |
| "grad_norm": 1.0, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2244, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 0.15401069518716579, | |
| "grad_norm": 0.388671875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1342, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.15454545454545454, | |
| "grad_norm": 0.1962890625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.206, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 0.15508021390374332, | |
| "grad_norm": 1.375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3573, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.15561497326203208, | |
| "grad_norm": 0.294921875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1673, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 0.15614973262032086, | |
| "grad_norm": 0.1708984375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0897, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.15668449197860962, | |
| "grad_norm": 0.1572265625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0698, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 0.1572192513368984, | |
| "grad_norm": 0.0849609375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0981, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 0.15775401069518716, | |
| "grad_norm": 0.1162109375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0483, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.15828877005347594, | |
| "grad_norm": 0.1923828125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1486, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 0.1588235294117647, | |
| "grad_norm": 0.146484375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0981, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 0.15935828877005348, | |
| "grad_norm": 0.94140625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.282, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 0.15989304812834224, | |
| "grad_norm": 0.318359375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1445, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 0.16042780748663102, | |
| "grad_norm": 0.25, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1235, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.16096256684491977, | |
| "grad_norm": 0.10693359375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0835, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 0.16149732620320856, | |
| "grad_norm": 0.375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0934, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 0.1620320855614973, | |
| "grad_norm": 0.59765625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3041, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 0.1625668449197861, | |
| "grad_norm": 0.06640625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1435, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 0.16310160427807488, | |
| "grad_norm": 0.66796875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1957, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.16363636363636364, | |
| "grad_norm": 0.84765625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3342, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 0.16417112299465242, | |
| "grad_norm": 0.7265625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3299, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 0.16470588235294117, | |
| "grad_norm": 0.515625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1717, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 0.16524064171122996, | |
| "grad_norm": 0.4296875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2635, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 0.1657754010695187, | |
| "grad_norm": 0.1318359375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0933, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.1663101604278075, | |
| "grad_norm": 0.08642578125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0547, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 0.16684491978609625, | |
| "grad_norm": 0.46484375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2123, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.16737967914438504, | |
| "grad_norm": 0.095703125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1559, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 0.1679144385026738, | |
| "grad_norm": 1.0703125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.243, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 0.16844919786096257, | |
| "grad_norm": 0.2578125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2638, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.16898395721925133, | |
| "grad_norm": 0.1435546875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1185, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 0.1695187165775401, | |
| "grad_norm": 0.5703125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2719, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 0.17005347593582887, | |
| "grad_norm": 1.375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4247, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 0.17058823529411765, | |
| "grad_norm": 0.34765625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3772, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 0.1711229946524064, | |
| "grad_norm": 0.10791015625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.212, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.1716577540106952, | |
| "grad_norm": 0.419921875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1685, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 0.17219251336898395, | |
| "grad_norm": 0.255859375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2688, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 0.17272727272727273, | |
| "grad_norm": 0.11376953125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2073, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 0.17326203208556148, | |
| "grad_norm": 0.51953125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2158, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 0.17379679144385027, | |
| "grad_norm": 1.2734375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3103, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.17433155080213902, | |
| "grad_norm": 0.2421875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2458, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 0.1748663101604278, | |
| "grad_norm": 0.193359375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2203, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 0.1754010695187166, | |
| "grad_norm": 0.392578125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2319, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 0.17593582887700535, | |
| "grad_norm": 0.625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2094, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 0.17647058823529413, | |
| "grad_norm": 0.396484375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2354, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.17700534759358288, | |
| "grad_norm": 0.73046875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3425, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 0.17754010695187167, | |
| "grad_norm": 0.37109375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1983, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 0.17807486631016042, | |
| "grad_norm": 0.08984375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1281, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 0.1786096256684492, | |
| "grad_norm": 0.103515625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1543, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 0.17914438502673796, | |
| "grad_norm": 0.62890625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2649, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.17967914438502675, | |
| "grad_norm": 0.11572265625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1132, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 0.1802139037433155, | |
| "grad_norm": 1.0, | |
| "learning_rate": 0.0002, | |
| "loss": 0.428, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 0.18074866310160428, | |
| "grad_norm": 0.283203125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2577, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.18128342245989304, | |
| "grad_norm": 0.1357421875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1888, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 0.18181818181818182, | |
| "grad_norm": 0.5859375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2504, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.18235294117647058, | |
| "grad_norm": 0.90625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1937, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 0.18288770053475936, | |
| "grad_norm": 0.3671875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1778, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 0.18342245989304812, | |
| "grad_norm": 0.06298828125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0472, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 0.1839572192513369, | |
| "grad_norm": 0.455078125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1782, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 0.18449197860962566, | |
| "grad_norm": 0.494140625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2222, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.18502673796791444, | |
| "grad_norm": 0.3359375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1891, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 0.1855614973262032, | |
| "grad_norm": 0.51953125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3089, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 0.18609625668449198, | |
| "grad_norm": 2.734375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4823, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 0.18663101604278076, | |
| "grad_norm": 1.2890625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1879, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 0.18716577540106952, | |
| "grad_norm": 0.10546875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1862, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.1877005347593583, | |
| "grad_norm": 1.2578125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2416, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 0.18823529411764706, | |
| "grad_norm": 0.16796875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2126, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 0.18877005347593584, | |
| "grad_norm": 1.1484375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1726, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 0.1893048128342246, | |
| "grad_norm": 0.58203125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2085, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 0.18983957219251338, | |
| "grad_norm": 0.263671875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1478, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.19037433155080213, | |
| "grad_norm": 0.328125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1796, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 0.19090909090909092, | |
| "grad_norm": 0.72265625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3885, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 0.19144385026737967, | |
| "grad_norm": 0.2275390625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1843, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 0.19197860962566846, | |
| "grad_norm": 0.333984375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1804, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 0.1925133689839572, | |
| "grad_norm": 0.55859375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.192, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.193048128342246, | |
| "grad_norm": 0.1220703125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0813, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 0.19358288770053475, | |
| "grad_norm": 0.1171875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0699, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 0.19411764705882353, | |
| "grad_norm": 0.453125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2945, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 0.1946524064171123, | |
| "grad_norm": 0.107421875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1997, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.19518716577540107, | |
| "grad_norm": 0.28125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2366, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.19572192513368983, | |
| "grad_norm": 0.58984375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2421, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 0.1962566844919786, | |
| "grad_norm": 0.2421875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1364, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 0.19679144385026737, | |
| "grad_norm": 0.45703125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1706, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 0.19732620320855615, | |
| "grad_norm": 0.1259765625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1628, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 0.19786096256684493, | |
| "grad_norm": 0.0927734375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1654, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.1983957219251337, | |
| "grad_norm": 0.29296875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0769, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 0.19893048128342247, | |
| "grad_norm": 0.255859375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2066, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 0.19946524064171123, | |
| "grad_norm": 0.376953125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.12, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 0.2080078125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.164, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 0.20053475935828877, | |
| "grad_norm": 0.375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.251, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.20106951871657755, | |
| "grad_norm": 0.6640625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1263, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 0.2016042780748663, | |
| "grad_norm": 0.431640625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2442, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 0.2021390374331551, | |
| "grad_norm": 0.470703125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0631, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 0.20267379679144384, | |
| "grad_norm": 0.326171875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2238, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 0.20320855614973263, | |
| "grad_norm": 0.2890625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.319, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.20374331550802138, | |
| "grad_norm": 0.298828125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1655, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 0.20427807486631017, | |
| "grad_norm": 0.205078125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1586, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 0.20481283422459892, | |
| "grad_norm": 0.318359375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2062, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 0.2053475935828877, | |
| "grad_norm": 0.365234375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2637, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 0.20588235294117646, | |
| "grad_norm": 0.271484375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.203, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.20641711229946524, | |
| "grad_norm": 0.4140625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3365, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 0.206951871657754, | |
| "grad_norm": 0.2578125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2558, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 0.20748663101604278, | |
| "grad_norm": 0.380859375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1074, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 0.20802139037433154, | |
| "grad_norm": 0.2470703125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1507, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 0.20855614973262032, | |
| "grad_norm": 0.1474609375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2006, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.20909090909090908, | |
| "grad_norm": 0.224609375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2195, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 0.20962566844919786, | |
| "grad_norm": 0.193359375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1366, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 0.21016042780748664, | |
| "grad_norm": 0.3359375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2126, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 0.2106951871657754, | |
| "grad_norm": 0.20703125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0699, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 0.21122994652406418, | |
| "grad_norm": 0.248046875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1594, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.21176470588235294, | |
| "grad_norm": 0.265625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1362, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 0.21229946524064172, | |
| "grad_norm": 0.1328125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1458, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 0.21283422459893048, | |
| "grad_norm": 0.67578125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2377, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 0.21336898395721926, | |
| "grad_norm": 1.1171875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.4138, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 0.21390374331550802, | |
| "grad_norm": 0.2197265625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1331, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.2144385026737968, | |
| "grad_norm": 0.703125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3063, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 0.21497326203208555, | |
| "grad_norm": 0.49609375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.15, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 0.21550802139037434, | |
| "grad_norm": 0.59765625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1727, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 0.2160427807486631, | |
| "grad_norm": 0.328125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1833, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 0.21657754010695188, | |
| "grad_norm": 0.146484375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0712, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.21711229946524063, | |
| "grad_norm": 0.134765625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1766, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 0.21764705882352942, | |
| "grad_norm": 0.3203125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2309, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 0.21818181818181817, | |
| "grad_norm": 0.1298828125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1819, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 0.21871657754010695, | |
| "grad_norm": 0.146484375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1536, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 0.2192513368983957, | |
| "grad_norm": 0.177734375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1115, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.2197860962566845, | |
| "grad_norm": 0.181640625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1675, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 0.22032085561497325, | |
| "grad_norm": 1.171875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1975, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 0.22085561497326203, | |
| "grad_norm": 0.609375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2091, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 0.22139037433155082, | |
| "grad_norm": 0.431640625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2046, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 0.22192513368983957, | |
| "grad_norm": 0.11474609375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1062, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.22245989304812835, | |
| "grad_norm": 0.328125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.252, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.2229946524064171, | |
| "grad_norm": 0.4296875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2299, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 0.2235294117647059, | |
| "grad_norm": 0.453125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1805, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 0.22406417112299465, | |
| "grad_norm": 0.2314453125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2925, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 0.22459893048128343, | |
| "grad_norm": 1.0546875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2782, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.2251336898395722, | |
| "grad_norm": 0.162109375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1329, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 0.22566844919786097, | |
| "grad_norm": 0.421875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2275, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 0.22620320855614973, | |
| "grad_norm": 0.2158203125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1434, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 0.2267379679144385, | |
| "grad_norm": 0.2119140625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1863, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 0.22727272727272727, | |
| "grad_norm": 0.69140625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2988, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.22780748663101605, | |
| "grad_norm": 0.5, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2001, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 0.2283422459893048, | |
| "grad_norm": 0.251953125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2114, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 0.2288770053475936, | |
| "grad_norm": 0.1796875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1191, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 0.22941176470588234, | |
| "grad_norm": 0.52734375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2393, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 0.22994652406417113, | |
| "grad_norm": 0.671875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1093, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.23048128342245988, | |
| "grad_norm": 0.51171875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2828, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 0.23101604278074866, | |
| "grad_norm": 0.26953125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2073, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 0.23155080213903742, | |
| "grad_norm": 0.578125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3156, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 0.2320855614973262, | |
| "grad_norm": 0.2294921875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1862, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 0.232620320855615, | |
| "grad_norm": 0.10791015625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1061, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.23315508021390374, | |
| "grad_norm": 0.466796875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2721, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 0.23368983957219253, | |
| "grad_norm": 0.12060546875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0834, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 0.23422459893048128, | |
| "grad_norm": 0.380859375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1377, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 0.23475935828877006, | |
| "grad_norm": 0.671875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.21, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 0.23529411764705882, | |
| "grad_norm": 0.2099609375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1004, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.2358288770053476, | |
| "grad_norm": 0.1611328125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1424, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 0.23636363636363636, | |
| "grad_norm": 0.1826171875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.12, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.23689839572192514, | |
| "grad_norm": 0.28515625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.155, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 0.2374331550802139, | |
| "grad_norm": 0.2255859375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1647, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 0.23796791443850268, | |
| "grad_norm": 0.2890625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1718, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.23850267379679144, | |
| "grad_norm": 0.103515625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0817, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 0.23903743315508022, | |
| "grad_norm": 0.25390625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1277, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 0.23957219251336898, | |
| "grad_norm": 0.353515625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1198, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 0.24010695187165776, | |
| "grad_norm": 0.06787109375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0802, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 0.24064171122994651, | |
| "grad_norm": 0.2490234375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0657, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.2411764705882353, | |
| "grad_norm": 0.1181640625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1076, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 0.24171122994652405, | |
| "grad_norm": 0.1181640625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0829, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 0.24224598930481284, | |
| "grad_norm": 0.1767578125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1498, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 0.2427807486631016, | |
| "grad_norm": 0.1025390625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0773, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 0.24331550802139038, | |
| "grad_norm": 0.1357421875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0283, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.24385026737967913, | |
| "grad_norm": 0.73046875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1534, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 0.24438502673796791, | |
| "grad_norm": 0.25390625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2006, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 0.2449197860962567, | |
| "grad_norm": 0.09814453125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0201, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 0.24545454545454545, | |
| "grad_norm": 0.458984375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3034, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 0.24598930481283424, | |
| "grad_norm": 0.63671875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3538, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.246524064171123, | |
| "grad_norm": 0.2001953125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.138, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 0.24705882352941178, | |
| "grad_norm": 1.3671875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2199, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 0.24759358288770053, | |
| "grad_norm": 0.58984375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3079, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 0.24812834224598931, | |
| "grad_norm": 0.51171875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.36, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 0.24866310160427807, | |
| "grad_norm": 0.91015625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3007, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.24919786096256685, | |
| "grad_norm": 1.0234375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2952, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 0.2497326203208556, | |
| "grad_norm": 0.2197265625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0584, | |
| "step": 467 | |
| }, | |
| { | |
| "epoch": 0.25026737967914436, | |
| "grad_norm": 0.1923828125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0912, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 0.25080213903743315, | |
| "grad_norm": 0.21875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1412, | |
| "step": 469 | |
| }, | |
| { | |
| "epoch": 0.25133689839572193, | |
| "grad_norm": 0.59765625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3287, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.2518716577540107, | |
| "grad_norm": 0.1708984375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2102, | |
| "step": 471 | |
| }, | |
| { | |
| "epoch": 0.25240641711229944, | |
| "grad_norm": 0.259765625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2511, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 0.2529411764705882, | |
| "grad_norm": 0.388671875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2573, | |
| "step": 473 | |
| }, | |
| { | |
| "epoch": 0.253475935828877, | |
| "grad_norm": 0.1025390625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1181, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 0.2540106951871658, | |
| "grad_norm": 0.291015625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1183, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.2545454545454545, | |
| "grad_norm": 0.330078125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1341, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 0.2550802139037433, | |
| "grad_norm": 0.31640625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.16, | |
| "step": 477 | |
| }, | |
| { | |
| "epoch": 0.2556149732620321, | |
| "grad_norm": 0.2392578125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2133, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 0.25614973262032087, | |
| "grad_norm": 0.373046875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2048, | |
| "step": 479 | |
| }, | |
| { | |
| "epoch": 0.25668449197860965, | |
| "grad_norm": 0.2353515625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2218, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.2572192513368984, | |
| "grad_norm": 0.36328125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1959, | |
| "step": 481 | |
| }, | |
| { | |
| "epoch": 0.25775401069518716, | |
| "grad_norm": 0.1318359375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1816, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 0.25828877005347595, | |
| "grad_norm": 0.458984375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1479, | |
| "step": 483 | |
| }, | |
| { | |
| "epoch": 0.25882352941176473, | |
| "grad_norm": 0.33984375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.189, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 0.25935828877005346, | |
| "grad_norm": 1.1953125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2001, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.25989304812834224, | |
| "grad_norm": 0.06591796875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0943, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 0.260427807486631, | |
| "grad_norm": 0.287109375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1934, | |
| "step": 487 | |
| }, | |
| { | |
| "epoch": 0.2609625668449198, | |
| "grad_norm": 0.51171875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2581, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 0.26149732620320854, | |
| "grad_norm": 0.15625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.0733, | |
| "step": 489 | |
| }, | |
| { | |
| "epoch": 0.2620320855614973, | |
| "grad_norm": 0.494140625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2736, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.2625668449197861, | |
| "grad_norm": 0.271484375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1819, | |
| "step": 491 | |
| }, | |
| { | |
| "epoch": 0.2631016042780749, | |
| "grad_norm": 0.302734375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.2011, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 0.2636363636363636, | |
| "grad_norm": 0.3984375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1678, | |
| "step": 493 | |
| }, | |
| { | |
| "epoch": 0.2641711229946524, | |
| "grad_norm": 0.94921875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1992, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 0.2647058823529412, | |
| "grad_norm": 1.0546875, | |
| "learning_rate": 0.0002, | |
| "loss": 0.3075, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.26524064171122996, | |
| "grad_norm": 0.21484375, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1961, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 0.2657754010695187, | |
| "grad_norm": 0.236328125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1905, | |
| "step": 497 | |
| }, | |
| { | |
| "epoch": 0.2663101604278075, | |
| "grad_norm": 0.1015625, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1238, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 0.26684491978609626, | |
| "grad_norm": 0.1767578125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.122, | |
| "step": 499 | |
| }, | |
| { | |
| "epoch": 0.26737967914438504, | |
| "grad_norm": 0.70703125, | |
| "learning_rate": 0.0002, | |
| "loss": 0.192, | |
| "step": 500 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 500, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 100, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.729911211451351e+17, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |