| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 75.0, | |
| "eval_steps": 1000, | |
| "global_step": 150, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 0.0, | |
| "learning_rate": 0, | |
| "loss": 1.3402, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 0.0, | |
| "learning_rate": 0, | |
| "loss": 1.286, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "grad_norm": 0.0, | |
| "learning_rate": 0, | |
| "loss": 1.3382, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "grad_norm": 6.078482627868652, | |
| "learning_rate": 0.0, | |
| "loss": 1.2555, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "grad_norm": 5.590641975402832, | |
| "learning_rate": 1.5051499783199055e-07, | |
| "loss": 1.2272, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "grad_norm": 5.590641975402832, | |
| "learning_rate": 1.5051499783199055e-07, | |
| "loss": 1.3339, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 3.5, | |
| "grad_norm": 5.590641975402832, | |
| "learning_rate": 1.5051499783199055e-07, | |
| "loss": 1.4161, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "grad_norm": 11.268411636352539, | |
| "learning_rate": 2.385606273598312e-07, | |
| "loss": 1.362, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 4.5, | |
| "grad_norm": 9.850939750671387, | |
| "learning_rate": 3.010299956639811e-07, | |
| "loss": 1.1579, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "grad_norm": 11.71530532836914, | |
| "learning_rate": 3.494850021680093e-07, | |
| "loss": 1.4335, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 5.5, | |
| "grad_norm": 5.690868377685547, | |
| "learning_rate": 3.8907562519182173e-07, | |
| "loss": 1.2658, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "grad_norm": 10.700072288513184, | |
| "learning_rate": 4.2254902000712834e-07, | |
| "loss": 1.3482, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 6.5, | |
| "grad_norm": 6.189509391784668, | |
| "learning_rate": 4.5154499349597166e-07, | |
| "loss": 1.3464, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 7.0, | |
| "grad_norm": 6.247044086456299, | |
| "learning_rate": 4.771212547196623e-07, | |
| "loss": 1.2738, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 7.5, | |
| "grad_norm": 6.324206829071045, | |
| "learning_rate": 4.999999999999999e-07, | |
| "loss": 1.2795, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "grad_norm": 5.119565486907959, | |
| "learning_rate": 5.206963425791124e-07, | |
| "loss": 1.281, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 8.5, | |
| "grad_norm": 5.067975044250488, | |
| "learning_rate": 5.395906230238123e-07, | |
| "loss": 1.4589, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 9.0, | |
| "grad_norm": 4.973406791687012, | |
| "learning_rate": 5.569716761534182e-07, | |
| "loss": 1.293, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 9.5, | |
| "grad_norm": 4.727417945861816, | |
| "learning_rate": 5.730640178391189e-07, | |
| "loss": 1.2502, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "grad_norm": 4.404863357543945, | |
| "learning_rate": 5.880456295278405e-07, | |
| "loss": 1.2186, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 10.5, | |
| "grad_norm": 6.084046840667725, | |
| "learning_rate": 6.020599913279622e-07, | |
| "loss": 1.4264, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 11.0, | |
| "grad_norm": 4.772565841674805, | |
| "learning_rate": 6.15224460689137e-07, | |
| "loss": 1.2514, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 11.5, | |
| "grad_norm": 4.340854167938232, | |
| "learning_rate": 6.276362525516529e-07, | |
| "loss": 1.2459, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 12.0, | |
| "grad_norm": 3.630235433578491, | |
| "learning_rate": 6.393768004764143e-07, | |
| "loss": 1.0926, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 12.5, | |
| "grad_norm": 7.564579963684082, | |
| "learning_rate": 6.505149978319905e-07, | |
| "loss": 1.3692, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 13.0, | |
| "grad_norm": 4.304481029510498, | |
| "learning_rate": 6.611096473669595e-07, | |
| "loss": 1.0156, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 13.5, | |
| "grad_norm": 5.493950366973877, | |
| "learning_rate": 6.712113404111031e-07, | |
| "loss": 1.2072, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 14.0, | |
| "grad_norm": 3.82694411277771, | |
| "learning_rate": 6.808639180087963e-07, | |
| "loss": 1.0592, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 14.5, | |
| "grad_norm": 3.7888131141662598, | |
| "learning_rate": 6.901056208558029e-07, | |
| "loss": 1.1043, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 15.0, | |
| "grad_norm": 3.5746841430664062, | |
| "learning_rate": 6.989700043360186e-07, | |
| "loss": 1.1532, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 15.5, | |
| "grad_norm": 4.591047286987305, | |
| "learning_rate": 7.074866739854088e-07, | |
| "loss": 1.2519, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 16.0, | |
| "grad_norm": 10.526527404785156, | |
| "learning_rate": 7.156818820794935e-07, | |
| "loss": 1.1504, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 16.5, | |
| "grad_norm": 4.354953289031982, | |
| "learning_rate": 7.235790156711094e-07, | |
| "loss": 1.098, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 17.0, | |
| "grad_norm": 5.967930316925049, | |
| "learning_rate": 7.311989989494779e-07, | |
| "loss": 1.1957, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 17.5, | |
| "grad_norm": 3.097174644470215, | |
| "learning_rate": 7.38560627359831e-07, | |
| "loss": 1.073, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 18.0, | |
| "grad_norm": 3.9502739906311035, | |
| "learning_rate": 7.456808469171361e-07, | |
| "loss": 1.0042, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 18.5, | |
| "grad_norm": 3.2742667198181152, | |
| "learning_rate": 7.525749891599529e-07, | |
| "loss": 1.004, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 19.0, | |
| "grad_norm": 4.2321953773498535, | |
| "learning_rate": 7.592569699389436e-07, | |
| "loss": 1.0532, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 19.5, | |
| "grad_norm": 4.322045803070068, | |
| "learning_rate": 7.657394585211274e-07, | |
| "loss": 1.0297, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "grad_norm": 3.5067355632781982, | |
| "learning_rate": 7.720340221751376e-07, | |
| "loss": 1.0541, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 20.5, | |
| "grad_norm": 3.1070497035980225, | |
| "learning_rate": 7.781512503836435e-07, | |
| "loss": 0.8917, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 21.0, | |
| "grad_norm": 3.413217067718506, | |
| "learning_rate": 7.841008620334974e-07, | |
| "loss": 1.0688, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 21.5, | |
| "grad_norm": 3.413217067718506, | |
| "learning_rate": 7.841008620334974e-07, | |
| "loss": 0.8956, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 22.0, | |
| "grad_norm": 31.582998275756836, | |
| "learning_rate": 7.89891798308405e-07, | |
| "loss": 1.1435, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 22.5, | |
| "grad_norm": 3.2715117931365967, | |
| "learning_rate": 7.955323035132494e-07, | |
| "loss": 1.023, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 23.0, | |
| "grad_norm": 3.081068754196167, | |
| "learning_rate": 8.01029995663981e-07, | |
| "loss": 0.9536, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 23.5, | |
| "grad_norm": 5.033097267150879, | |
| "learning_rate": 8.063919283598676e-07, | |
| "loss": 1.0542, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 24.0, | |
| "grad_norm": 3.2547528743743896, | |
| "learning_rate": 8.116246451989502e-07, | |
| "loss": 0.8736, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 24.5, | |
| "grad_norm": 3.2148780822753906, | |
| "learning_rate": 8.16734227789793e-07, | |
| "loss": 0.8366, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 25.0, | |
| "grad_norm": 3.271756887435913, | |
| "learning_rate": 8.217263382430935e-07, | |
| "loss": 0.9887, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 25.5, | |
| "grad_norm": 3.1897552013397217, | |
| "learning_rate": 8.266062568876716e-07, | |
| "loss": 0.9634, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 26.0, | |
| "grad_norm": 2.999152421951294, | |
| "learning_rate": 8.313789158407869e-07, | |
| "loss": 0.8479, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 26.5, | |
| "grad_norm": 3.126117467880249, | |
| "learning_rate": 8.360489289678585e-07, | |
| "loss": 0.9963, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 27.0, | |
| "grad_norm": 2.808807611465454, | |
| "learning_rate": 8.406206186877934e-07, | |
| "loss": 0.7693, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 27.5, | |
| "grad_norm": 3.8470795154571533, | |
| "learning_rate": 8.450980400142567e-07, | |
| "loss": 0.8929, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 28.0, | |
| "grad_norm": 4.059876918792725, | |
| "learning_rate": 8.494850021680092e-07, | |
| "loss": 0.9874, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 28.5, | |
| "grad_norm": 3.1297614574432373, | |
| "learning_rate": 8.53785088048968e-07, | |
| "loss": 0.8148, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 29.0, | |
| "grad_norm": 4.2221150398254395, | |
| "learning_rate": 8.580016718173995e-07, | |
| "loss": 0.9519, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 29.5, | |
| "grad_norm": 3.0946404933929443, | |
| "learning_rate": 8.621379348003944e-07, | |
| "loss": 0.8067, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 30.0, | |
| "grad_norm": 3.16498064994812, | |
| "learning_rate": 8.661968799114842e-07, | |
| "loss": 0.759, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 30.5, | |
| "grad_norm": 3.2981464862823486, | |
| "learning_rate": 8.701813447471218e-07, | |
| "loss": 0.8196, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 31.0, | |
| "grad_norm": 2.9492411613464355, | |
| "learning_rate": 8.740940135031001e-07, | |
| "loss": 0.8733, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 31.5, | |
| "grad_norm": 4.684609413146973, | |
| "learning_rate": 8.779374278362456e-07, | |
| "loss": 0.9314, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 32.0, | |
| "grad_norm": 2.6447761058807373, | |
| "learning_rate": 8.817139967814684e-07, | |
| "loss": 0.7206, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 32.5, | |
| "grad_norm": 3.544161319732666, | |
| "learning_rate": 8.854260058210719e-07, | |
| "loss": 0.7301, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 33.0, | |
| "grad_norm": 2.7136151790618896, | |
| "learning_rate": 8.890756251918216e-07, | |
| "loss": 0.7486, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 33.5, | |
| "grad_norm": 3.6095616817474365, | |
| "learning_rate": 8.926649175053833e-07, | |
| "loss": 0.6688, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 34.0, | |
| "grad_norm": 2.742135763168335, | |
| "learning_rate": 8.961958447491268e-07, | |
| "loss": 0.7577, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 34.5, | |
| "grad_norm": 3.166233777999878, | |
| "learning_rate": 8.996702747267907e-07, | |
| "loss": 0.6201, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 35.0, | |
| "grad_norm": 6.629209518432617, | |
| "learning_rate": 9.030899869919433e-07, | |
| "loss": 0.7546, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 35.5, | |
| "grad_norm": 3.2534568309783936, | |
| "learning_rate": 9.064566783214276e-07, | |
| "loss": 0.6966, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 36.0, | |
| "grad_norm": 2.9836554527282715, | |
| "learning_rate": 9.097719677709341e-07, | |
| "loss": 0.669, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 36.5, | |
| "grad_norm": 2.8371522426605225, | |
| "learning_rate": 9.13037401350413e-07, | |
| "loss": 0.602, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 37.0, | |
| "grad_norm": 3.1005606651306152, | |
| "learning_rate": 9.162544563531181e-07, | |
| "loss": 0.7351, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 37.5, | |
| "grad_norm": 2.771270751953125, | |
| "learning_rate": 9.194245453686276e-07, | |
| "loss": 0.5892, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 38.0, | |
| "grad_norm": 4.207065105438232, | |
| "learning_rate": 9.225490200071283e-07, | |
| "loss": 0.7159, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 38.5, | |
| "grad_norm": 4.0643439292907715, | |
| "learning_rate": 9.256291743595375e-07, | |
| "loss": 0.6424, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 39.0, | |
| "grad_norm": 3.463486433029175, | |
| "learning_rate": 9.28666248215634e-07, | |
| "loss": 0.6164, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 39.5, | |
| "grad_norm": 3.14841890335083, | |
| "learning_rate": 9.316614300602277e-07, | |
| "loss": 0.5615, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 40.0, | |
| "grad_norm": 2.958975076675415, | |
| "learning_rate": 9.346158598654879e-07, | |
| "loss": 0.5521, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 40.5, | |
| "grad_norm": 2.7486228942871094, | |
| "learning_rate": 9.375306316958498e-07, | |
| "loss": 0.5937, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 41.0, | |
| "grad_norm": 2.736597776412964, | |
| "learning_rate": 9.404067961403955e-07, | |
| "loss": 0.5238, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 41.5, | |
| "grad_norm": 2.9119350910186768, | |
| "learning_rate": 9.432453625862408e-07, | |
| "loss": 0.5296, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 42.0, | |
| "grad_norm": 3.870023727416992, | |
| "learning_rate": 9.4604730134524e-07, | |
| "loss": 0.5716, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 42.5, | |
| "grad_norm": 3.2765142917633057, | |
| "learning_rate": 9.488135456452205e-07, | |
| "loss": 0.597, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 43.0, | |
| "grad_norm": 3.575230598449707, | |
| "learning_rate": 9.515449934959715e-07, | |
| "loss": 0.5601, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 43.5, | |
| "grad_norm": 3.2691493034362793, | |
| "learning_rate": 9.542425094393247e-07, | |
| "loss": 0.5662, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 44.0, | |
| "grad_norm": 2.8934061527252197, | |
| "learning_rate": 9.569069261918583e-07, | |
| "loss": 0.5026, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 44.5, | |
| "grad_norm": 3.517810821533203, | |
| "learning_rate": 9.59539046188037e-07, | |
| "loss": 0.5107, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 45.0, | |
| "grad_norm": 3.1376760005950928, | |
| "learning_rate": 9.621396430309406e-07, | |
| "loss": 0.5166, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 45.5, | |
| "grad_norm": 3.4940109252929688, | |
| "learning_rate": 9.647094628571462e-07, | |
| "loss": 0.4549, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 46.0, | |
| "grad_norm": 6.771224498748779, | |
| "learning_rate": 9.672492256217836e-07, | |
| "loss": 0.4991, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 46.5, | |
| "grad_norm": 4.691510200500488, | |
| "learning_rate": 9.69759626309309e-07, | |
| "loss": 0.4521, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 47.0, | |
| "grad_norm": 3.1397156715393066, | |
| "learning_rate": 9.722413360750842e-07, | |
| "loss": 0.4287, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 47.5, | |
| "grad_norm": 3.4504847526550293, | |
| "learning_rate": 9.74695003322456e-07, | |
| "loss": 0.3875, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 48.0, | |
| "grad_norm": 2.7338361740112305, | |
| "learning_rate": 9.771212547196622e-07, | |
| "loss": 0.44, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 48.5, | |
| "grad_norm": 4.007640838623047, | |
| "learning_rate": 9.795206961605466e-07, | |
| "loss": 0.3561, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 49.0, | |
| "grad_norm": 3.637798309326172, | |
| "learning_rate": 9.818939136727774e-07, | |
| "loss": 0.4562, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 49.5, | |
| "grad_norm": 4.988349914550781, | |
| "learning_rate": 9.842414742769674e-07, | |
| "loss": 0.4038, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 50.0, | |
| "grad_norm": 3.512204885482788, | |
| "learning_rate": 9.865639267998492e-07, | |
| "loss": 0.3882, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 50.5, | |
| "grad_norm": 2.9533920288085938, | |
| "learning_rate": 9.888618026444236e-07, | |
| "loss": 0.3843, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 51.0, | |
| "grad_norm": 2.7523856163024902, | |
| "learning_rate": 9.91135616519784e-07, | |
| "loss": 0.3811, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 51.5, | |
| "grad_norm": 6.668997764587402, | |
| "learning_rate": 9.933858671331222e-07, | |
| "loss": 0.4223, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 52.0, | |
| "grad_norm": 3.053255319595337, | |
| "learning_rate": 9.956130378462473e-07, | |
| "loss": 0.3066, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 52.5, | |
| "grad_norm": 3.2109737396240234, | |
| "learning_rate": 9.978175972987748e-07, | |
| "loss": 0.3446, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 53.0, | |
| "grad_norm": 2.961881399154663, | |
| "learning_rate": 9.999999999999997e-07, | |
| "loss": 0.2935, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 53.5, | |
| "grad_norm": 3.6308419704437256, | |
| "learning_rate": 1e-06, | |
| "loss": 0.3619, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 54.0, | |
| "grad_norm": 3.809734582901001, | |
| "learning_rate": 1e-06, | |
| "loss": 0.3063, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 54.5, | |
| "grad_norm": 2.6650562286376953, | |
| "learning_rate": 1e-06, | |
| "loss": 0.3504, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 55.0, | |
| "grad_norm": 3.361117362976074, | |
| "learning_rate": 1e-06, | |
| "loss": 0.3084, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 55.5, | |
| "grad_norm": 2.6281044483184814, | |
| "learning_rate": 1e-06, | |
| "loss": 0.3191, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 56.0, | |
| "grad_norm": 3.0685670375823975, | |
| "learning_rate": 1e-06, | |
| "loss": 0.2472, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 56.5, | |
| "grad_norm": 3.149151563644409, | |
| "learning_rate": 1e-06, | |
| "loss": 0.255, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 57.0, | |
| "grad_norm": 3.0701425075531006, | |
| "learning_rate": 1e-06, | |
| "loss": 0.2392, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 57.5, | |
| "grad_norm": 2.449301242828369, | |
| "learning_rate": 1e-06, | |
| "loss": 0.2239, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 58.0, | |
| "grad_norm": 2.5579042434692383, | |
| "learning_rate": 1e-06, | |
| "loss": 0.265, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 58.5, | |
| "grad_norm": 3.4325478076934814, | |
| "learning_rate": 1e-06, | |
| "loss": 0.2522, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 59.0, | |
| "grad_norm": 3.4678075313568115, | |
| "learning_rate": 1e-06, | |
| "loss": 0.1957, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 59.5, | |
| "grad_norm": 3.3002102375030518, | |
| "learning_rate": 1e-06, | |
| "loss": 0.2152, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 60.0, | |
| "grad_norm": 2.939915180206299, | |
| "learning_rate": 1e-06, | |
| "loss": 0.228, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 60.5, | |
| "grad_norm": 2.763664484024048, | |
| "learning_rate": 1e-06, | |
| "loss": 0.2261, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 61.0, | |
| "grad_norm": 2.337585210800171, | |
| "learning_rate": 1e-06, | |
| "loss": 0.1851, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 61.5, | |
| "grad_norm": 2.545644760131836, | |
| "learning_rate": 1e-06, | |
| "loss": 0.2182, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 62.0, | |
| "grad_norm": 2.564753770828247, | |
| "learning_rate": 1e-06, | |
| "loss": 0.1962, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 62.5, | |
| "grad_norm": 2.338416814804077, | |
| "learning_rate": 1e-06, | |
| "loss": 0.2051, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 63.0, | |
| "grad_norm": 2.154878854751587, | |
| "learning_rate": 1e-06, | |
| "loss": 0.1879, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 63.5, | |
| "grad_norm": 3.405646800994873, | |
| "learning_rate": 1e-06, | |
| "loss": 0.1655, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 64.0, | |
| "grad_norm": 1.8845716714859009, | |
| "learning_rate": 1e-06, | |
| "loss": 0.161, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 64.5, | |
| "grad_norm": 1.707482099533081, | |
| "learning_rate": 1e-06, | |
| "loss": 0.1342, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 65.0, | |
| "grad_norm": 3.2093379497528076, | |
| "learning_rate": 1e-06, | |
| "loss": 0.1416, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 65.5, | |
| "grad_norm": 3.438514232635498, | |
| "learning_rate": 1e-06, | |
| "loss": 0.1486, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 66.0, | |
| "grad_norm": 1.9615154266357422, | |
| "learning_rate": 1e-06, | |
| "loss": 0.1461, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 66.5, | |
| "grad_norm": 2.3800694942474365, | |
| "learning_rate": 1e-06, | |
| "loss": 0.1133, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 67.0, | |
| "grad_norm": 2.241901159286499, | |
| "learning_rate": 1e-06, | |
| "loss": 0.1317, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 67.5, | |
| "grad_norm": 2.024240255355835, | |
| "learning_rate": 1e-06, | |
| "loss": 0.1195, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 68.0, | |
| "grad_norm": 2.2146880626678467, | |
| "learning_rate": 1e-06, | |
| "loss": 0.1413, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 68.5, | |
| "grad_norm": 2.4345452785491943, | |
| "learning_rate": 1e-06, | |
| "loss": 0.1313, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 69.0, | |
| "grad_norm": 3.0751936435699463, | |
| "learning_rate": 1e-06, | |
| "loss": 0.1231, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 69.5, | |
| "grad_norm": 2.2487804889678955, | |
| "learning_rate": 1e-06, | |
| "loss": 0.1131, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 70.0, | |
| "grad_norm": 3.5457677841186523, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0812, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 70.5, | |
| "grad_norm": 1.4189667701721191, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0574, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 71.0, | |
| "grad_norm": 2.4439244270324707, | |
| "learning_rate": 1e-06, | |
| "loss": 0.1146, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 71.5, | |
| "grad_norm": 1.7094790935516357, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0937, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 72.0, | |
| "grad_norm": 1.4799623489379883, | |
| "learning_rate": 1e-06, | |
| "loss": 0.085, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 72.5, | |
| "grad_norm": 1.4534505605697632, | |
| "learning_rate": 1e-06, | |
| "loss": 0.074, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 73.0, | |
| "grad_norm": 1.7459020614624023, | |
| "learning_rate": 1e-06, | |
| "loss": 0.1062, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 73.5, | |
| "grad_norm": 1.3827319145202637, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0597, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 74.0, | |
| "grad_norm": 1.4655884504318237, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0749, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 74.5, | |
| "grad_norm": 1.6126585006713867, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0812, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 75.0, | |
| "grad_norm": 1.4007526636123657, | |
| "learning_rate": 1e-06, | |
| "loss": 0.0724, | |
| "step": 150 | |
| } | |
| ], | |
| "logging_steps": 1.0, | |
| "max_steps": 10000, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 5000, | |
| "save_steps": 50, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 2.098224817063854e+17, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |