| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.9997915363769022, | |
| "eval_steps": 400, | |
| "global_step": 1199, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0008338544923910778, | |
| "grad_norm": 0.35252463817596436, | |
| "learning_rate": 2e-08, | |
| "loss": 0.8273, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.0008338544923910778, | |
| "eval_loss": 0.861491858959198, | |
| "eval_runtime": 319.044, | |
| "eval_samples_per_second": 16.872, | |
| "eval_steps_per_second": 2.815, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.0016677089847821555, | |
| "grad_norm": 0.3189675807952881, | |
| "learning_rate": 4e-08, | |
| "loss": 0.8572, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.0025015634771732333, | |
| "grad_norm": 0.3477325141429901, | |
| "learning_rate": 6e-08, | |
| "loss": 0.9059, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.003335417969564311, | |
| "grad_norm": 0.32423585653305054, | |
| "learning_rate": 8e-08, | |
| "loss": 0.8386, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.004169272461955389, | |
| "grad_norm": 0.32164493203163147, | |
| "learning_rate": 1e-07, | |
| "loss": 0.8753, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.0050031269543464665, | |
| "grad_norm": 0.33231157064437866, | |
| "learning_rate": 1.2e-07, | |
| "loss": 0.8665, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.005836981446737544, | |
| "grad_norm": 0.31154102087020874, | |
| "learning_rate": 1.4e-07, | |
| "loss": 0.8813, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.006670835939128622, | |
| "grad_norm": 0.3176712393760681, | |
| "learning_rate": 1.6e-07, | |
| "loss": 0.8683, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.0075046904315197, | |
| "grad_norm": 0.3419969975948334, | |
| "learning_rate": 1.8e-07, | |
| "loss": 0.849, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.008338544923910778, | |
| "grad_norm": 0.33408281207084656, | |
| "learning_rate": 2e-07, | |
| "loss": 0.8778, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.009172399416301855, | |
| "grad_norm": 0.3125899136066437, | |
| "learning_rate": 2.1999999999999998e-07, | |
| "loss": 0.8051, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.010006253908692933, | |
| "grad_norm": 0.32648083567619324, | |
| "learning_rate": 2.4e-07, | |
| "loss": 0.8297, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.01084010840108401, | |
| "grad_norm": 0.3143678605556488, | |
| "learning_rate": 2.6e-07, | |
| "loss": 0.8401, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.011673962893475089, | |
| "grad_norm": 0.3291407525539398, | |
| "learning_rate": 2.8e-07, | |
| "loss": 0.8668, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.012507817385866166, | |
| "grad_norm": 0.3211856484413147, | |
| "learning_rate": 3e-07, | |
| "loss": 0.84, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.013341671878257244, | |
| "grad_norm": 0.31169289350509644, | |
| "learning_rate": 3.2e-07, | |
| "loss": 0.8659, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.014175526370648322, | |
| "grad_norm": 0.3279392719268799, | |
| "learning_rate": 3.4000000000000003e-07, | |
| "loss": 0.8317, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.0150093808630394, | |
| "grad_norm": 0.3028814494609833, | |
| "learning_rate": 3.6e-07, | |
| "loss": 0.8246, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.015843235355430477, | |
| "grad_norm": 0.307066410779953, | |
| "learning_rate": 3.7999999999999996e-07, | |
| "loss": 0.8025, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.016677089847821555, | |
| "grad_norm": 0.3293515741825104, | |
| "learning_rate": 4e-07, | |
| "loss": 0.8674, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.017510944340212633, | |
| "grad_norm": 0.3215535581111908, | |
| "learning_rate": 4.1999999999999995e-07, | |
| "loss": 0.8533, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.01834479883260371, | |
| "grad_norm": 0.3236245810985565, | |
| "learning_rate": 4.3999999999999997e-07, | |
| "loss": 0.8957, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.01917865332499479, | |
| "grad_norm": 0.33956846594810486, | |
| "learning_rate": 4.6e-07, | |
| "loss": 0.8293, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.020012507817385866, | |
| "grad_norm": 0.33766457438468933, | |
| "learning_rate": 4.8e-07, | |
| "loss": 0.8381, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.020846362309776944, | |
| "grad_norm": 0.3336811065673828, | |
| "learning_rate": 5e-07, | |
| "loss": 0.8475, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.02168021680216802, | |
| "grad_norm": 0.307309627532959, | |
| "learning_rate": 5.2e-07, | |
| "loss": 0.852, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.0225140712945591, | |
| "grad_norm": 0.29773110151290894, | |
| "learning_rate": 5.4e-07, | |
| "loss": 0.8042, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.023347925786950177, | |
| "grad_norm": 0.3210948407649994, | |
| "learning_rate": 5.6e-07, | |
| "loss": 0.9099, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.024181780279341255, | |
| "grad_norm": 0.3149116337299347, | |
| "learning_rate": 5.8e-07, | |
| "loss": 0.8, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.025015634771732333, | |
| "grad_norm": 0.29307302832603455, | |
| "learning_rate": 6e-07, | |
| "loss": 0.8583, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.02584948926412341, | |
| "grad_norm": 0.3131772577762604, | |
| "learning_rate": 6.2e-07, | |
| "loss": 0.8432, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.026683343756514488, | |
| "grad_norm": 0.28058937191963196, | |
| "learning_rate": 6.4e-07, | |
| "loss": 0.8099, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.027517198248905566, | |
| "grad_norm": 0.2649301290512085, | |
| "learning_rate": 6.6e-07, | |
| "loss": 0.8715, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.028351052741296644, | |
| "grad_norm": 0.2659442126750946, | |
| "learning_rate": 6.800000000000001e-07, | |
| "loss": 0.7432, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.02918490723368772, | |
| "grad_norm": 0.24875527620315552, | |
| "learning_rate": 7e-07, | |
| "loss": 0.8502, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.0300187617260788, | |
| "grad_norm": 0.21730971336364746, | |
| "learning_rate": 7.2e-07, | |
| "loss": 0.813, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.030852616218469877, | |
| "grad_norm": 0.19625824689865112, | |
| "learning_rate": 7.4e-07, | |
| "loss": 0.7839, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.031686470710860955, | |
| "grad_norm": 0.19287347793579102, | |
| "learning_rate": 7.599999999999999e-07, | |
| "loss": 0.7949, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.032520325203252036, | |
| "grad_norm": 0.1588095873594284, | |
| "learning_rate": 7.799999999999999e-07, | |
| "loss": 0.7137, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.03335417969564311, | |
| "grad_norm": 0.17520244419574738, | |
| "learning_rate": 8e-07, | |
| "loss": 0.8112, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.03418803418803419, | |
| "grad_norm": 0.16394659876823425, | |
| "learning_rate": 8.199999999999999e-07, | |
| "loss": 0.8096, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.035021888680425266, | |
| "grad_norm": 0.1522587239742279, | |
| "learning_rate": 8.399999999999999e-07, | |
| "loss": 0.708, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.03585574317281635, | |
| "grad_norm": 0.16340652108192444, | |
| "learning_rate": 8.599999999999999e-07, | |
| "loss": 0.7816, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.03668959766520742, | |
| "grad_norm": 0.15452717244625092, | |
| "learning_rate": 8.799999999999999e-07, | |
| "loss": 0.7946, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.0375234521575985, | |
| "grad_norm": 0.1545872688293457, | |
| "learning_rate": 9e-07, | |
| "loss": 0.7908, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.03835730664998958, | |
| "grad_norm": 0.13799017667770386, | |
| "learning_rate": 9.2e-07, | |
| "loss": 0.7056, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.03919116114238066, | |
| "grad_norm": 0.14599646627902985, | |
| "learning_rate": 9.399999999999999e-07, | |
| "loss": 0.8098, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.04002501563477173, | |
| "grad_norm": 0.13951805233955383, | |
| "learning_rate": 9.6e-07, | |
| "loss": 0.7938, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.040858870127162814, | |
| "grad_norm": 0.14058919250965118, | |
| "learning_rate": 9.8e-07, | |
| "loss": 0.7806, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.04169272461955389, | |
| "grad_norm": 0.14300380647182465, | |
| "learning_rate": 1e-06, | |
| "loss": 0.8192, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.04252657911194497, | |
| "grad_norm": 0.13719584047794342, | |
| "learning_rate": 1.02e-06, | |
| "loss": 0.7939, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.04336043360433604, | |
| "grad_norm": 0.14002980291843414, | |
| "learning_rate": 1.04e-06, | |
| "loss": 0.8163, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.044194288096727125, | |
| "grad_norm": 0.13815677165985107, | |
| "learning_rate": 1.06e-06, | |
| "loss": 0.8137, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.0450281425891182, | |
| "grad_norm": 0.13772878050804138, | |
| "learning_rate": 1.08e-06, | |
| "loss": 0.771, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.04586199708150928, | |
| "grad_norm": 0.1414576917886734, | |
| "learning_rate": 1.1e-06, | |
| "loss": 0.7989, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.046695851573900354, | |
| "grad_norm": 0.11953306943178177, | |
| "learning_rate": 1.12e-06, | |
| "loss": 0.7074, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.047529706066291436, | |
| "grad_norm": 0.13051201403141022, | |
| "learning_rate": 1.1399999999999999e-06, | |
| "loss": 0.8034, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.04836356055868251, | |
| "grad_norm": 0.12802985310554504, | |
| "learning_rate": 1.16e-06, | |
| "loss": 0.7581, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.04919741505107359, | |
| "grad_norm": 0.13593435287475586, | |
| "learning_rate": 1.18e-06, | |
| "loss": 0.7978, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.050031269543464665, | |
| "grad_norm": 0.12595658004283905, | |
| "learning_rate": 1.2e-06, | |
| "loss": 0.7997, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.05086512403585575, | |
| "grad_norm": 0.1257220059633255, | |
| "learning_rate": 1.22e-06, | |
| "loss": 0.7194, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.05169897852824682, | |
| "grad_norm": 0.12169249355792999, | |
| "learning_rate": 1.24e-06, | |
| "loss": 0.7412, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.0525328330206379, | |
| "grad_norm": 0.12391054630279541, | |
| "learning_rate": 1.26e-06, | |
| "loss": 0.7849, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.053366687513028976, | |
| "grad_norm": 0.12134575098752975, | |
| "learning_rate": 1.28e-06, | |
| "loss": 0.6859, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.05420054200542006, | |
| "grad_norm": 0.11605742573738098, | |
| "learning_rate": 1.3e-06, | |
| "loss": 0.8112, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.05503439649781113, | |
| "grad_norm": 0.11593819409608841, | |
| "learning_rate": 1.32e-06, | |
| "loss": 0.6993, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.05586825099020221, | |
| "grad_norm": 0.12103109061717987, | |
| "learning_rate": 1.34e-06, | |
| "loss": 0.7595, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.05670210548259329, | |
| "grad_norm": 0.12391602247953415, | |
| "learning_rate": 1.3600000000000001e-06, | |
| "loss": 0.7676, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.05753595997498437, | |
| "grad_norm": 0.10895267128944397, | |
| "learning_rate": 1.38e-06, | |
| "loss": 0.7256, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.05836981446737544, | |
| "grad_norm": 0.1129627376794815, | |
| "learning_rate": 1.4e-06, | |
| "loss": 0.7097, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.059203668959766524, | |
| "grad_norm": 0.1078682616353035, | |
| "learning_rate": 1.42e-06, | |
| "loss": 0.7052, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.0600375234521576, | |
| "grad_norm": 0.1140650063753128, | |
| "learning_rate": 1.44e-06, | |
| "loss": 0.8067, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.06087137794454868, | |
| "grad_norm": 0.10528790205717087, | |
| "learning_rate": 1.46e-06, | |
| "loss": 0.7241, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.061705232436939754, | |
| "grad_norm": 0.11815836280584335, | |
| "learning_rate": 1.48e-06, | |
| "loss": 0.7344, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.06253908692933083, | |
| "grad_norm": 0.11140462756156921, | |
| "learning_rate": 1.5e-06, | |
| "loss": 0.7664, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.06337294142172191, | |
| "grad_norm": 0.11665944010019302, | |
| "learning_rate": 1.5199999999999998e-06, | |
| "loss": 0.7178, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.06420679591411299, | |
| "grad_norm": 0.10751160234212875, | |
| "learning_rate": 1.5399999999999999e-06, | |
| "loss": 0.7269, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.06504065040650407, | |
| "grad_norm": 0.09768356382846832, | |
| "learning_rate": 1.5599999999999999e-06, | |
| "loss": 0.7449, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.06587450489889514, | |
| "grad_norm": 0.09686972200870514, | |
| "learning_rate": 1.58e-06, | |
| "loss": 0.7097, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.06670835939128622, | |
| "grad_norm": 0.09527275711297989, | |
| "learning_rate": 1.6e-06, | |
| "loss": 0.7301, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.0675422138836773, | |
| "grad_norm": 0.0931195393204689, | |
| "learning_rate": 1.62e-06, | |
| "loss": 0.7524, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.06837606837606838, | |
| "grad_norm": 0.0892946794629097, | |
| "learning_rate": 1.6399999999999998e-06, | |
| "loss": 0.7156, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.06920992286845945, | |
| "grad_norm": 0.08845791220664978, | |
| "learning_rate": 1.6599999999999998e-06, | |
| "loss": 0.7349, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.07004377736085053, | |
| "grad_norm": 0.0896732360124588, | |
| "learning_rate": 1.6799999999999998e-06, | |
| "loss": 0.7145, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.07087763185324161, | |
| "grad_norm": 0.09997335076332092, | |
| "learning_rate": 1.6999999999999998e-06, | |
| "loss": 0.7103, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.0717114863456327, | |
| "grad_norm": 0.08821084350347519, | |
| "learning_rate": 1.7199999999999998e-06, | |
| "loss": 0.725, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.07254534083802376, | |
| "grad_norm": 0.08937980234622955, | |
| "learning_rate": 1.7399999999999999e-06, | |
| "loss": 0.7194, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.07337919533041484, | |
| "grad_norm": 0.09307857602834702, | |
| "learning_rate": 1.7599999999999999e-06, | |
| "loss": 0.7595, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.07421304982280592, | |
| "grad_norm": 0.08865829557180405, | |
| "learning_rate": 1.78e-06, | |
| "loss": 0.7464, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.075046904315197, | |
| "grad_norm": 0.09038495272397995, | |
| "learning_rate": 1.8e-06, | |
| "loss": 0.6698, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.07588075880758807, | |
| "grad_norm": 0.08254078030586243, | |
| "learning_rate": 1.82e-06, | |
| "loss": 0.7426, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.07671461329997915, | |
| "grad_norm": 0.08807505667209625, | |
| "learning_rate": 1.84e-06, | |
| "loss": 0.6789, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.07754846779237023, | |
| "grad_norm": 0.09523889422416687, | |
| "learning_rate": 1.86e-06, | |
| "loss": 0.7218, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.07838232228476132, | |
| "grad_norm": 0.08782277256250381, | |
| "learning_rate": 1.8799999999999998e-06, | |
| "loss": 0.7599, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.07921617677715238, | |
| "grad_norm": 0.08968065679073334, | |
| "learning_rate": 1.8999999999999998e-06, | |
| "loss": 0.7971, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.08005003126954346, | |
| "grad_norm": 0.08655782043933868, | |
| "learning_rate": 1.92e-06, | |
| "loss": 0.7479, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.08088388576193455, | |
| "grad_norm": 0.0854155421257019, | |
| "learning_rate": 1.94e-06, | |
| "loss": 0.7327, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.08171774025432563, | |
| "grad_norm": 0.08121508359909058, | |
| "learning_rate": 1.96e-06, | |
| "loss": 0.6749, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.0825515947467167, | |
| "grad_norm": 0.08429264277219772, | |
| "learning_rate": 1.98e-06, | |
| "loss": 0.6402, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.08338544923910778, | |
| "grad_norm": 0.08390086144208908, | |
| "learning_rate": 2e-06, | |
| "loss": 0.7434, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.08421930373149886, | |
| "grad_norm": 0.08662284910678864, | |
| "learning_rate": 1.9999995964675577e-06, | |
| "loss": 0.6681, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.08505315822388994, | |
| "grad_norm": 0.08700251579284668, | |
| "learning_rate": 1.9999983858705566e-06, | |
| "loss": 0.763, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.085887012716281, | |
| "grad_norm": 0.08504343777894974, | |
| "learning_rate": 1.9999963682099734e-06, | |
| "loss": 0.77, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.08672086720867209, | |
| "grad_norm": 0.08222745358943939, | |
| "learning_rate": 1.999993543487437e-06, | |
| "loss": 0.7103, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.08755472170106317, | |
| "grad_norm": 0.08124719560146332, | |
| "learning_rate": 1.9999899117052263e-06, | |
| "loss": 0.6913, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.08838857619345425, | |
| "grad_norm": 0.09033836424350739, | |
| "learning_rate": 1.9999854728662734e-06, | |
| "loss": 0.7065, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.08922243068584532, | |
| "grad_norm": 0.09045730531215668, | |
| "learning_rate": 1.99998022697416e-06, | |
| "loss": 0.7624, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.0900562851782364, | |
| "grad_norm": 0.0832991823554039, | |
| "learning_rate": 1.9999741740331203e-06, | |
| "loss": 0.7399, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.09089013967062748, | |
| "grad_norm": 0.07666011154651642, | |
| "learning_rate": 1.9999673140480388e-06, | |
| "loss": 0.6477, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.09172399416301856, | |
| "grad_norm": 0.08195007592439651, | |
| "learning_rate": 1.9999596470244527e-06, | |
| "loss": 0.7068, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.09255784865540963, | |
| "grad_norm": 0.07755164802074432, | |
| "learning_rate": 1.99995117296855e-06, | |
| "loss": 0.7021, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.09339170314780071, | |
| "grad_norm": 0.0782230794429779, | |
| "learning_rate": 1.9999418918871685e-06, | |
| "loss": 0.7039, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.09422555764019179, | |
| "grad_norm": 0.08311248570680618, | |
| "learning_rate": 1.9999318037877995e-06, | |
| "loss": 0.7238, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.09505941213258287, | |
| "grad_norm": 0.08455490320920944, | |
| "learning_rate": 1.999920908678585e-06, | |
| "loss": 0.7184, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.09589326662497394, | |
| "grad_norm": 0.08463476598262787, | |
| "learning_rate": 1.9999092065683174e-06, | |
| "loss": 0.6162, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.09672712111736502, | |
| "grad_norm": 0.07902689278125763, | |
| "learning_rate": 1.999896697466442e-06, | |
| "loss": 0.6649, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.0975609756097561, | |
| "grad_norm": 0.08383440971374512, | |
| "learning_rate": 1.999883381383053e-06, | |
| "loss": 0.703, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.09839483010214718, | |
| "grad_norm": 0.07288578897714615, | |
| "learning_rate": 1.999869258328899e-06, | |
| "loss": 0.7286, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.09922868459453825, | |
| "grad_norm": 0.08177065849304199, | |
| "learning_rate": 1.999854328315377e-06, | |
| "loss": 0.6899, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.10006253908692933, | |
| "grad_norm": 0.07529062032699585, | |
| "learning_rate": 1.999838591354537e-06, | |
| "loss": 0.6781, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.10089639357932041, | |
| "grad_norm": 0.08230678737163544, | |
| "learning_rate": 1.9998220474590795e-06, | |
| "loss": 0.7059, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.1017302480717115, | |
| "grad_norm": 0.08000528067350388, | |
| "learning_rate": 1.9998046966423567e-06, | |
| "loss": 0.6561, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.10256410256410256, | |
| "grad_norm": 0.07669328153133392, | |
| "learning_rate": 1.9997865389183717e-06, | |
| "loss": 0.6452, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.10339795705649364, | |
| "grad_norm": 0.07872146368026733, | |
| "learning_rate": 1.9997675743017794e-06, | |
| "loss": 0.7337, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.10423181154888472, | |
| "grad_norm": 0.07842138409614563, | |
| "learning_rate": 1.9997478028078853e-06, | |
| "loss": 0.7126, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.1050656660412758, | |
| "grad_norm": 0.07890679687261581, | |
| "learning_rate": 1.9997272244526453e-06, | |
| "loss": 0.7049, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.10589952053366687, | |
| "grad_norm": 0.07983887195587158, | |
| "learning_rate": 1.999705839252669e-06, | |
| "loss": 0.7578, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.10673337502605795, | |
| "grad_norm": 0.07721222192049026, | |
| "learning_rate": 1.9996836472252144e-06, | |
| "loss": 0.678, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.10756722951844903, | |
| "grad_norm": 0.07825150340795517, | |
| "learning_rate": 1.9996606483881927e-06, | |
| "loss": 0.6763, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.10840108401084012, | |
| "grad_norm": 0.08063995838165283, | |
| "learning_rate": 1.999636842760165e-06, | |
| "loss": 0.758, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.10923493850323118, | |
| "grad_norm": 0.0724928081035614, | |
| "learning_rate": 1.9996122303603443e-06, | |
| "loss": 0.7312, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.11006879299562226, | |
| "grad_norm": 0.08180397748947144, | |
| "learning_rate": 1.999586811208594e-06, | |
| "loss": 0.6794, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.11090264748801335, | |
| "grad_norm": 0.0856177881360054, | |
| "learning_rate": 1.99956058532543e-06, | |
| "loss": 0.6885, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.11173650198040443, | |
| "grad_norm": 0.0763324424624443, | |
| "learning_rate": 1.999533552732017e-06, | |
| "loss": 0.6821, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.1125703564727955, | |
| "grad_norm": 0.07964632660150528, | |
| "learning_rate": 1.9995057134501725e-06, | |
| "loss": 0.7362, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.11340421096518657, | |
| "grad_norm": 0.07633720338344574, | |
| "learning_rate": 1.999477067502365e-06, | |
| "loss": 0.7396, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.11423806545757766, | |
| "grad_norm": 0.07655656337738037, | |
| "learning_rate": 1.9994476149117132e-06, | |
| "loss": 0.728, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.11507191994996874, | |
| "grad_norm": 0.07770884782075882, | |
| "learning_rate": 1.9994173557019877e-06, | |
| "loss": 0.6878, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.1159057744423598, | |
| "grad_norm": 0.07332731038331985, | |
| "learning_rate": 1.999386289897609e-06, | |
| "loss": 0.7115, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.11673962893475089, | |
| "grad_norm": 0.0784095972776413, | |
| "learning_rate": 1.9993544175236497e-06, | |
| "loss": 0.6653, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.11757348342714197, | |
| "grad_norm": 0.08311771601438522, | |
| "learning_rate": 1.9993217386058326e-06, | |
| "loss": 0.6504, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.11840733791953305, | |
| "grad_norm": 0.07529395818710327, | |
| "learning_rate": 1.999288253170532e-06, | |
| "loss": 0.7083, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.11924119241192412, | |
| "grad_norm": 0.07608146965503693, | |
| "learning_rate": 1.999253961244773e-06, | |
| "loss": 0.6856, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.1200750469043152, | |
| "grad_norm": 0.07125360518693924, | |
| "learning_rate": 1.9992188628562303e-06, | |
| "loss": 0.6095, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.12090890139670628, | |
| "grad_norm": 0.07715660333633423, | |
| "learning_rate": 1.999182958033232e-06, | |
| "loss": 0.7337, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.12174275588909736, | |
| "grad_norm": 0.07756403833627701, | |
| "learning_rate": 1.999146246804755e-06, | |
| "loss": 0.6929, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.12257661038148843, | |
| "grad_norm": 0.07667449861764908, | |
| "learning_rate": 1.9991087292004273e-06, | |
| "loss": 0.67, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.12341046487387951, | |
| "grad_norm": 0.08285968005657196, | |
| "learning_rate": 1.9990704052505284e-06, | |
| "loss": 0.7483, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.12424431936627059, | |
| "grad_norm": 0.07476554065942764, | |
| "learning_rate": 1.9990312749859887e-06, | |
| "loss": 0.6846, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.12507817385866166, | |
| "grad_norm": 0.0818118005990982, | |
| "learning_rate": 1.998991338438388e-06, | |
| "loss": 0.6709, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.12591202835105275, | |
| "grad_norm": 0.08350253850221634, | |
| "learning_rate": 1.998950595639958e-06, | |
| "loss": 0.7379, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.12674588284344382, | |
| "grad_norm": 0.07916589826345444, | |
| "learning_rate": 1.9989090466235806e-06, | |
| "loss": 0.7206, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.1275797373358349, | |
| "grad_norm": 0.07860454171895981, | |
| "learning_rate": 1.998866691422789e-06, | |
| "loss": 0.6777, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.12841359182822598, | |
| "grad_norm": 0.07861131429672241, | |
| "learning_rate": 1.998823530071766e-06, | |
| "loss": 0.6938, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.12924744632061705, | |
| "grad_norm": 0.07819496840238571, | |
| "learning_rate": 1.9987795626053465e-06, | |
| "loss": 0.7193, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.13008130081300814, | |
| "grad_norm": 0.07828415185213089, | |
| "learning_rate": 1.9987347890590144e-06, | |
| "loss": 0.6809, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.1309151553053992, | |
| "grad_norm": 0.07755222916603088, | |
| "learning_rate": 1.998689209468905e-06, | |
| "loss": 0.7213, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.13174900979779028, | |
| "grad_norm": 0.07672612369060516, | |
| "learning_rate": 1.998642823871804e-06, | |
| "loss": 0.7196, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.13258286429018137, | |
| "grad_norm": 0.07963719964027405, | |
| "learning_rate": 1.9985956323051475e-06, | |
| "loss": 0.7465, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.13341671878257244, | |
| "grad_norm": 0.0795946940779686, | |
| "learning_rate": 1.9985476348070223e-06, | |
| "loss": 0.717, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.1342505732749635, | |
| "grad_norm": 0.07344063371419907, | |
| "learning_rate": 1.9984988314161656e-06, | |
| "loss": 0.6861, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.1350844277673546, | |
| "grad_norm": 0.07722171396017075, | |
| "learning_rate": 1.9984492221719645e-06, | |
| "loss": 0.6825, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.13591828225974567, | |
| "grad_norm": 0.0812341496348381, | |
| "learning_rate": 1.9983988071144572e-06, | |
| "loss": 0.7245, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.13675213675213677, | |
| "grad_norm": 0.08472032845020294, | |
| "learning_rate": 1.998347586284332e-06, | |
| "loss": 0.7315, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.13758599124452783, | |
| "grad_norm": 0.07785464823246002, | |
| "learning_rate": 1.9982955597229275e-06, | |
| "loss": 0.6906, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.1384198457369189, | |
| "grad_norm": 0.07382892072200775, | |
| "learning_rate": 1.998242727472232e-06, | |
| "loss": 0.6488, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.13925370022931, | |
| "grad_norm": 0.07445723563432693, | |
| "learning_rate": 1.9981890895748845e-06, | |
| "loss": 0.692, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.14008755472170106, | |
| "grad_norm": 0.07889281213283539, | |
| "learning_rate": 1.998134646074175e-06, | |
| "loss": 0.6438, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.14092140921409213, | |
| "grad_norm": 0.07664698362350464, | |
| "learning_rate": 1.9980793970140426e-06, | |
| "loss": 0.6993, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.14175526370648323, | |
| "grad_norm": 0.07807335257530212, | |
| "learning_rate": 1.9980233424390773e-06, | |
| "loss": 0.7334, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.1425891181988743, | |
| "grad_norm": 0.0683249682188034, | |
| "learning_rate": 1.9979664823945174e-06, | |
| "loss": 0.6325, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.1434229726912654, | |
| "grad_norm": 0.08339341729879379, | |
| "learning_rate": 1.997908816926254e-06, | |
| "loss": 0.697, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.14425682718365646, | |
| "grad_norm": 0.07781082391738892, | |
| "learning_rate": 1.997850346080827e-06, | |
| "loss": 0.6748, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.14509068167604752, | |
| "grad_norm": 0.07718750089406967, | |
| "learning_rate": 1.9977910699054247e-06, | |
| "loss": 0.7002, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.14592453616843862, | |
| "grad_norm": 0.07779830694198608, | |
| "learning_rate": 1.997730988447888e-06, | |
| "loss": 0.723, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.14675839066082969, | |
| "grad_norm": 0.07958701252937317, | |
| "learning_rate": 1.997670101756706e-06, | |
| "loss": 0.6651, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.14759224515322075, | |
| "grad_norm": 0.08142640441656113, | |
| "learning_rate": 1.997608409881019e-06, | |
| "loss": 0.6652, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.14842609964561185, | |
| "grad_norm": 0.07898704707622528, | |
| "learning_rate": 1.9975459128706152e-06, | |
| "loss": 0.6842, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.14925995413800291, | |
| "grad_norm": 0.07870691269636154, | |
| "learning_rate": 1.997482610775935e-06, | |
| "loss": 0.7167, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.150093808630394, | |
| "grad_norm": 0.07766727358102798, | |
| "learning_rate": 1.9974185036480658e-06, | |
| "loss": 0.6145, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.15092766312278508, | |
| "grad_norm": 0.07913653552532196, | |
| "learning_rate": 1.997353591538748e-06, | |
| "loss": 0.7598, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.15176151761517614, | |
| "grad_norm": 0.07850881665945053, | |
| "learning_rate": 1.9972878745003684e-06, | |
| "loss": 0.7492, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.15259537210756724, | |
| "grad_norm": 0.07599500566720963, | |
| "learning_rate": 1.9972213525859656e-06, | |
| "loss": 0.6802, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.1534292265999583, | |
| "grad_norm": 0.08165914565324783, | |
| "learning_rate": 1.997154025849227e-06, | |
| "loss": 0.6343, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.15426308109234937, | |
| "grad_norm": 0.0763566642999649, | |
| "learning_rate": 1.9970858943444893e-06, | |
| "loss": 0.7214, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.15509693558474047, | |
| "grad_norm": 0.07993835210800171, | |
| "learning_rate": 1.99701695812674e-06, | |
| "loss": 0.6975, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.15593079007713154, | |
| "grad_norm": 0.07139826565980911, | |
| "learning_rate": 1.996947217251614e-06, | |
| "loss": 0.6714, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.15676464456952263, | |
| "grad_norm": 0.0777943804860115, | |
| "learning_rate": 1.996876671775397e-06, | |
| "loss": 0.6684, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.1575984990619137, | |
| "grad_norm": 0.07822203636169434, | |
| "learning_rate": 1.9968053217550242e-06, | |
| "loss": 0.7161, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.15843235355430477, | |
| "grad_norm": 0.07645302265882492, | |
| "learning_rate": 1.9967331672480794e-06, | |
| "loss": 0.647, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.15926620804669586, | |
| "grad_norm": 0.07744666188955307, | |
| "learning_rate": 1.996660208312796e-06, | |
| "loss": 0.6887, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.16010006253908693, | |
| "grad_norm": 0.07616739720106125, | |
| "learning_rate": 1.9965864450080565e-06, | |
| "loss": 0.7209, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.160933917031478, | |
| "grad_norm": 0.0808212161064148, | |
| "learning_rate": 1.996511877393393e-06, | |
| "loss": 0.7229, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.1617677715238691, | |
| "grad_norm": 0.07879424095153809, | |
| "learning_rate": 1.996436505528986e-06, | |
| "loss": 0.6753, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.16260162601626016, | |
| "grad_norm": 0.07811419665813446, | |
| "learning_rate": 1.9963603294756654e-06, | |
| "loss": 0.6427, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.16343548050865125, | |
| "grad_norm": 0.08347994834184647, | |
| "learning_rate": 1.996283349294911e-06, | |
| "loss": 0.6575, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.16426933500104232, | |
| "grad_norm": 0.08415097743272781, | |
| "learning_rate": 1.99620556504885e-06, | |
| "loss": 0.6859, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.1651031894934334, | |
| "grad_norm": 0.08112502098083496, | |
| "learning_rate": 1.9961269768002596e-06, | |
| "loss": 0.6601, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.16593704398582448, | |
| "grad_norm": 0.07908611744642258, | |
| "learning_rate": 1.9960475846125656e-06, | |
| "loss": 0.6697, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.16677089847821555, | |
| "grad_norm": 0.07735829800367355, | |
| "learning_rate": 1.995967388549843e-06, | |
| "loss": 0.6631, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.16760475297060662, | |
| "grad_norm": 0.07564505934715271, | |
| "learning_rate": 1.9958863886768145e-06, | |
| "loss": 0.7004, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.1684386074629977, | |
| "grad_norm": 0.08117896318435669, | |
| "learning_rate": 1.9958045850588527e-06, | |
| "loss": 0.62, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.16927246195538878, | |
| "grad_norm": 0.0761483758687973, | |
| "learning_rate": 1.9957219777619784e-06, | |
| "loss": 0.7145, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.17010631644777988, | |
| "grad_norm": 0.08249073475599289, | |
| "learning_rate": 1.995638566852861e-06, | |
| "loss": 0.703, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.17094017094017094, | |
| "grad_norm": 0.08197219669818878, | |
| "learning_rate": 1.9955543523988187e-06, | |
| "loss": 0.695, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.171774025432562, | |
| "grad_norm": 0.07933253049850464, | |
| "learning_rate": 1.995469334467818e-06, | |
| "loss": 0.6575, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.1726078799249531, | |
| "grad_norm": 0.07888033986091614, | |
| "learning_rate": 1.9953835131284735e-06, | |
| "loss": 0.6504, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.17344173441734417, | |
| "grad_norm": 0.080783411860466, | |
| "learning_rate": 1.995296888450049e-06, | |
| "loss": 0.6586, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.17427558890973524, | |
| "grad_norm": 0.07863239198923111, | |
| "learning_rate": 1.995209460502456e-06, | |
| "loss": 0.7204, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.17510944340212634, | |
| "grad_norm": 0.07864314317703247, | |
| "learning_rate": 1.9951212293562544e-06, | |
| "loss": 0.6867, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.1759432978945174, | |
| "grad_norm": 0.07574496418237686, | |
| "learning_rate": 1.995032195082653e-06, | |
| "loss": 0.6806, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.1767771523869085, | |
| "grad_norm": 0.08079942315816879, | |
| "learning_rate": 1.9949423577535078e-06, | |
| "loss": 0.6929, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.17761100687929957, | |
| "grad_norm": 0.08361431211233139, | |
| "learning_rate": 1.9948517174413235e-06, | |
| "loss": 0.6948, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.17844486137169063, | |
| "grad_norm": 0.07974692434072495, | |
| "learning_rate": 1.994760274219253e-06, | |
| "loss": 0.7269, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.17927871586408173, | |
| "grad_norm": 0.08275672793388367, | |
| "learning_rate": 1.9946680281610957e-06, | |
| "loss": 0.6268, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.1801125703564728, | |
| "grad_norm": 0.08361789584159851, | |
| "learning_rate": 1.9945749793413017e-06, | |
| "loss": 0.6145, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.18094642484886386, | |
| "grad_norm": 0.0829552635550499, | |
| "learning_rate": 1.9944811278349665e-06, | |
| "loss": 0.7203, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.18178027934125496, | |
| "grad_norm": 0.07961665093898773, | |
| "learning_rate": 1.994386473717835e-06, | |
| "loss": 0.6598, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.18261413383364603, | |
| "grad_norm": 0.08015939593315125, | |
| "learning_rate": 1.9942910170662986e-06, | |
| "loss": 0.6465, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.18344798832603712, | |
| "grad_norm": 0.08055105805397034, | |
| "learning_rate": 1.994194757957397e-06, | |
| "loss": 0.6543, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.1842818428184282, | |
| "grad_norm": 0.07763723284006119, | |
| "learning_rate": 1.994097696468818e-06, | |
| "loss": 0.6984, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 0.18511569731081925, | |
| "grad_norm": 0.07882793247699738, | |
| "learning_rate": 1.9939998326788966e-06, | |
| "loss": 0.6672, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.18594955180321035, | |
| "grad_norm": 0.08098764717578888, | |
| "learning_rate": 1.993901166666615e-06, | |
| "loss": 0.65, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 0.18678340629560142, | |
| "grad_norm": 0.07791118323802948, | |
| "learning_rate": 1.9938016985116027e-06, | |
| "loss": 0.6954, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.18761726078799248, | |
| "grad_norm": 0.07814016193151474, | |
| "learning_rate": 1.993701428294137e-06, | |
| "loss": 0.6788, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.18845111528038358, | |
| "grad_norm": 0.07884248346090317, | |
| "learning_rate": 1.9936003560951433e-06, | |
| "loss": 0.6577, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.18928496977277465, | |
| "grad_norm": 0.07807621359825134, | |
| "learning_rate": 1.9934984819961924e-06, | |
| "loss": 0.6424, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 0.19011882426516574, | |
| "grad_norm": 0.08191139996051788, | |
| "learning_rate": 1.9933958060795043e-06, | |
| "loss": 0.6809, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.1909526787575568, | |
| "grad_norm": 0.07597807049751282, | |
| "learning_rate": 1.9932923284279445e-06, | |
| "loss": 0.7109, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 0.19178653324994788, | |
| "grad_norm": 0.08211328089237213, | |
| "learning_rate": 1.9931880491250263e-06, | |
| "loss": 0.6916, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.19262038774233897, | |
| "grad_norm": 0.07775567471981049, | |
| "learning_rate": 1.993082968254909e-06, | |
| "loss": 0.6458, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.19345424223473004, | |
| "grad_norm": 0.08438362181186676, | |
| "learning_rate": 1.992977085902402e-06, | |
| "loss": 0.6946, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.1942880967271211, | |
| "grad_norm": 0.08631843328475952, | |
| "learning_rate": 1.9928704021529564e-06, | |
| "loss": 0.6643, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 0.1951219512195122, | |
| "grad_norm": 0.07766929268836975, | |
| "learning_rate": 1.992762917092675e-06, | |
| "loss": 0.6753, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.19595580571190327, | |
| "grad_norm": 0.07574877887964249, | |
| "learning_rate": 1.9926546308083044e-06, | |
| "loss": 0.6346, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.19678966020429436, | |
| "grad_norm": 0.07815206050872803, | |
| "learning_rate": 1.9925455433872386e-06, | |
| "loss": 0.6946, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.19762351469668543, | |
| "grad_norm": 0.07457253336906433, | |
| "learning_rate": 1.9924356549175185e-06, | |
| "loss": 0.7108, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.1984573691890765, | |
| "grad_norm": 0.07688765227794647, | |
| "learning_rate": 1.992324965487831e-06, | |
| "loss": 0.6764, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.1992912236814676, | |
| "grad_norm": 0.08050046861171722, | |
| "learning_rate": 1.99221347518751e-06, | |
| "loss": 0.6263, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 0.20012507817385866, | |
| "grad_norm": 0.07763144373893738, | |
| "learning_rate": 1.9921011841065347e-06, | |
| "loss": 0.6618, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.20095893266624973, | |
| "grad_norm": 0.07744456082582474, | |
| "learning_rate": 1.991988092335532e-06, | |
| "loss": 0.6922, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 0.20179278715864082, | |
| "grad_norm": 0.07596256583929062, | |
| "learning_rate": 1.9918741999657743e-06, | |
| "loss": 0.7267, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.2026266416510319, | |
| "grad_norm": 0.08003415167331696, | |
| "learning_rate": 1.9917595070891793e-06, | |
| "loss": 0.6796, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.203460496143423, | |
| "grad_norm": 0.08418845385313034, | |
| "learning_rate": 1.9916440137983125e-06, | |
| "loss": 0.6815, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.20429435063581405, | |
| "grad_norm": 0.0794699415564537, | |
| "learning_rate": 1.991527720186384e-06, | |
| "loss": 0.6962, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.20512820512820512, | |
| "grad_norm": 0.07894127070903778, | |
| "learning_rate": 1.991410626347251e-06, | |
| "loss": 0.6898, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.20596205962059622, | |
| "grad_norm": 0.0819648951292038, | |
| "learning_rate": 1.9912927323754145e-06, | |
| "loss": 0.726, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 0.20679591411298728, | |
| "grad_norm": 0.08042652159929276, | |
| "learning_rate": 1.991174038366024e-06, | |
| "loss": 0.6398, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.20762976860537835, | |
| "grad_norm": 0.07848809659481049, | |
| "learning_rate": 1.9910545444148722e-06, | |
| "loss": 0.6562, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.20846362309776945, | |
| "grad_norm": 0.08295590430498123, | |
| "learning_rate": 1.9909342506183987e-06, | |
| "loss": 0.6767, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.2092974775901605, | |
| "grad_norm": 0.0840587317943573, | |
| "learning_rate": 1.9908131570736887e-06, | |
| "loss": 0.7274, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 0.2101313320825516, | |
| "grad_norm": 0.08375398814678192, | |
| "learning_rate": 1.9906912638784727e-06, | |
| "loss": 0.6716, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.21096518657494268, | |
| "grad_norm": 0.08229872584342957, | |
| "learning_rate": 1.990568571131126e-06, | |
| "loss": 0.7385, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 0.21179904106733374, | |
| "grad_norm": 0.07736154645681381, | |
| "learning_rate": 1.99044507893067e-06, | |
| "loss": 0.71, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 0.21263289555972484, | |
| "grad_norm": 0.08812221139669418, | |
| "learning_rate": 1.9903207873767704e-06, | |
| "loss": 0.7045, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.2134667500521159, | |
| "grad_norm": 0.08660853654146194, | |
| "learning_rate": 1.9901956965697386e-06, | |
| "loss": 0.7218, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.21430060454450697, | |
| "grad_norm": 0.08562801033258438, | |
| "learning_rate": 1.9900698066105318e-06, | |
| "loss": 0.6376, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 0.21513445903689807, | |
| "grad_norm": 0.07764414697885513, | |
| "learning_rate": 1.9899431176007503e-06, | |
| "loss": 0.6629, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.21596831352928914, | |
| "grad_norm": 0.07547292113304138, | |
| "learning_rate": 1.989815629642641e-06, | |
| "loss": 0.6887, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 0.21680216802168023, | |
| "grad_norm": 0.08463111519813538, | |
| "learning_rate": 1.9896873428390948e-06, | |
| "loss": 0.6169, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.2176360225140713, | |
| "grad_norm": 0.08064530789852142, | |
| "learning_rate": 1.9895582572936473e-06, | |
| "loss": 0.6905, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 0.21846987700646237, | |
| "grad_norm": 0.07929642498493195, | |
| "learning_rate": 1.9894283731104796e-06, | |
| "loss": 0.6415, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 0.21930373149885346, | |
| "grad_norm": 0.07858303934335709, | |
| "learning_rate": 1.9892976903944154e-06, | |
| "loss": 0.6955, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 0.22013758599124453, | |
| "grad_norm": 0.07967871427536011, | |
| "learning_rate": 1.9891662092509255e-06, | |
| "loss": 0.7004, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.2209714404836356, | |
| "grad_norm": 0.08009492605924606, | |
| "learning_rate": 1.989033929786123e-06, | |
| "loss": 0.6206, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.2218052949760267, | |
| "grad_norm": 0.08245829492807388, | |
| "learning_rate": 1.988900852106766e-06, | |
| "loss": 0.7182, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 0.22263914946841776, | |
| "grad_norm": 0.07844138890504837, | |
| "learning_rate": 1.9887669763202568e-06, | |
| "loss": 0.6372, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 0.22347300396080885, | |
| "grad_norm": 0.0783441960811615, | |
| "learning_rate": 1.9886323025346417e-06, | |
| "loss": 0.6182, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.22430685845319992, | |
| "grad_norm": 0.0803346335887909, | |
| "learning_rate": 1.9884968308586117e-06, | |
| "loss": 0.6436, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 0.225140712945591, | |
| "grad_norm": 0.0813342034816742, | |
| "learning_rate": 1.9883605614015015e-06, | |
| "loss": 0.6472, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.22597456743798208, | |
| "grad_norm": 0.07891476154327393, | |
| "learning_rate": 1.988223494273288e-06, | |
| "loss": 0.6625, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 0.22680842193037315, | |
| "grad_norm": 0.08087664842605591, | |
| "learning_rate": 1.9880856295845945e-06, | |
| "loss": 0.7226, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.22764227642276422, | |
| "grad_norm": 0.08254389464855194, | |
| "learning_rate": 1.9879469674466865e-06, | |
| "loss": 0.6992, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 0.2284761309151553, | |
| "grad_norm": 0.08207474648952484, | |
| "learning_rate": 1.987807507971473e-06, | |
| "loss": 0.6827, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 0.22930998540754638, | |
| "grad_norm": 0.08394856005907059, | |
| "learning_rate": 1.9876672512715075e-06, | |
| "loss": 0.7445, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.23014383989993747, | |
| "grad_norm": 0.07934897392988205, | |
| "learning_rate": 1.9875261974599856e-06, | |
| "loss": 0.6988, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.23097769439232854, | |
| "grad_norm": 0.08756977319717407, | |
| "learning_rate": 1.987384346650747e-06, | |
| "loss": 0.6588, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 0.2318115488847196, | |
| "grad_norm": 0.08198294788599014, | |
| "learning_rate": 1.9872416989582747e-06, | |
| "loss": 0.6729, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 0.2326454033771107, | |
| "grad_norm": 0.07671192288398743, | |
| "learning_rate": 1.9870982544976948e-06, | |
| "loss": 0.6685, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 0.23347925786950177, | |
| "grad_norm": 0.07943634688854218, | |
| "learning_rate": 1.986954013384776e-06, | |
| "loss": 0.7352, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.23431311236189284, | |
| "grad_norm": 0.0829697921872139, | |
| "learning_rate": 1.98680897573593e-06, | |
| "loss": 0.6645, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 0.23514696685428393, | |
| "grad_norm": 0.08583538234233856, | |
| "learning_rate": 1.986663141668212e-06, | |
| "loss": 0.713, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.235980821346675, | |
| "grad_norm": 0.08206543326377869, | |
| "learning_rate": 1.9865165112993192e-06, | |
| "loss": 0.7211, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 0.2368146758390661, | |
| "grad_norm": 0.07899665087461472, | |
| "learning_rate": 1.9863690847475924e-06, | |
| "loss": 0.6757, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.23764853033145716, | |
| "grad_norm": 0.0813852995634079, | |
| "learning_rate": 1.986220862132014e-06, | |
| "loss": 0.692, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.23848238482384823, | |
| "grad_norm": 0.09079395979642868, | |
| "learning_rate": 1.9860718435722093e-06, | |
| "loss": 0.6756, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.23931623931623933, | |
| "grad_norm": 0.07518535107374191, | |
| "learning_rate": 1.985922029188446e-06, | |
| "loss": 0.626, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 0.2401500938086304, | |
| "grad_norm": 0.08228084444999695, | |
| "learning_rate": 1.9857714191016337e-06, | |
| "loss": 0.7394, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.24098394830102146, | |
| "grad_norm": 0.08036590367555618, | |
| "learning_rate": 1.985620013433325e-06, | |
| "loss": 0.6012, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 0.24181780279341256, | |
| "grad_norm": 0.08429574221372604, | |
| "learning_rate": 1.985467812305714e-06, | |
| "loss": 0.6881, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.24265165728580362, | |
| "grad_norm": 0.07866602391004562, | |
| "learning_rate": 1.9853148158416364e-06, | |
| "loss": 0.6708, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 0.24348551177819472, | |
| "grad_norm": 0.0796622484922409, | |
| "learning_rate": 1.9851610241645714e-06, | |
| "loss": 0.7129, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.24431936627058579, | |
| "grad_norm": 0.08136799931526184, | |
| "learning_rate": 1.9850064373986374e-06, | |
| "loss": 0.6806, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 0.24515322076297685, | |
| "grad_norm": 0.07808342576026917, | |
| "learning_rate": 1.9848510556685973e-06, | |
| "loss": 0.6217, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 0.24598707525536795, | |
| "grad_norm": 0.07293733209371567, | |
| "learning_rate": 1.984694879099853e-06, | |
| "loss": 0.662, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.24682092974775902, | |
| "grad_norm": 0.07730741053819656, | |
| "learning_rate": 1.98453790781845e-06, | |
| "loss": 0.6386, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 0.24765478424015008, | |
| "grad_norm": 0.08601437509059906, | |
| "learning_rate": 1.9843801419510743e-06, | |
| "loss": 0.6639, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 0.24848863873254118, | |
| "grad_norm": 0.0815718024969101, | |
| "learning_rate": 1.9842215816250525e-06, | |
| "loss": 0.6891, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 0.24932249322493225, | |
| "grad_norm": 0.08793067187070847, | |
| "learning_rate": 1.9840622269683537e-06, | |
| "loss": 0.684, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 0.2501563477173233, | |
| "grad_norm": 0.07922165095806122, | |
| "learning_rate": 1.9839020781095873e-06, | |
| "loss": 0.6913, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.2509902022097144, | |
| "grad_norm": 0.0789838656783104, | |
| "learning_rate": 1.9837411351780035e-06, | |
| "loss": 0.6556, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 0.2518240567021055, | |
| "grad_norm": 0.07958388328552246, | |
| "learning_rate": 1.9835793983034944e-06, | |
| "loss": 0.6301, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 0.25265791119449654, | |
| "grad_norm": 0.08036024868488312, | |
| "learning_rate": 1.9834168676165915e-06, | |
| "loss": 0.592, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 0.25349176568688764, | |
| "grad_norm": 0.08134333044290543, | |
| "learning_rate": 1.9832535432484677e-06, | |
| "loss": 0.7142, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 0.25432562017927873, | |
| "grad_norm": 0.08333199471235275, | |
| "learning_rate": 1.983089425330937e-06, | |
| "loss": 0.7021, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.2551594746716698, | |
| "grad_norm": 0.08486991375684738, | |
| "learning_rate": 1.982924513996452e-06, | |
| "loss": 0.6724, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 0.25599332916406087, | |
| "grad_norm": 0.08286510407924652, | |
| "learning_rate": 1.982758809378108e-06, | |
| "loss": 0.6856, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 0.25682718365645196, | |
| "grad_norm": 0.08175141364336014, | |
| "learning_rate": 1.982592311609639e-06, | |
| "loss": 0.6605, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 0.257661038148843, | |
| "grad_norm": 0.08094872534275055, | |
| "learning_rate": 1.982425020825419e-06, | |
| "loss": 0.7136, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 0.2584948926412341, | |
| "grad_norm": 0.07797659933567047, | |
| "learning_rate": 1.9822569371604632e-06, | |
| "loss": 0.6975, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.2593287471336252, | |
| "grad_norm": 0.08523422479629517, | |
| "learning_rate": 1.982088060750426e-06, | |
| "loss": 0.6487, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 0.2601626016260163, | |
| "grad_norm": 0.09064790606498718, | |
| "learning_rate": 1.981918391731601e-06, | |
| "loss": 0.7006, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.2609964561184073, | |
| "grad_norm": 0.08136545121669769, | |
| "learning_rate": 1.9817479302409225e-06, | |
| "loss": 0.67, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 0.2618303106107984, | |
| "grad_norm": 0.07847245037555695, | |
| "learning_rate": 1.9815766764159642e-06, | |
| "loss": 0.6692, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 0.2626641651031895, | |
| "grad_norm": 0.07690361142158508, | |
| "learning_rate": 1.9814046303949387e-06, | |
| "loss": 0.5999, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.26349801959558056, | |
| "grad_norm": 0.08166715502738953, | |
| "learning_rate": 1.9812317923166987e-06, | |
| "loss": 0.6666, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 0.26433187408797165, | |
| "grad_norm": 0.08074827492237091, | |
| "learning_rate": 1.981058162320735e-06, | |
| "loss": 0.657, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 0.26516572858036275, | |
| "grad_norm": 0.08311966806650162, | |
| "learning_rate": 1.980883740547179e-06, | |
| "loss": 0.6679, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 0.2659995830727538, | |
| "grad_norm": 0.08272409439086914, | |
| "learning_rate": 1.9807085271368e-06, | |
| "loss": 0.6467, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 0.2668334375651449, | |
| "grad_norm": 0.08190209418535233, | |
| "learning_rate": 1.980532522231007e-06, | |
| "loss": 0.623, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.267667292057536, | |
| "grad_norm": 0.0819978415966034, | |
| "learning_rate": 1.980355725971847e-06, | |
| "loss": 0.677, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 0.268501146549927, | |
| "grad_norm": 0.07838036119937897, | |
| "learning_rate": 1.980178138502006e-06, | |
| "loss": 0.6545, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 0.2693350010423181, | |
| "grad_norm": 0.08094768226146698, | |
| "learning_rate": 1.979999759964809e-06, | |
| "loss": 0.6429, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 0.2701688555347092, | |
| "grad_norm": 0.0796407088637352, | |
| "learning_rate": 1.9798205905042184e-06, | |
| "loss": 0.7114, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 0.27100271002710025, | |
| "grad_norm": 0.07667295634746552, | |
| "learning_rate": 1.9796406302648366e-06, | |
| "loss": 0.7059, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.27183656451949134, | |
| "grad_norm": 0.07456690818071365, | |
| "learning_rate": 1.9794598793919023e-06, | |
| "loss": 0.5275, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 0.27267041901188244, | |
| "grad_norm": 0.08084844797849655, | |
| "learning_rate": 1.9792783380312935e-06, | |
| "loss": 0.6449, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 0.27350427350427353, | |
| "grad_norm": 0.08692745119333267, | |
| "learning_rate": 1.9790960063295254e-06, | |
| "loss": 0.6831, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 0.27433812799666457, | |
| "grad_norm": 0.0852961540222168, | |
| "learning_rate": 1.9789128844337524e-06, | |
| "loss": 0.6518, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 0.27517198248905567, | |
| "grad_norm": 0.08140264451503754, | |
| "learning_rate": 1.978728972491765e-06, | |
| "loss": 0.606, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.27600583698144676, | |
| "grad_norm": 0.08339618146419525, | |
| "learning_rate": 1.9785442706519927e-06, | |
| "loss": 0.6792, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 0.2768396914738378, | |
| "grad_norm": 0.08711510896682739, | |
| "learning_rate": 1.9783587790635017e-06, | |
| "loss": 0.6903, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 0.2776735459662289, | |
| "grad_norm": 0.08612138777971268, | |
| "learning_rate": 1.978172497875995e-06, | |
| "loss": 0.6998, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 0.27850740045862, | |
| "grad_norm": 0.07922092080116272, | |
| "learning_rate": 1.9779854272398146e-06, | |
| "loss": 0.654, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 0.27934125495101103, | |
| "grad_norm": 0.08098642528057098, | |
| "learning_rate": 1.9777975673059382e-06, | |
| "loss": 0.6706, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.2801751094434021, | |
| "grad_norm": 0.0873970165848732, | |
| "learning_rate": 1.977608918225981e-06, | |
| "loss": 0.6856, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 0.2810089639357932, | |
| "grad_norm": 0.082435242831707, | |
| "learning_rate": 1.9774194801521947e-06, | |
| "loss": 0.6375, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 0.28184281842818426, | |
| "grad_norm": 0.08199653774499893, | |
| "learning_rate": 1.9772292532374687e-06, | |
| "loss": 0.6785, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.28267667292057536, | |
| "grad_norm": 0.08817990124225616, | |
| "learning_rate": 1.9770382376353285e-06, | |
| "loss": 0.6642, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 0.28351052741296645, | |
| "grad_norm": 0.08359472453594208, | |
| "learning_rate": 1.976846433499935e-06, | |
| "loss": 0.6555, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.2843443819053575, | |
| "grad_norm": 0.08499454706907272, | |
| "learning_rate": 1.976653840986088e-06, | |
| "loss": 0.687, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 0.2851782363977486, | |
| "grad_norm": 0.08286130428314209, | |
| "learning_rate": 1.976460460249222e-06, | |
| "loss": 0.6904, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 0.2860120908901397, | |
| "grad_norm": 0.08244482427835464, | |
| "learning_rate": 1.976266291445406e-06, | |
| "loss": 0.6546, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 0.2868459453825308, | |
| "grad_norm": 0.07950432598590851, | |
| "learning_rate": 1.976071334731349e-06, | |
| "loss": 0.6683, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 0.2876797998749218, | |
| "grad_norm": 0.0854944959282875, | |
| "learning_rate": 1.975875590264393e-06, | |
| "loss": 0.6947, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.2885136543673129, | |
| "grad_norm": 0.08143036812543869, | |
| "learning_rate": 1.975679058202516e-06, | |
| "loss": 0.6422, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 0.289347508859704, | |
| "grad_norm": 0.0797310546040535, | |
| "learning_rate": 1.9754817387043325e-06, | |
| "loss": 0.6939, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 0.29018136335209505, | |
| "grad_norm": 0.08961952477693558, | |
| "learning_rate": 1.9752836319290926e-06, | |
| "loss": 0.6848, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 0.29101521784448614, | |
| "grad_norm": 0.08825525641441345, | |
| "learning_rate": 1.9750847380366806e-06, | |
| "loss": 0.7088, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 0.29184907233687724, | |
| "grad_norm": 0.08601095527410507, | |
| "learning_rate": 1.9748850571876168e-06, | |
| "loss": 0.6283, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.2926829268292683, | |
| "grad_norm": 0.08371152728796005, | |
| "learning_rate": 1.9746845895430568e-06, | |
| "loss": 0.6755, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 0.29351678132165937, | |
| "grad_norm": 0.08786465227603912, | |
| "learning_rate": 1.974483335264791e-06, | |
| "loss": 0.6149, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 0.29435063581405047, | |
| "grad_norm": 0.083786241710186, | |
| "learning_rate": 1.974281294515245e-06, | |
| "loss": 0.7104, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 0.2951844903064415, | |
| "grad_norm": 0.0846792683005333, | |
| "learning_rate": 1.9740784674574785e-06, | |
| "loss": 0.7032, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 0.2960183447988326, | |
| "grad_norm": 0.08477747440338135, | |
| "learning_rate": 1.973874854255186e-06, | |
| "loss": 0.6244, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.2968521992912237, | |
| "grad_norm": 0.09119998663663864, | |
| "learning_rate": 1.9736704550726965e-06, | |
| "loss": 0.6373, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 0.29768605378361473, | |
| "grad_norm": 0.08474856615066528, | |
| "learning_rate": 1.9734652700749733e-06, | |
| "loss": 0.736, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 0.29851990827600583, | |
| "grad_norm": 0.08070901036262512, | |
| "learning_rate": 1.9732592994276145e-06, | |
| "loss": 0.6442, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 0.2993537627683969, | |
| "grad_norm": 0.08306030929088593, | |
| "learning_rate": 1.9730525432968517e-06, | |
| "loss": 0.6863, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 0.300187617260788, | |
| "grad_norm": 0.08554356545209885, | |
| "learning_rate": 1.97284500184955e-06, | |
| "loss": 0.6782, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.30102147175317906, | |
| "grad_norm": 0.08350827544927597, | |
| "learning_rate": 1.9726366752532096e-06, | |
| "loss": 0.6656, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 0.30185532624557015, | |
| "grad_norm": 0.08501608669757843, | |
| "learning_rate": 1.972427563675963e-06, | |
| "loss": 0.6651, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 0.30268918073796125, | |
| "grad_norm": 0.08401606976985931, | |
| "learning_rate": 1.972217667286577e-06, | |
| "loss": 0.7153, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 0.3035230352303523, | |
| "grad_norm": 0.08713418990373611, | |
| "learning_rate": 1.9720069862544513e-06, | |
| "loss": 0.6623, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.3043568897227434, | |
| "grad_norm": 0.0871342197060585, | |
| "learning_rate": 1.9717955207496197e-06, | |
| "loss": 0.6571, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.3051907442151345, | |
| "grad_norm": 0.08501847088336945, | |
| "learning_rate": 1.9715832709427477e-06, | |
| "loss": 0.6714, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 0.3060245987075255, | |
| "grad_norm": 0.07894369214773178, | |
| "learning_rate": 1.971370237005136e-06, | |
| "loss": 0.6737, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 0.3068584531999166, | |
| "grad_norm": 0.08392394334077835, | |
| "learning_rate": 1.9711564191087157e-06, | |
| "loss": 0.6807, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 0.3076923076923077, | |
| "grad_norm": 0.08396563678979874, | |
| "learning_rate": 1.970941817426052e-06, | |
| "loss": 0.6766, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 0.30852616218469875, | |
| "grad_norm": 0.07790801674127579, | |
| "learning_rate": 1.9707264321303425e-06, | |
| "loss": 0.6293, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.30936001667708984, | |
| "grad_norm": 0.08699779212474823, | |
| "learning_rate": 1.970510263395417e-06, | |
| "loss": 0.681, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 0.31019387116948094, | |
| "grad_norm": 0.08176835626363754, | |
| "learning_rate": 1.9702933113957384e-06, | |
| "loss": 0.6931, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 0.311027725661872, | |
| "grad_norm": 0.08946527540683746, | |
| "learning_rate": 1.9700755763063997e-06, | |
| "loss": 0.6062, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 0.3118615801542631, | |
| "grad_norm": 0.08260657638311386, | |
| "learning_rate": 1.9698570583031284e-06, | |
| "loss": 0.6466, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 0.31269543464665417, | |
| "grad_norm": 0.08951833099126816, | |
| "learning_rate": 1.969637757562282e-06, | |
| "loss": 0.6458, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.31352928913904526, | |
| "grad_norm": 0.08605215698480606, | |
| "learning_rate": 1.9694176742608507e-06, | |
| "loss": 0.6919, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 0.3143631436314363, | |
| "grad_norm": 0.08389750868082047, | |
| "learning_rate": 1.969196808576456e-06, | |
| "loss": 0.6941, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 0.3151969981238274, | |
| "grad_norm": 0.08429732173681259, | |
| "learning_rate": 1.968975160687351e-06, | |
| "loss": 0.622, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 0.3160308526162185, | |
| "grad_norm": 0.08689334988594055, | |
| "learning_rate": 1.9687527307724195e-06, | |
| "loss": 0.6235, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 0.31686470710860953, | |
| "grad_norm": 0.08494516462087631, | |
| "learning_rate": 1.9685295190111774e-06, | |
| "loss": 0.6491, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.31769856160100063, | |
| "grad_norm": 0.0890749841928482, | |
| "learning_rate": 1.9683055255837708e-06, | |
| "loss": 0.6791, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 0.3185324160933917, | |
| "grad_norm": 0.07894483208656311, | |
| "learning_rate": 1.968080750670977e-06, | |
| "loss": 0.7075, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 0.31936627058578276, | |
| "grad_norm": 0.08575093746185303, | |
| "learning_rate": 1.9678551944542033e-06, | |
| "loss": 0.7027, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 0.32020012507817386, | |
| "grad_norm": 0.08360203355550766, | |
| "learning_rate": 1.9676288571154895e-06, | |
| "loss": 0.5894, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 0.32103397957056495, | |
| "grad_norm": 0.08341158181428909, | |
| "learning_rate": 1.9674017388375036e-06, | |
| "loss": 0.658, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.321867834062956, | |
| "grad_norm": 0.0844089537858963, | |
| "learning_rate": 1.9671738398035446e-06, | |
| "loss": 0.6691, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 0.3227016885553471, | |
| "grad_norm": 0.08446256816387177, | |
| "learning_rate": 1.9669451601975426e-06, | |
| "loss": 0.7211, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 0.3235355430477382, | |
| "grad_norm": 0.08008047193288803, | |
| "learning_rate": 1.9667157002040565e-06, | |
| "loss": 0.5843, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 0.3243693975401292, | |
| "grad_norm": 0.08821803331375122, | |
| "learning_rate": 1.9664854600082754e-06, | |
| "loss": 0.6444, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 0.3252032520325203, | |
| "grad_norm": 0.08548033237457275, | |
| "learning_rate": 1.966254439796018e-06, | |
| "loss": 0.6643, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.3260371065249114, | |
| "grad_norm": 0.08261854946613312, | |
| "learning_rate": 1.9660226397537322e-06, | |
| "loss": 0.6469, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 0.3268709610173025, | |
| "grad_norm": 0.07960055768489838, | |
| "learning_rate": 1.965790060068497e-06, | |
| "loss": 0.6149, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 0.32770481550969355, | |
| "grad_norm": 0.0834561213850975, | |
| "learning_rate": 1.9655567009280177e-06, | |
| "loss": 0.6798, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 0.32853867000208464, | |
| "grad_norm": 0.08524155616760254, | |
| "learning_rate": 1.9653225625206313e-06, | |
| "loss": 0.6999, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 0.32937252449447574, | |
| "grad_norm": 0.08503536134958267, | |
| "learning_rate": 1.965087645035302e-06, | |
| "loss": 0.6674, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.3302063789868668, | |
| "grad_norm": 0.09243112802505493, | |
| "learning_rate": 1.964851948661624e-06, | |
| "loss": 0.6391, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 0.3310402334792579, | |
| "grad_norm": 0.08015627413988113, | |
| "learning_rate": 1.96461547358982e-06, | |
| "loss": 0.611, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 0.33187408797164897, | |
| "grad_norm": 0.08681736886501312, | |
| "learning_rate": 1.9643782200107395e-06, | |
| "loss": 0.6967, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 0.33270794246404, | |
| "grad_norm": 0.08576110005378723, | |
| "learning_rate": 1.9641401881158623e-06, | |
| "loss": 0.6753, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 0.3335417969564311, | |
| "grad_norm": 0.08432309329509735, | |
| "learning_rate": 1.9639013780972954e-06, | |
| "loss": 0.6312, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.3335417969564311, | |
| "eval_loss": 0.6676867008209229, | |
| "eval_runtime": 321.9965, | |
| "eval_samples_per_second": 16.718, | |
| "eval_steps_per_second": 2.789, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.3343756514488222, | |
| "grad_norm": 0.08781838417053223, | |
| "learning_rate": 1.9636617901477742e-06, | |
| "loss": 0.698, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 0.33520950594121324, | |
| "grad_norm": 0.08723490685224533, | |
| "learning_rate": 1.9634214244606613e-06, | |
| "loss": 0.6642, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 0.33604336043360433, | |
| "grad_norm": 0.08249269425868988, | |
| "learning_rate": 1.963180281229948e-06, | |
| "loss": 0.705, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 0.3368772149259954, | |
| "grad_norm": 0.08193965256214142, | |
| "learning_rate": 1.962938360650252e-06, | |
| "loss": 0.6813, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 0.33771106941838647, | |
| "grad_norm": 0.08125729858875275, | |
| "learning_rate": 1.962695662916819e-06, | |
| "loss": 0.6571, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.33854492391077756, | |
| "grad_norm": 0.08508846163749695, | |
| "learning_rate": 1.962452188225522e-06, | |
| "loss": 0.6212, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 0.33937877840316866, | |
| "grad_norm": 0.0881478488445282, | |
| "learning_rate": 1.962207936772861e-06, | |
| "loss": 0.6582, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 0.34021263289555975, | |
| "grad_norm": 0.08948707580566406, | |
| "learning_rate": 1.9619629087559622e-06, | |
| "loss": 0.6924, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 0.3410464873879508, | |
| "grad_norm": 0.08214866369962692, | |
| "learning_rate": 1.9617171043725797e-06, | |
| "loss": 0.6266, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 0.3418803418803419, | |
| "grad_norm": 0.09406815469264984, | |
| "learning_rate": 1.961470523821093e-06, | |
| "loss": 0.6754, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.342714196372733, | |
| "grad_norm": 0.08819446712732315, | |
| "learning_rate": 1.961223167300509e-06, | |
| "loss": 0.6858, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 0.343548050865124, | |
| "grad_norm": 0.0902935117483139, | |
| "learning_rate": 1.9609750350104606e-06, | |
| "loss": 0.6654, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 0.3443819053575151, | |
| "grad_norm": 0.0852714329957962, | |
| "learning_rate": 1.9607261271512065e-06, | |
| "loss": 0.6604, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 0.3452157598499062, | |
| "grad_norm": 0.08480080962181091, | |
| "learning_rate": 1.9604764439236313e-06, | |
| "loss": 0.7252, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 0.34604961434229725, | |
| "grad_norm": 0.08467400074005127, | |
| "learning_rate": 1.960225985529246e-06, | |
| "loss": 0.6182, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.34688346883468835, | |
| "grad_norm": 0.09190120548009872, | |
| "learning_rate": 1.9599747521701865e-06, | |
| "loss": 0.6605, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.34771732332707944, | |
| "grad_norm": 0.08994212001562119, | |
| "learning_rate": 1.9597227440492143e-06, | |
| "loss": 0.6674, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 0.3485511778194705, | |
| "grad_norm": 0.08530712872743607, | |
| "learning_rate": 1.959469961369716e-06, | |
| "loss": 0.673, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 0.3493850323118616, | |
| "grad_norm": 0.0814647376537323, | |
| "learning_rate": 1.9592164043357046e-06, | |
| "loss": 0.6224, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 0.35021888680425267, | |
| "grad_norm": 0.08769946545362473, | |
| "learning_rate": 1.9589620731518164e-06, | |
| "loss": 0.6446, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.3510527412966437, | |
| "grad_norm": 0.0855315700173378, | |
| "learning_rate": 1.958706968023313e-06, | |
| "loss": 0.6665, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 0.3518865957890348, | |
| "grad_norm": 0.08764835447072983, | |
| "learning_rate": 1.958451089156082e-06, | |
| "loss": 0.6888, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 0.3527204502814259, | |
| "grad_norm": 0.08674878627061844, | |
| "learning_rate": 1.9581944367566324e-06, | |
| "loss": 0.6894, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 0.353554304773817, | |
| "grad_norm": 0.07654520124197006, | |
| "learning_rate": 1.9579370110321005e-06, | |
| "loss": 0.5972, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 0.35438815926620804, | |
| "grad_norm": 0.08671235293149948, | |
| "learning_rate": 1.9576788121902454e-06, | |
| "loss": 0.6705, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.35522201375859913, | |
| "grad_norm": 0.08799296617507935, | |
| "learning_rate": 1.9574198404394505e-06, | |
| "loss": 0.66, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 0.3560558682509902, | |
| "grad_norm": 0.08908937871456146, | |
| "learning_rate": 1.9571600959887224e-06, | |
| "loss": 0.6687, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 0.35688972274338127, | |
| "grad_norm": 0.08397683501243591, | |
| "learning_rate": 1.9568995790476915e-06, | |
| "loss": 0.5915, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 0.35772357723577236, | |
| "grad_norm": 0.08805875480175018, | |
| "learning_rate": 1.9566382898266126e-06, | |
| "loss": 0.6706, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 0.35855743172816346, | |
| "grad_norm": 0.08281008154153824, | |
| "learning_rate": 1.956376228536363e-06, | |
| "loss": 0.5692, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.3593912862205545, | |
| "grad_norm": 0.08514556288719177, | |
| "learning_rate": 1.9561133953884427e-06, | |
| "loss": 0.5938, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 0.3602251407129456, | |
| "grad_norm": 0.09112891554832458, | |
| "learning_rate": 1.955849790594975e-06, | |
| "loss": 0.7044, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 0.3610589952053367, | |
| "grad_norm": 0.08578234165906906, | |
| "learning_rate": 1.9555854143687064e-06, | |
| "loss": 0.6901, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 0.3618928496977277, | |
| "grad_norm": 0.08405105024576187, | |
| "learning_rate": 1.955320266923006e-06, | |
| "loss": 0.6307, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 0.3627267041901188, | |
| "grad_norm": 0.085452601313591, | |
| "learning_rate": 1.9550543484718644e-06, | |
| "loss": 0.6147, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.3635605586825099, | |
| "grad_norm": 0.0871538370847702, | |
| "learning_rate": 1.9547876592298955e-06, | |
| "loss": 0.6415, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 0.36439441317490096, | |
| "grad_norm": 0.08855723589658737, | |
| "learning_rate": 1.954520199412334e-06, | |
| "loss": 0.7026, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 0.36522826766729205, | |
| "grad_norm": 0.08221881091594696, | |
| "learning_rate": 1.9542519692350387e-06, | |
| "loss": 0.6813, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 0.36606212215968315, | |
| "grad_norm": 0.08595414459705353, | |
| "learning_rate": 1.9539829689144876e-06, | |
| "loss": 0.6809, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 0.36689597665207424, | |
| "grad_norm": 0.08485814183950424, | |
| "learning_rate": 1.9537131986677816e-06, | |
| "loss": 0.6352, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.3677298311444653, | |
| "grad_norm": 0.08620929718017578, | |
| "learning_rate": 1.9534426587126433e-06, | |
| "loss": 0.7072, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 0.3685636856368564, | |
| "grad_norm": 0.0919066071510315, | |
| "learning_rate": 1.9531713492674154e-06, | |
| "loss": 0.6809, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.36939754012924747, | |
| "grad_norm": 0.08376951515674591, | |
| "learning_rate": 1.9528992705510628e-06, | |
| "loss": 0.6486, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 0.3702313946216385, | |
| "grad_norm": 0.088445745408535, | |
| "learning_rate": 1.95262642278317e-06, | |
| "loss": 0.709, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 0.3710652491140296, | |
| "grad_norm": 0.08183503895998001, | |
| "learning_rate": 1.952352806183943e-06, | |
| "loss": 0.6277, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.3718991036064207, | |
| "grad_norm": 0.08638439327478409, | |
| "learning_rate": 1.9520784209742094e-06, | |
| "loss": 0.6665, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 0.37273295809881174, | |
| "grad_norm": 0.08837208896875381, | |
| "learning_rate": 1.951803267375414e-06, | |
| "loss": 0.6103, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 0.37356681259120283, | |
| "grad_norm": 0.08010973036289215, | |
| "learning_rate": 1.9515273456096247e-06, | |
| "loss": 0.662, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 0.37440066708359393, | |
| "grad_norm": 0.08395659178495407, | |
| "learning_rate": 1.9512506558995283e-06, | |
| "loss": 0.687, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 0.37523452157598497, | |
| "grad_norm": 0.08506524562835693, | |
| "learning_rate": 1.950973198468431e-06, | |
| "loss": 0.6695, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.37606837606837606, | |
| "grad_norm": 0.08150207251310349, | |
| "learning_rate": 1.950694973540259e-06, | |
| "loss": 0.6258, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 0.37690223056076716, | |
| "grad_norm": 0.08509568125009537, | |
| "learning_rate": 1.9504159813395576e-06, | |
| "loss": 0.5774, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 0.3777360850531582, | |
| "grad_norm": 0.08139500021934509, | |
| "learning_rate": 1.950136222091492e-06, | |
| "loss": 0.6652, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 0.3785699395455493, | |
| "grad_norm": 0.08891302347183228, | |
| "learning_rate": 1.949855696021846e-06, | |
| "loss": 0.6602, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 0.3794037940379404, | |
| "grad_norm": 0.08729292452335358, | |
| "learning_rate": 1.949574403357022e-06, | |
| "loss": 0.6232, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.3802376485303315, | |
| "grad_norm": 0.0855722650885582, | |
| "learning_rate": 1.9492923443240425e-06, | |
| "loss": 0.7249, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 0.3810715030227225, | |
| "grad_norm": 0.09225167334079742, | |
| "learning_rate": 1.949009519150546e-06, | |
| "loss": 0.7061, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 0.3819053575151136, | |
| "grad_norm": 0.08833880722522736, | |
| "learning_rate": 1.9487259280647918e-06, | |
| "loss": 0.6683, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 0.3827392120075047, | |
| "grad_norm": 0.08648336678743362, | |
| "learning_rate": 1.948441571295656e-06, | |
| "loss": 0.6716, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 0.38357306649989575, | |
| "grad_norm": 0.08678124845027924, | |
| "learning_rate": 1.9481564490726324e-06, | |
| "loss": 0.6869, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.38440692099228685, | |
| "grad_norm": 0.0866696834564209, | |
| "learning_rate": 1.9478705616258344e-06, | |
| "loss": 0.6346, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 0.38524077548467794, | |
| "grad_norm": 0.09009408950805664, | |
| "learning_rate": 1.9475839091859904e-06, | |
| "loss": 0.6507, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 0.386074629977069, | |
| "grad_norm": 0.09393912553787231, | |
| "learning_rate": 1.9472964919844484e-06, | |
| "loss": 0.6624, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 0.3869084844694601, | |
| "grad_norm": 0.09663153439760208, | |
| "learning_rate": 1.9470083102531722e-06, | |
| "loss": 0.6378, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 0.3877423389618512, | |
| "grad_norm": 0.08551555126905441, | |
| "learning_rate": 1.9467193642247435e-06, | |
| "loss": 0.6717, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.3885761934542422, | |
| "grad_norm": 0.08986209332942963, | |
| "learning_rate": 1.9464296541323602e-06, | |
| "loss": 0.7313, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 0.3894100479466333, | |
| "grad_norm": 0.08851300179958344, | |
| "learning_rate": 1.9461391802098373e-06, | |
| "loss": 0.6783, | |
| "step": 467 | |
| }, | |
| { | |
| "epoch": 0.3902439024390244, | |
| "grad_norm": 0.09088604897260666, | |
| "learning_rate": 1.9458479426916068e-06, | |
| "loss": 0.5952, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 0.39107775693141544, | |
| "grad_norm": 0.08506208658218384, | |
| "learning_rate": 1.9455559418127144e-06, | |
| "loss": 0.6912, | |
| "step": 469 | |
| }, | |
| { | |
| "epoch": 0.39191161142380654, | |
| "grad_norm": 0.087041936814785, | |
| "learning_rate": 1.945263177808826e-06, | |
| "loss": 0.6722, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.39274546591619763, | |
| "grad_norm": 0.09510099142789841, | |
| "learning_rate": 1.9449696509162193e-06, | |
| "loss": 0.697, | |
| "step": 471 | |
| }, | |
| { | |
| "epoch": 0.39357932040858873, | |
| "grad_norm": 0.0864386260509491, | |
| "learning_rate": 1.944675361371791e-06, | |
| "loss": 0.6882, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 0.39441317490097977, | |
| "grad_norm": 0.08964356780052185, | |
| "learning_rate": 1.944380309413051e-06, | |
| "loss": 0.6403, | |
| "step": 473 | |
| }, | |
| { | |
| "epoch": 0.39524702939337086, | |
| "grad_norm": 0.09786627441644669, | |
| "learning_rate": 1.9440844952781253e-06, | |
| "loss": 0.7037, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 0.39608088388576196, | |
| "grad_norm": 0.08709276467561722, | |
| "learning_rate": 1.9437879192057556e-06, | |
| "loss": 0.6619, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.396914738378153, | |
| "grad_norm": 0.08677015453577042, | |
| "learning_rate": 1.9434905814352976e-06, | |
| "loss": 0.6609, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 0.3977485928705441, | |
| "grad_norm": 0.08886487782001495, | |
| "learning_rate": 1.9431924822067225e-06, | |
| "loss": 0.6045, | |
| "step": 477 | |
| }, | |
| { | |
| "epoch": 0.3985824473629352, | |
| "grad_norm": 0.08575446903705597, | |
| "learning_rate": 1.942893621760616e-06, | |
| "loss": 0.6886, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 0.39941630185532623, | |
| "grad_norm": 0.08386967331171036, | |
| "learning_rate": 1.9425940003381767e-06, | |
| "loss": 0.6721, | |
| "step": 479 | |
| }, | |
| { | |
| "epoch": 0.4002501563477173, | |
| "grad_norm": 0.08702324330806732, | |
| "learning_rate": 1.9422936181812197e-06, | |
| "loss": 0.6959, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.4010840108401084, | |
| "grad_norm": 0.09090246260166168, | |
| "learning_rate": 1.9419924755321727e-06, | |
| "loss": 0.7094, | |
| "step": 481 | |
| }, | |
| { | |
| "epoch": 0.40191786533249946, | |
| "grad_norm": 0.09213659167289734, | |
| "learning_rate": 1.9416905726340767e-06, | |
| "loss": 0.6446, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 0.40275171982489055, | |
| "grad_norm": 0.09240783751010895, | |
| "learning_rate": 1.9413879097305878e-06, | |
| "loss": 0.6401, | |
| "step": 483 | |
| }, | |
| { | |
| "epoch": 0.40358557431728165, | |
| "grad_norm": 0.09058364480733871, | |
| "learning_rate": 1.941084487065974e-06, | |
| "loss": 0.7035, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 0.4044194288096727, | |
| "grad_norm": 0.0906047448515892, | |
| "learning_rate": 1.940780304885117e-06, | |
| "loss": 0.6519, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.4052532833020638, | |
| "grad_norm": 0.09101078659296036, | |
| "learning_rate": 1.940475363433512e-06, | |
| "loss": 0.6934, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 0.4060871377944549, | |
| "grad_norm": 0.08729363977909088, | |
| "learning_rate": 1.940169662957266e-06, | |
| "loss": 0.5822, | |
| "step": 487 | |
| }, | |
| { | |
| "epoch": 0.406920992286846, | |
| "grad_norm": 0.0946895033121109, | |
| "learning_rate": 1.9398632037031003e-06, | |
| "loss": 0.6534, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 0.407754846779237, | |
| "grad_norm": 0.09069128334522247, | |
| "learning_rate": 1.9395559859183463e-06, | |
| "loss": 0.6706, | |
| "step": 489 | |
| }, | |
| { | |
| "epoch": 0.4085887012716281, | |
| "grad_norm": 0.08318338543176651, | |
| "learning_rate": 1.9392480098509483e-06, | |
| "loss": 0.6368, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.4094225557640192, | |
| "grad_norm": 0.0826449766755104, | |
| "learning_rate": 1.9389392757494645e-06, | |
| "loss": 0.667, | |
| "step": 491 | |
| }, | |
| { | |
| "epoch": 0.41025641025641024, | |
| "grad_norm": 0.09276958554983139, | |
| "learning_rate": 1.938629783863062e-06, | |
| "loss": 0.6709, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 0.41109026474880134, | |
| "grad_norm": 0.08281219005584717, | |
| "learning_rate": 1.9383195344415215e-06, | |
| "loss": 0.584, | |
| "step": 493 | |
| }, | |
| { | |
| "epoch": 0.41192411924119243, | |
| "grad_norm": 0.08632051944732666, | |
| "learning_rate": 1.938008527735234e-06, | |
| "loss": 0.6743, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 0.41275797373358347, | |
| "grad_norm": 0.08970591425895691, | |
| "learning_rate": 1.937696763995202e-06, | |
| "loss": 0.6611, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.41359182822597457, | |
| "grad_norm": 0.0902903825044632, | |
| "learning_rate": 1.93738424347304e-06, | |
| "loss": 0.6455, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 0.41442568271836566, | |
| "grad_norm": 0.08964331448078156, | |
| "learning_rate": 1.9370709664209715e-06, | |
| "loss": 0.6561, | |
| "step": 497 | |
| }, | |
| { | |
| "epoch": 0.4152595372107567, | |
| "grad_norm": 0.08478616923093796, | |
| "learning_rate": 1.9367569330918317e-06, | |
| "loss": 0.595, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 0.4160933917031478, | |
| "grad_norm": 0.09221872687339783, | |
| "learning_rate": 1.9364421437390658e-06, | |
| "loss": 0.6363, | |
| "step": 499 | |
| }, | |
| { | |
| "epoch": 0.4169272461955389, | |
| "grad_norm": 0.09467694163322449, | |
| "learning_rate": 1.936126598616729e-06, | |
| "loss": 0.6696, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.41776110068792993, | |
| "grad_norm": 0.09464087337255478, | |
| "learning_rate": 1.935810297979487e-06, | |
| "loss": 0.6224, | |
| "step": 501 | |
| }, | |
| { | |
| "epoch": 0.418594955180321, | |
| "grad_norm": 0.08994987607002258, | |
| "learning_rate": 1.9354932420826147e-06, | |
| "loss": 0.6383, | |
| "step": 502 | |
| }, | |
| { | |
| "epoch": 0.4194288096727121, | |
| "grad_norm": 0.0921832025051117, | |
| "learning_rate": 1.9351754311819974e-06, | |
| "loss": 0.685, | |
| "step": 503 | |
| }, | |
| { | |
| "epoch": 0.4202626641651032, | |
| "grad_norm": 0.08833765983581543, | |
| "learning_rate": 1.934856865534129e-06, | |
| "loss": 0.6233, | |
| "step": 504 | |
| }, | |
| { | |
| "epoch": 0.42109651865749426, | |
| "grad_norm": 0.08863073587417603, | |
| "learning_rate": 1.934537545396111e-06, | |
| "loss": 0.6756, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 0.42193037314988535, | |
| "grad_norm": 0.08703204244375229, | |
| "learning_rate": 1.934217471025658e-06, | |
| "loss": 0.6372, | |
| "step": 506 | |
| }, | |
| { | |
| "epoch": 0.42276422764227645, | |
| "grad_norm": 0.09010464698076248, | |
| "learning_rate": 1.9338966426810887e-06, | |
| "loss": 0.6677, | |
| "step": 507 | |
| }, | |
| { | |
| "epoch": 0.4235980821346675, | |
| "grad_norm": 0.09014829248189926, | |
| "learning_rate": 1.9335750606213336e-06, | |
| "loss": 0.6839, | |
| "step": 508 | |
| }, | |
| { | |
| "epoch": 0.4244319366270586, | |
| "grad_norm": 0.08871164172887802, | |
| "learning_rate": 1.93325272510593e-06, | |
| "loss": 0.7012, | |
| "step": 509 | |
| }, | |
| { | |
| "epoch": 0.4252657911194497, | |
| "grad_norm": 0.08788250386714935, | |
| "learning_rate": 1.9329296363950236e-06, | |
| "loss": 0.6169, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.4260996456118407, | |
| "grad_norm": 0.08845999836921692, | |
| "learning_rate": 1.9326057947493675e-06, | |
| "loss": 0.6424, | |
| "step": 511 | |
| }, | |
| { | |
| "epoch": 0.4269335001042318, | |
| "grad_norm": 0.0858997106552124, | |
| "learning_rate": 1.9322812004303238e-06, | |
| "loss": 0.6389, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 0.4277673545966229, | |
| "grad_norm": 0.08850055187940598, | |
| "learning_rate": 1.9319558536998603e-06, | |
| "loss": 0.6975, | |
| "step": 513 | |
| }, | |
| { | |
| "epoch": 0.42860120908901395, | |
| "grad_norm": 0.09360900521278381, | |
| "learning_rate": 1.9316297548205534e-06, | |
| "loss": 0.686, | |
| "step": 514 | |
| }, | |
| { | |
| "epoch": 0.42943506358140504, | |
| "grad_norm": 0.08473866432905197, | |
| "learning_rate": 1.931302904055586e-06, | |
| "loss": 0.6517, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 0.43026891807379614, | |
| "grad_norm": 0.08656331151723862, | |
| "learning_rate": 1.9309753016687477e-06, | |
| "loss": 0.7003, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 0.4311027725661872, | |
| "grad_norm": 0.08407936990261078, | |
| "learning_rate": 1.9306469479244347e-06, | |
| "loss": 0.6611, | |
| "step": 517 | |
| }, | |
| { | |
| "epoch": 0.43193662705857827, | |
| "grad_norm": 0.09026416391134262, | |
| "learning_rate": 1.930317843087651e-06, | |
| "loss": 0.6749, | |
| "step": 518 | |
| }, | |
| { | |
| "epoch": 0.43277048155096937, | |
| "grad_norm": 0.10031198710203171, | |
| "learning_rate": 1.929987987424004e-06, | |
| "loss": 0.6403, | |
| "step": 519 | |
| }, | |
| { | |
| "epoch": 0.43360433604336046, | |
| "grad_norm": 0.09739411622285843, | |
| "learning_rate": 1.9296573811997086e-06, | |
| "loss": 0.6893, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.4344381905357515, | |
| "grad_norm": 0.08792299032211304, | |
| "learning_rate": 1.929326024681587e-06, | |
| "loss": 0.6268, | |
| "step": 521 | |
| }, | |
| { | |
| "epoch": 0.4352720450281426, | |
| "grad_norm": 0.0874318853020668, | |
| "learning_rate": 1.9289939181370646e-06, | |
| "loss": 0.6659, | |
| "step": 522 | |
| }, | |
| { | |
| "epoch": 0.4361058995205337, | |
| "grad_norm": 0.09086845815181732, | |
| "learning_rate": 1.9286610618341724e-06, | |
| "loss": 0.6399, | |
| "step": 523 | |
| }, | |
| { | |
| "epoch": 0.43693975401292473, | |
| "grad_norm": 0.09616074711084366, | |
| "learning_rate": 1.9283274560415477e-06, | |
| "loss": 0.6325, | |
| "step": 524 | |
| }, | |
| { | |
| "epoch": 0.4377736085053158, | |
| "grad_norm": 0.08674857765436172, | |
| "learning_rate": 1.9279931010284322e-06, | |
| "loss": 0.6487, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.4386074629977069, | |
| "grad_norm": 0.09142550826072693, | |
| "learning_rate": 1.927657997064671e-06, | |
| "loss": 0.6509, | |
| "step": 526 | |
| }, | |
| { | |
| "epoch": 0.43944131749009796, | |
| "grad_norm": 0.0838690921664238, | |
| "learning_rate": 1.9273221444207158e-06, | |
| "loss": 0.6087, | |
| "step": 527 | |
| }, | |
| { | |
| "epoch": 0.44027517198248906, | |
| "grad_norm": 0.0922529399394989, | |
| "learning_rate": 1.9269855433676213e-06, | |
| "loss": 0.6815, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 0.44110902647488015, | |
| "grad_norm": 0.08945809304714203, | |
| "learning_rate": 1.926648194177046e-06, | |
| "loss": 0.623, | |
| "step": 529 | |
| }, | |
| { | |
| "epoch": 0.4419428809672712, | |
| "grad_norm": 0.09663840383291245, | |
| "learning_rate": 1.926310097121253e-06, | |
| "loss": 0.6244, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.4427767354596623, | |
| "grad_norm": 0.0872710794210434, | |
| "learning_rate": 1.9259712524731083e-06, | |
| "loss": 0.6877, | |
| "step": 531 | |
| }, | |
| { | |
| "epoch": 0.4436105899520534, | |
| "grad_norm": 0.0860145092010498, | |
| "learning_rate": 1.9256316605060813e-06, | |
| "loss": 0.6458, | |
| "step": 532 | |
| }, | |
| { | |
| "epoch": 0.4444444444444444, | |
| "grad_norm": 0.08691411465406418, | |
| "learning_rate": 1.9252913214942456e-06, | |
| "loss": 0.6286, | |
| "step": 533 | |
| }, | |
| { | |
| "epoch": 0.4452782989368355, | |
| "grad_norm": 0.0899488553404808, | |
| "learning_rate": 1.9249502357122757e-06, | |
| "loss": 0.6154, | |
| "step": 534 | |
| }, | |
| { | |
| "epoch": 0.4461121534292266, | |
| "grad_norm": 0.08670518547296524, | |
| "learning_rate": 1.924608403435451e-06, | |
| "loss": 0.6189, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 0.4469460079216177, | |
| "grad_norm": 0.08927959948778152, | |
| "learning_rate": 1.9242658249396514e-06, | |
| "loss": 0.6591, | |
| "step": 536 | |
| }, | |
| { | |
| "epoch": 0.44777986241400874, | |
| "grad_norm": 0.09092804044485092, | |
| "learning_rate": 1.9239225005013607e-06, | |
| "loss": 0.6756, | |
| "step": 537 | |
| }, | |
| { | |
| "epoch": 0.44861371690639984, | |
| "grad_norm": 0.08876843005418777, | |
| "learning_rate": 1.923578430397664e-06, | |
| "loss": 0.6221, | |
| "step": 538 | |
| }, | |
| { | |
| "epoch": 0.44944757139879093, | |
| "grad_norm": 0.09023214131593704, | |
| "learning_rate": 1.923233614906248e-06, | |
| "loss": 0.6478, | |
| "step": 539 | |
| }, | |
| { | |
| "epoch": 0.450281425891182, | |
| "grad_norm": 0.09291035681962967, | |
| "learning_rate": 1.9228880543054006e-06, | |
| "loss": 0.6568, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.45111528038357307, | |
| "grad_norm": 0.09169748425483704, | |
| "learning_rate": 1.9225417488740127e-06, | |
| "loss": 0.6966, | |
| "step": 541 | |
| }, | |
| { | |
| "epoch": 0.45194913487596416, | |
| "grad_norm": 0.0960812196135521, | |
| "learning_rate": 1.922194698891574e-06, | |
| "loss": 0.6426, | |
| "step": 542 | |
| }, | |
| { | |
| "epoch": 0.4527829893683552, | |
| "grad_norm": 0.09381034225225449, | |
| "learning_rate": 1.9218469046381778e-06, | |
| "loss": 0.7138, | |
| "step": 543 | |
| }, | |
| { | |
| "epoch": 0.4536168438607463, | |
| "grad_norm": 0.09605712443590164, | |
| "learning_rate": 1.9214983663945157e-06, | |
| "loss": 0.6677, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 0.4544506983531374, | |
| "grad_norm": 0.09246627241373062, | |
| "learning_rate": 1.921149084441881e-06, | |
| "loss": 0.5995, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 0.45528455284552843, | |
| "grad_norm": 0.09615227580070496, | |
| "learning_rate": 1.9207990590621663e-06, | |
| "loss": 0.6543, | |
| "step": 546 | |
| }, | |
| { | |
| "epoch": 0.45611840733791953, | |
| "grad_norm": 0.09620420634746552, | |
| "learning_rate": 1.9204482905378654e-06, | |
| "loss": 0.5895, | |
| "step": 547 | |
| }, | |
| { | |
| "epoch": 0.4569522618303106, | |
| "grad_norm": 0.08878765255212784, | |
| "learning_rate": 1.920096779152071e-06, | |
| "loss": 0.599, | |
| "step": 548 | |
| }, | |
| { | |
| "epoch": 0.45778611632270166, | |
| "grad_norm": 0.09067462384700775, | |
| "learning_rate": 1.9197445251884763e-06, | |
| "loss": 0.6897, | |
| "step": 549 | |
| }, | |
| { | |
| "epoch": 0.45861997081509276, | |
| "grad_norm": 0.09180627018213272, | |
| "learning_rate": 1.9193915289313724e-06, | |
| "loss": 0.6647, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.45945382530748385, | |
| "grad_norm": 0.08824802190065384, | |
| "learning_rate": 1.91903779066565e-06, | |
| "loss": 0.6376, | |
| "step": 551 | |
| }, | |
| { | |
| "epoch": 0.46028767979987495, | |
| "grad_norm": 0.08835819363594055, | |
| "learning_rate": 1.918683310676799e-06, | |
| "loss": 0.6922, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 0.461121534292266, | |
| "grad_norm": 0.09096652269363403, | |
| "learning_rate": 1.918328089250908e-06, | |
| "loss": 0.7169, | |
| "step": 553 | |
| }, | |
| { | |
| "epoch": 0.4619553887846571, | |
| "grad_norm": 0.09415201842784882, | |
| "learning_rate": 1.917972126674664e-06, | |
| "loss": 0.659, | |
| "step": 554 | |
| }, | |
| { | |
| "epoch": 0.4627892432770482, | |
| "grad_norm": 0.08964891731739044, | |
| "learning_rate": 1.9176154232353513e-06, | |
| "loss": 0.6869, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 0.4636230977694392, | |
| "grad_norm": 0.08688930422067642, | |
| "learning_rate": 1.917257979220853e-06, | |
| "loss": 0.627, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 0.4644569522618303, | |
| "grad_norm": 0.09115028381347656, | |
| "learning_rate": 1.9168997949196496e-06, | |
| "loss": 0.64, | |
| "step": 557 | |
| }, | |
| { | |
| "epoch": 0.4652908067542214, | |
| "grad_norm": 0.09430664777755737, | |
| "learning_rate": 1.9165408706208184e-06, | |
| "loss": 0.6646, | |
| "step": 558 | |
| }, | |
| { | |
| "epoch": 0.46612466124661245, | |
| "grad_norm": 0.09062688797712326, | |
| "learning_rate": 1.916181206614036e-06, | |
| "loss": 0.6392, | |
| "step": 559 | |
| }, | |
| { | |
| "epoch": 0.46695851573900354, | |
| "grad_norm": 0.09099205583333969, | |
| "learning_rate": 1.9158208031895737e-06, | |
| "loss": 0.7133, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.46779237023139464, | |
| "grad_norm": 0.09944413602352142, | |
| "learning_rate": 1.9154596606383002e-06, | |
| "loss": 0.6958, | |
| "step": 561 | |
| }, | |
| { | |
| "epoch": 0.4686262247237857, | |
| "grad_norm": 0.09643740206956863, | |
| "learning_rate": 1.9150977792516816e-06, | |
| "loss": 0.6813, | |
| "step": 562 | |
| }, | |
| { | |
| "epoch": 0.4694600792161768, | |
| "grad_norm": 0.09532240033149719, | |
| "learning_rate": 1.914735159321779e-06, | |
| "loss": 0.6528, | |
| "step": 563 | |
| }, | |
| { | |
| "epoch": 0.47029393370856787, | |
| "grad_norm": 0.08927876502275467, | |
| "learning_rate": 1.914371801141251e-06, | |
| "loss": 0.639, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 0.4711277882009589, | |
| "grad_norm": 0.09561195224523544, | |
| "learning_rate": 1.914007705003351e-06, | |
| "loss": 0.6795, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 0.47196164269335, | |
| "grad_norm": 0.0882195308804512, | |
| "learning_rate": 1.9136428712019275e-06, | |
| "loss": 0.6437, | |
| "step": 566 | |
| }, | |
| { | |
| "epoch": 0.4727954971857411, | |
| "grad_norm": 0.08940508961677551, | |
| "learning_rate": 1.913277300031426e-06, | |
| "loss": 0.6485, | |
| "step": 567 | |
| }, | |
| { | |
| "epoch": 0.4736293516781322, | |
| "grad_norm": 0.0933694839477539, | |
| "learning_rate": 1.912910991786886e-06, | |
| "loss": 0.6876, | |
| "step": 568 | |
| }, | |
| { | |
| "epoch": 0.47446320617052323, | |
| "grad_norm": 0.09212841838598251, | |
| "learning_rate": 1.9125439467639414e-06, | |
| "loss": 0.5669, | |
| "step": 569 | |
| }, | |
| { | |
| "epoch": 0.47529706066291433, | |
| "grad_norm": 0.09037458896636963, | |
| "learning_rate": 1.9121761652588214e-06, | |
| "loss": 0.6489, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.4761309151553054, | |
| "grad_norm": 0.08847213536500931, | |
| "learning_rate": 1.9118076475683506e-06, | |
| "loss": 0.6456, | |
| "step": 571 | |
| }, | |
| { | |
| "epoch": 0.47696476964769646, | |
| "grad_norm": 0.08496098965406418, | |
| "learning_rate": 1.9114383939899455e-06, | |
| "loss": 0.6174, | |
| "step": 572 | |
| }, | |
| { | |
| "epoch": 0.47779862414008756, | |
| "grad_norm": 0.09430284053087234, | |
| "learning_rate": 1.911068404821618e-06, | |
| "loss": 0.7102, | |
| "step": 573 | |
| }, | |
| { | |
| "epoch": 0.47863247863247865, | |
| "grad_norm": 0.09211906790733337, | |
| "learning_rate": 1.910697680361974e-06, | |
| "loss": 0.6161, | |
| "step": 574 | |
| }, | |
| { | |
| "epoch": 0.4794663331248697, | |
| "grad_norm": 0.09092090278863907, | |
| "learning_rate": 1.910326220910211e-06, | |
| "loss": 0.6595, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 0.4803001876172608, | |
| "grad_norm": 0.09773701429367065, | |
| "learning_rate": 1.909954026766122e-06, | |
| "loss": 0.6303, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 0.4811340421096519, | |
| "grad_norm": 0.09343478083610535, | |
| "learning_rate": 1.909581098230091e-06, | |
| "loss": 0.6578, | |
| "step": 577 | |
| }, | |
| { | |
| "epoch": 0.4819678966020429, | |
| "grad_norm": 0.09686136245727539, | |
| "learning_rate": 1.9092074356030966e-06, | |
| "loss": 0.6504, | |
| "step": 578 | |
| }, | |
| { | |
| "epoch": 0.482801751094434, | |
| "grad_norm": 0.09207551926374435, | |
| "learning_rate": 1.9088330391867076e-06, | |
| "loss": 0.6092, | |
| "step": 579 | |
| }, | |
| { | |
| "epoch": 0.4836356055868251, | |
| "grad_norm": 0.093394935131073, | |
| "learning_rate": 1.908457909283087e-06, | |
| "loss": 0.645, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.48446946007921615, | |
| "grad_norm": 0.10280878096818924, | |
| "learning_rate": 1.9080820461949886e-06, | |
| "loss": 0.7288, | |
| "step": 581 | |
| }, | |
| { | |
| "epoch": 0.48530331457160725, | |
| "grad_norm": 0.09714383631944656, | |
| "learning_rate": 1.9077054502257585e-06, | |
| "loss": 0.6213, | |
| "step": 582 | |
| }, | |
| { | |
| "epoch": 0.48613716906399834, | |
| "grad_norm": 0.09004565328359604, | |
| "learning_rate": 1.9073281216793337e-06, | |
| "loss": 0.6459, | |
| "step": 583 | |
| }, | |
| { | |
| "epoch": 0.48697102355638944, | |
| "grad_norm": 0.12440581619739532, | |
| "learning_rate": 1.906950060860243e-06, | |
| "loss": 0.6823, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 0.4878048780487805, | |
| "grad_norm": 0.09314379096031189, | |
| "learning_rate": 1.9065712680736066e-06, | |
| "loss": 0.6475, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 0.48863873254117157, | |
| "grad_norm": 0.09512604027986526, | |
| "learning_rate": 1.906191743625134e-06, | |
| "loss": 0.5932, | |
| "step": 586 | |
| }, | |
| { | |
| "epoch": 0.48947258703356267, | |
| "grad_norm": 0.08977750688791275, | |
| "learning_rate": 1.9058114878211265e-06, | |
| "loss": 0.6546, | |
| "step": 587 | |
| }, | |
| { | |
| "epoch": 0.4903064415259537, | |
| "grad_norm": 0.09166161715984344, | |
| "learning_rate": 1.905430500968475e-06, | |
| "loss": 0.6772, | |
| "step": 588 | |
| }, | |
| { | |
| "epoch": 0.4911402960183448, | |
| "grad_norm": 0.10014794021844864, | |
| "learning_rate": 1.9050487833746609e-06, | |
| "loss": 0.6945, | |
| "step": 589 | |
| }, | |
| { | |
| "epoch": 0.4919741505107359, | |
| "grad_norm": 0.09376100450754166, | |
| "learning_rate": 1.9046663353477548e-06, | |
| "loss": 0.6581, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.49280800500312694, | |
| "grad_norm": 0.09505487233400345, | |
| "learning_rate": 1.9042831571964171e-06, | |
| "loss": 0.6785, | |
| "step": 591 | |
| }, | |
| { | |
| "epoch": 0.49364185949551803, | |
| "grad_norm": 0.09371201694011688, | |
| "learning_rate": 1.9038992492298975e-06, | |
| "loss": 0.6331, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 0.4944757139879091, | |
| "grad_norm": 0.09442009776830673, | |
| "learning_rate": 1.9035146117580348e-06, | |
| "loss": 0.5728, | |
| "step": 593 | |
| }, | |
| { | |
| "epoch": 0.49530956848030017, | |
| "grad_norm": 0.09772226214408875, | |
| "learning_rate": 1.903129245091256e-06, | |
| "loss": 0.6635, | |
| "step": 594 | |
| }, | |
| { | |
| "epoch": 0.49614342297269126, | |
| "grad_norm": 0.09484317898750305, | |
| "learning_rate": 1.9027431495405776e-06, | |
| "loss": 0.6327, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 0.49697727746508236, | |
| "grad_norm": 0.09658980369567871, | |
| "learning_rate": 1.9023563254176032e-06, | |
| "loss": 0.6608, | |
| "step": 596 | |
| }, | |
| { | |
| "epoch": 0.4978111319574734, | |
| "grad_norm": 0.09717069566249847, | |
| "learning_rate": 1.9019687730345251e-06, | |
| "loss": 0.6605, | |
| "step": 597 | |
| }, | |
| { | |
| "epoch": 0.4986449864498645, | |
| "grad_norm": 0.09019803255796432, | |
| "learning_rate": 1.9015804927041233e-06, | |
| "loss": 0.6867, | |
| "step": 598 | |
| }, | |
| { | |
| "epoch": 0.4994788409422556, | |
| "grad_norm": 0.09377589821815491, | |
| "learning_rate": 1.9011914847397653e-06, | |
| "loss": 0.6419, | |
| "step": 599 | |
| }, | |
| { | |
| "epoch": 0.5003126954346466, | |
| "grad_norm": 0.09151824563741684, | |
| "learning_rate": 1.9008017494554055e-06, | |
| "loss": 0.7184, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.5011465499270378, | |
| "grad_norm": 0.0963352620601654, | |
| "learning_rate": 1.900411287165586e-06, | |
| "loss": 0.6897, | |
| "step": 601 | |
| }, | |
| { | |
| "epoch": 0.5019804044194288, | |
| "grad_norm": 0.08974531292915344, | |
| "learning_rate": 1.9000200981854347e-06, | |
| "loss": 0.5847, | |
| "step": 602 | |
| }, | |
| { | |
| "epoch": 0.5028142589118199, | |
| "grad_norm": 0.09220883995294571, | |
| "learning_rate": 1.8996281828306665e-06, | |
| "loss": 0.6652, | |
| "step": 603 | |
| }, | |
| { | |
| "epoch": 0.503648113404211, | |
| "grad_norm": 0.09491855651140213, | |
| "learning_rate": 1.899235541417583e-06, | |
| "loss": 0.595, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 0.504481967896602, | |
| "grad_norm": 0.09616294503211975, | |
| "learning_rate": 1.898842174263071e-06, | |
| "loss": 0.6378, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 0.5053158223889931, | |
| "grad_norm": 0.08942185342311859, | |
| "learning_rate": 1.8984480816846035e-06, | |
| "loss": 0.6459, | |
| "step": 606 | |
| }, | |
| { | |
| "epoch": 0.5061496768813842, | |
| "grad_norm": 0.09059803187847137, | |
| "learning_rate": 1.8980532640002388e-06, | |
| "loss": 0.6019, | |
| "step": 607 | |
| }, | |
| { | |
| "epoch": 0.5069835313737753, | |
| "grad_norm": 0.09913185238838196, | |
| "learning_rate": 1.89765772152862e-06, | |
| "loss": 0.669, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 0.5078173858661663, | |
| "grad_norm": 0.10223888605833054, | |
| "learning_rate": 1.8972614545889756e-06, | |
| "loss": 0.6454, | |
| "step": 609 | |
| }, | |
| { | |
| "epoch": 0.5086512403585575, | |
| "grad_norm": 0.09271295368671417, | |
| "learning_rate": 1.896864463501119e-06, | |
| "loss": 0.6467, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.5094850948509485, | |
| "grad_norm": 0.08862827718257904, | |
| "learning_rate": 1.896466748585448e-06, | |
| "loss": 0.6343, | |
| "step": 611 | |
| }, | |
| { | |
| "epoch": 0.5103189493433395, | |
| "grad_norm": 0.09303930401802063, | |
| "learning_rate": 1.8960683101629435e-06, | |
| "loss": 0.6389, | |
| "step": 612 | |
| }, | |
| { | |
| "epoch": 0.5111528038357307, | |
| "grad_norm": 0.09331949055194855, | |
| "learning_rate": 1.8956691485551719e-06, | |
| "loss": 0.6507, | |
| "step": 613 | |
| }, | |
| { | |
| "epoch": 0.5119866583281217, | |
| "grad_norm": 0.0993635281920433, | |
| "learning_rate": 1.8952692640842825e-06, | |
| "loss": 0.63, | |
| "step": 614 | |
| }, | |
| { | |
| "epoch": 0.5128205128205128, | |
| "grad_norm": 0.09348301589488983, | |
| "learning_rate": 1.8948686570730074e-06, | |
| "loss": 0.676, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 0.5136543673129039, | |
| "grad_norm": 0.0950765460729599, | |
| "learning_rate": 1.894467327844663e-06, | |
| "loss": 0.7035, | |
| "step": 616 | |
| }, | |
| { | |
| "epoch": 0.514488221805295, | |
| "grad_norm": 0.08676007390022278, | |
| "learning_rate": 1.8940652767231483e-06, | |
| "loss": 0.6067, | |
| "step": 617 | |
| }, | |
| { | |
| "epoch": 0.515322076297686, | |
| "grad_norm": 0.09285329282283783, | |
| "learning_rate": 1.8936625040329438e-06, | |
| "loss": 0.6441, | |
| "step": 618 | |
| }, | |
| { | |
| "epoch": 0.5161559307900772, | |
| "grad_norm": 0.09361077845096588, | |
| "learning_rate": 1.8932590100991136e-06, | |
| "loss": 0.6767, | |
| "step": 619 | |
| }, | |
| { | |
| "epoch": 0.5169897852824682, | |
| "grad_norm": 0.09173277020454407, | |
| "learning_rate": 1.8928547952473033e-06, | |
| "loss": 0.6395, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.5178236397748592, | |
| "grad_norm": 0.09198316186666489, | |
| "learning_rate": 1.892449859803741e-06, | |
| "loss": 0.7006, | |
| "step": 621 | |
| }, | |
| { | |
| "epoch": 0.5186574942672504, | |
| "grad_norm": 0.09348779171705246, | |
| "learning_rate": 1.8920442040952356e-06, | |
| "loss": 0.6641, | |
| "step": 622 | |
| }, | |
| { | |
| "epoch": 0.5194913487596414, | |
| "grad_norm": 0.08798382431268692, | |
| "learning_rate": 1.8916378284491774e-06, | |
| "loss": 0.6478, | |
| "step": 623 | |
| }, | |
| { | |
| "epoch": 0.5203252032520326, | |
| "grad_norm": 0.09442038089036942, | |
| "learning_rate": 1.8912307331935383e-06, | |
| "loss": 0.6201, | |
| "step": 624 | |
| }, | |
| { | |
| "epoch": 0.5211590577444236, | |
| "grad_norm": 0.09519796818494797, | |
| "learning_rate": 1.8908229186568705e-06, | |
| "loss": 0.5917, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 0.5219929122368147, | |
| "grad_norm": 0.09593921154737473, | |
| "learning_rate": 1.8904143851683062e-06, | |
| "loss": 0.6859, | |
| "step": 626 | |
| }, | |
| { | |
| "epoch": 0.5228267667292058, | |
| "grad_norm": 0.09510339796543121, | |
| "learning_rate": 1.890005133057559e-06, | |
| "loss": 0.6679, | |
| "step": 627 | |
| }, | |
| { | |
| "epoch": 0.5236606212215968, | |
| "grad_norm": 0.09473798424005508, | |
| "learning_rate": 1.8895951626549222e-06, | |
| "loss": 0.6108, | |
| "step": 628 | |
| }, | |
| { | |
| "epoch": 0.5244944757139879, | |
| "grad_norm": 0.10544616729021072, | |
| "learning_rate": 1.8891844742912678e-06, | |
| "loss": 0.6007, | |
| "step": 629 | |
| }, | |
| { | |
| "epoch": 0.525328330206379, | |
| "grad_norm": 0.09652596712112427, | |
| "learning_rate": 1.8887730682980482e-06, | |
| "loss": 0.6858, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.5261621846987701, | |
| "grad_norm": 0.09385030716657639, | |
| "learning_rate": 1.888360945007295e-06, | |
| "loss": 0.692, | |
| "step": 631 | |
| }, | |
| { | |
| "epoch": 0.5269960391911611, | |
| "grad_norm": 0.10086268931627274, | |
| "learning_rate": 1.887948104751618e-06, | |
| "loss": 0.7302, | |
| "step": 632 | |
| }, | |
| { | |
| "epoch": 0.5278298936835523, | |
| "grad_norm": 0.09164178371429443, | |
| "learning_rate": 1.8875345478642065e-06, | |
| "loss": 0.5793, | |
| "step": 633 | |
| }, | |
| { | |
| "epoch": 0.5286637481759433, | |
| "grad_norm": 0.08686941117048264, | |
| "learning_rate": 1.8871202746788274e-06, | |
| "loss": 0.6304, | |
| "step": 634 | |
| }, | |
| { | |
| "epoch": 0.5294976026683343, | |
| "grad_norm": 0.09579890966415405, | |
| "learning_rate": 1.8867052855298264e-06, | |
| "loss": 0.6663, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 0.5303314571607255, | |
| "grad_norm": 0.09054259210824966, | |
| "learning_rate": 1.8862895807521264e-06, | |
| "loss": 0.6395, | |
| "step": 636 | |
| }, | |
| { | |
| "epoch": 0.5311653116531165, | |
| "grad_norm": 0.0913679376244545, | |
| "learning_rate": 1.8858731606812284e-06, | |
| "loss": 0.6481, | |
| "step": 637 | |
| }, | |
| { | |
| "epoch": 0.5319991661455076, | |
| "grad_norm": 0.09573955833911896, | |
| "learning_rate": 1.8854560256532098e-06, | |
| "loss": 0.6543, | |
| "step": 638 | |
| }, | |
| { | |
| "epoch": 0.5328330206378987, | |
| "grad_norm": 0.09667269140481949, | |
| "learning_rate": 1.8850381760047262e-06, | |
| "loss": 0.6532, | |
| "step": 639 | |
| }, | |
| { | |
| "epoch": 0.5336668751302898, | |
| "grad_norm": 0.09536033123731613, | |
| "learning_rate": 1.8846196120730093e-06, | |
| "loss": 0.6429, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.5345007296226808, | |
| "grad_norm": 0.09058564901351929, | |
| "learning_rate": 1.8842003341958673e-06, | |
| "loss": 0.579, | |
| "step": 641 | |
| }, | |
| { | |
| "epoch": 0.535334584115072, | |
| "grad_norm": 0.09930071234703064, | |
| "learning_rate": 1.8837803427116844e-06, | |
| "loss": 0.5868, | |
| "step": 642 | |
| }, | |
| { | |
| "epoch": 0.536168438607463, | |
| "grad_norm": 0.0978417843580246, | |
| "learning_rate": 1.8833596379594211e-06, | |
| "loss": 0.6822, | |
| "step": 643 | |
| }, | |
| { | |
| "epoch": 0.537002293099854, | |
| "grad_norm": 0.09346287697553635, | |
| "learning_rate": 1.882938220278614e-06, | |
| "loss": 0.6116, | |
| "step": 644 | |
| }, | |
| { | |
| "epoch": 0.5378361475922452, | |
| "grad_norm": 0.09151905030012131, | |
| "learning_rate": 1.8825160900093738e-06, | |
| "loss": 0.6496, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 0.5386700020846362, | |
| "grad_norm": 0.0970907062292099, | |
| "learning_rate": 1.8820932474923871e-06, | |
| "loss": 0.6504, | |
| "step": 646 | |
| }, | |
| { | |
| "epoch": 0.5395038565770273, | |
| "grad_norm": 0.09408881515264511, | |
| "learning_rate": 1.8816696930689159e-06, | |
| "loss": 0.6829, | |
| "step": 647 | |
| }, | |
| { | |
| "epoch": 0.5403377110694184, | |
| "grad_norm": 0.09966862946748734, | |
| "learning_rate": 1.8812454270807951e-06, | |
| "loss": 0.5898, | |
| "step": 648 | |
| }, | |
| { | |
| "epoch": 0.5411715655618095, | |
| "grad_norm": 0.0900130420923233, | |
| "learning_rate": 1.8808204498704356e-06, | |
| "loss": 0.6042, | |
| "step": 649 | |
| }, | |
| { | |
| "epoch": 0.5420054200542005, | |
| "grad_norm": 0.09694640338420868, | |
| "learning_rate": 1.8803947617808214e-06, | |
| "loss": 0.6727, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.5428392745465916, | |
| "grad_norm": 0.09688873589038849, | |
| "learning_rate": 1.8799683631555103e-06, | |
| "loss": 0.6137, | |
| "step": 651 | |
| }, | |
| { | |
| "epoch": 0.5436731290389827, | |
| "grad_norm": 0.09676310420036316, | |
| "learning_rate": 1.8795412543386337e-06, | |
| "loss": 0.612, | |
| "step": 652 | |
| }, | |
| { | |
| "epoch": 0.5445069835313737, | |
| "grad_norm": 0.09581848233938217, | |
| "learning_rate": 1.8791134356748962e-06, | |
| "loss": 0.6632, | |
| "step": 653 | |
| }, | |
| { | |
| "epoch": 0.5453408380237649, | |
| "grad_norm": 0.099350206553936, | |
| "learning_rate": 1.8786849075095753e-06, | |
| "loss": 0.6652, | |
| "step": 654 | |
| }, | |
| { | |
| "epoch": 0.5461746925161559, | |
| "grad_norm": 0.09312117099761963, | |
| "learning_rate": 1.878255670188521e-06, | |
| "loss": 0.6184, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 0.5470085470085471, | |
| "grad_norm": 0.09312383830547333, | |
| "learning_rate": 1.8778257240581554e-06, | |
| "loss": 0.6595, | |
| "step": 656 | |
| }, | |
| { | |
| "epoch": 0.5478424015009381, | |
| "grad_norm": 0.09485237300395966, | |
| "learning_rate": 1.8773950694654732e-06, | |
| "loss": 0.716, | |
| "step": 657 | |
| }, | |
| { | |
| "epoch": 0.5486762559933291, | |
| "grad_norm": 0.09585438668727875, | |
| "learning_rate": 1.8769637067580405e-06, | |
| "loss": 0.7058, | |
| "step": 658 | |
| }, | |
| { | |
| "epoch": 0.5495101104857203, | |
| "grad_norm": 0.09024535119533539, | |
| "learning_rate": 1.876531636283995e-06, | |
| "loss": 0.594, | |
| "step": 659 | |
| }, | |
| { | |
| "epoch": 0.5503439649781113, | |
| "grad_norm": 0.09285992383956909, | |
| "learning_rate": 1.8760988583920456e-06, | |
| "loss": 0.648, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.5511778194705024, | |
| "grad_norm": 0.08939266204833984, | |
| "learning_rate": 1.8756653734314722e-06, | |
| "loss": 0.5866, | |
| "step": 661 | |
| }, | |
| { | |
| "epoch": 0.5520116739628935, | |
| "grad_norm": 0.10212475061416626, | |
| "learning_rate": 1.8752311817521254e-06, | |
| "loss": 0.5485, | |
| "step": 662 | |
| }, | |
| { | |
| "epoch": 0.5528455284552846, | |
| "grad_norm": 0.0938698872923851, | |
| "learning_rate": 1.8747962837044256e-06, | |
| "loss": 0.692, | |
| "step": 663 | |
| }, | |
| { | |
| "epoch": 0.5536793829476756, | |
| "grad_norm": 0.09794861823320389, | |
| "learning_rate": 1.8743606796393644e-06, | |
| "loss": 0.6449, | |
| "step": 664 | |
| }, | |
| { | |
| "epoch": 0.5545132374400668, | |
| "grad_norm": 0.09574401378631592, | |
| "learning_rate": 1.8739243699085019e-06, | |
| "loss": 0.5878, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 0.5553470919324578, | |
| "grad_norm": 0.10169863700866699, | |
| "learning_rate": 1.8734873548639687e-06, | |
| "loss": 0.6968, | |
| "step": 666 | |
| }, | |
| { | |
| "epoch": 0.5561809464248488, | |
| "grad_norm": 0.09560302644968033, | |
| "learning_rate": 1.8730496348584642e-06, | |
| "loss": 0.6448, | |
| "step": 667 | |
| }, | |
| { | |
| "epoch": 0.55701480091724, | |
| "grad_norm": 0.10035623610019684, | |
| "learning_rate": 1.8726112102452572e-06, | |
| "loss": 0.6893, | |
| "step": 668 | |
| }, | |
| { | |
| "epoch": 0.557848655409631, | |
| "grad_norm": 0.093927301466465, | |
| "learning_rate": 1.872172081378184e-06, | |
| "loss": 0.6541, | |
| "step": 669 | |
| }, | |
| { | |
| "epoch": 0.5586825099020221, | |
| "grad_norm": 0.0979999378323555, | |
| "learning_rate": 1.8717322486116508e-06, | |
| "loss": 0.6651, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.5595163643944132, | |
| "grad_norm": 0.10020069032907486, | |
| "learning_rate": 1.8712917123006312e-06, | |
| "loss": 0.6289, | |
| "step": 671 | |
| }, | |
| { | |
| "epoch": 0.5603502188868043, | |
| "grad_norm": 0.10121606290340424, | |
| "learning_rate": 1.8708504728006664e-06, | |
| "loss": 0.6547, | |
| "step": 672 | |
| }, | |
| { | |
| "epoch": 0.5611840733791953, | |
| "grad_norm": 0.09832029789686203, | |
| "learning_rate": 1.870408530467865e-06, | |
| "loss": 0.651, | |
| "step": 673 | |
| }, | |
| { | |
| "epoch": 0.5620179278715864, | |
| "grad_norm": 0.0978146642446518, | |
| "learning_rate": 1.8699658856589037e-06, | |
| "loss": 0.6453, | |
| "step": 674 | |
| }, | |
| { | |
| "epoch": 0.5628517823639775, | |
| "grad_norm": 0.10008279979228973, | |
| "learning_rate": 1.8695225387310249e-06, | |
| "loss": 0.6945, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 0.5636856368563685, | |
| "grad_norm": 0.09760496020317078, | |
| "learning_rate": 1.8690784900420388e-06, | |
| "loss": 0.6222, | |
| "step": 676 | |
| }, | |
| { | |
| "epoch": 0.5645194913487597, | |
| "grad_norm": 0.09725689142942429, | |
| "learning_rate": 1.8686337399503215e-06, | |
| "loss": 0.6464, | |
| "step": 677 | |
| }, | |
| { | |
| "epoch": 0.5653533458411507, | |
| "grad_norm": 0.09347565472126007, | |
| "learning_rate": 1.8681882888148152e-06, | |
| "loss": 0.6282, | |
| "step": 678 | |
| }, | |
| { | |
| "epoch": 0.5661872003335418, | |
| "grad_norm": 0.09704628586769104, | |
| "learning_rate": 1.867742136995028e-06, | |
| "loss": 0.6529, | |
| "step": 679 | |
| }, | |
| { | |
| "epoch": 0.5670210548259329, | |
| "grad_norm": 0.09918136149644852, | |
| "learning_rate": 1.8672952848510328e-06, | |
| "loss": 0.739, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.5678549093183239, | |
| "grad_norm": 0.09772541373968124, | |
| "learning_rate": 1.8668477327434686e-06, | |
| "loss": 0.6383, | |
| "step": 681 | |
| }, | |
| { | |
| "epoch": 0.568688763810715, | |
| "grad_norm": 0.09866555780172348, | |
| "learning_rate": 1.866399481033539e-06, | |
| "loss": 0.6721, | |
| "step": 682 | |
| }, | |
| { | |
| "epoch": 0.5695226183031061, | |
| "grad_norm": 0.09793423861265182, | |
| "learning_rate": 1.8659505300830123e-06, | |
| "loss": 0.6055, | |
| "step": 683 | |
| }, | |
| { | |
| "epoch": 0.5703564727954972, | |
| "grad_norm": 0.10061061382293701, | |
| "learning_rate": 1.865500880254221e-06, | |
| "loss": 0.6544, | |
| "step": 684 | |
| }, | |
| { | |
| "epoch": 0.5711903272878882, | |
| "grad_norm": 0.09208130091428757, | |
| "learning_rate": 1.8650505319100617e-06, | |
| "loss": 0.6014, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 0.5720241817802794, | |
| "grad_norm": 0.0995427742600441, | |
| "learning_rate": 1.8645994854139948e-06, | |
| "loss": 0.6965, | |
| "step": 686 | |
| }, | |
| { | |
| "epoch": 0.5728580362726704, | |
| "grad_norm": 0.09244947880506516, | |
| "learning_rate": 1.8641477411300438e-06, | |
| "loss": 0.6272, | |
| "step": 687 | |
| }, | |
| { | |
| "epoch": 0.5736918907650616, | |
| "grad_norm": 0.09937774389982224, | |
| "learning_rate": 1.863695299422796e-06, | |
| "loss": 0.6648, | |
| "step": 688 | |
| }, | |
| { | |
| "epoch": 0.5745257452574526, | |
| "grad_norm": 0.09111293405294418, | |
| "learning_rate": 1.8632421606574007e-06, | |
| "loss": 0.5476, | |
| "step": 689 | |
| }, | |
| { | |
| "epoch": 0.5753595997498436, | |
| "grad_norm": 0.10068730264902115, | |
| "learning_rate": 1.8627883251995708e-06, | |
| "loss": 0.6535, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.5761934542422348, | |
| "grad_norm": 0.09240923076868057, | |
| "learning_rate": 1.862333793415581e-06, | |
| "loss": 0.575, | |
| "step": 691 | |
| }, | |
| { | |
| "epoch": 0.5770273087346258, | |
| "grad_norm": 0.09765233844518661, | |
| "learning_rate": 1.8618785656722676e-06, | |
| "loss": 0.6323, | |
| "step": 692 | |
| }, | |
| { | |
| "epoch": 0.5778611632270169, | |
| "grad_norm": 0.09996625781059265, | |
| "learning_rate": 1.861422642337029e-06, | |
| "loss": 0.639, | |
| "step": 693 | |
| }, | |
| { | |
| "epoch": 0.578695017719408, | |
| "grad_norm": 0.10148533433675766, | |
| "learning_rate": 1.860966023777825e-06, | |
| "loss": 0.7196, | |
| "step": 694 | |
| }, | |
| { | |
| "epoch": 0.579528872211799, | |
| "grad_norm": 0.09370430558919907, | |
| "learning_rate": 1.8605087103631764e-06, | |
| "loss": 0.6385, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 0.5803627267041901, | |
| "grad_norm": 0.09228570014238358, | |
| "learning_rate": 1.8600507024621648e-06, | |
| "loss": 0.6242, | |
| "step": 696 | |
| }, | |
| { | |
| "epoch": 0.5811965811965812, | |
| "grad_norm": 0.09689746052026749, | |
| "learning_rate": 1.8595920004444324e-06, | |
| "loss": 0.6135, | |
| "step": 697 | |
| }, | |
| { | |
| "epoch": 0.5820304356889723, | |
| "grad_norm": 0.09802035987377167, | |
| "learning_rate": 1.8591326046801812e-06, | |
| "loss": 0.5789, | |
| "step": 698 | |
| }, | |
| { | |
| "epoch": 0.5828642901813633, | |
| "grad_norm": 0.10172217339277267, | |
| "learning_rate": 1.8586725155401735e-06, | |
| "loss": 0.5921, | |
| "step": 699 | |
| }, | |
| { | |
| "epoch": 0.5836981446737545, | |
| "grad_norm": 0.10300835222005844, | |
| "learning_rate": 1.858211733395731e-06, | |
| "loss": 0.6647, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.5845319991661455, | |
| "grad_norm": 0.09017598628997803, | |
| "learning_rate": 1.8577502586187353e-06, | |
| "loss": 0.6409, | |
| "step": 701 | |
| }, | |
| { | |
| "epoch": 0.5853658536585366, | |
| "grad_norm": 0.10058888792991638, | |
| "learning_rate": 1.8572880915816259e-06, | |
| "loss": 0.5902, | |
| "step": 702 | |
| }, | |
| { | |
| "epoch": 0.5861997081509277, | |
| "grad_norm": 0.10125189274549484, | |
| "learning_rate": 1.856825232657402e-06, | |
| "loss": 0.6392, | |
| "step": 703 | |
| }, | |
| { | |
| "epoch": 0.5870335626433187, | |
| "grad_norm": 0.0983993262052536, | |
| "learning_rate": 1.8563616822196204e-06, | |
| "loss": 0.5915, | |
| "step": 704 | |
| }, | |
| { | |
| "epoch": 0.5878674171357098, | |
| "grad_norm": 0.09559126943349838, | |
| "learning_rate": 1.8558974406423965e-06, | |
| "loss": 0.6386, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 0.5887012716281009, | |
| "grad_norm": 0.09088745713233948, | |
| "learning_rate": 1.8554325083004034e-06, | |
| "loss": 0.6483, | |
| "step": 706 | |
| }, | |
| { | |
| "epoch": 0.589535126120492, | |
| "grad_norm": 0.09516370296478271, | |
| "learning_rate": 1.854966885568872e-06, | |
| "loss": 0.6556, | |
| "step": 707 | |
| }, | |
| { | |
| "epoch": 0.590368980612883, | |
| "grad_norm": 0.09569456428289413, | |
| "learning_rate": 1.8545005728235896e-06, | |
| "loss": 0.6394, | |
| "step": 708 | |
| }, | |
| { | |
| "epoch": 0.5912028351052742, | |
| "grad_norm": 0.0980261042714119, | |
| "learning_rate": 1.854033570440901e-06, | |
| "loss": 0.6415, | |
| "step": 709 | |
| }, | |
| { | |
| "epoch": 0.5920366895976652, | |
| "grad_norm": 0.09796703606843948, | |
| "learning_rate": 1.8535658787977075e-06, | |
| "loss": 0.6736, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.5928705440900562, | |
| "grad_norm": 0.09802737832069397, | |
| "learning_rate": 1.8530974982714664e-06, | |
| "loss": 0.5802, | |
| "step": 711 | |
| }, | |
| { | |
| "epoch": 0.5937043985824474, | |
| "grad_norm": 0.09888147562742233, | |
| "learning_rate": 1.8526284292401914e-06, | |
| "loss": 0.6648, | |
| "step": 712 | |
| }, | |
| { | |
| "epoch": 0.5945382530748384, | |
| "grad_norm": 0.10190586000680923, | |
| "learning_rate": 1.8521586720824515e-06, | |
| "loss": 0.7032, | |
| "step": 713 | |
| }, | |
| { | |
| "epoch": 0.5953721075672295, | |
| "grad_norm": 0.09422854334115982, | |
| "learning_rate": 1.8516882271773712e-06, | |
| "loss": 0.6551, | |
| "step": 714 | |
| }, | |
| { | |
| "epoch": 0.5962059620596206, | |
| "grad_norm": 0.09025522321462631, | |
| "learning_rate": 1.8512170949046305e-06, | |
| "loss": 0.6231, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 0.5970398165520117, | |
| "grad_norm": 0.102678582072258, | |
| "learning_rate": 1.850745275644463e-06, | |
| "loss": 0.6924, | |
| "step": 716 | |
| }, | |
| { | |
| "epoch": 0.5978736710444027, | |
| "grad_norm": 0.09951562434434891, | |
| "learning_rate": 1.8502727697776579e-06, | |
| "loss": 0.5763, | |
| "step": 717 | |
| }, | |
| { | |
| "epoch": 0.5987075255367938, | |
| "grad_norm": 0.10048934072256088, | |
| "learning_rate": 1.8497995776855579e-06, | |
| "loss": 0.6369, | |
| "step": 718 | |
| }, | |
| { | |
| "epoch": 0.5995413800291849, | |
| "grad_norm": 0.11139318346977234, | |
| "learning_rate": 1.84932569975006e-06, | |
| "loss": 0.6086, | |
| "step": 719 | |
| }, | |
| { | |
| "epoch": 0.600375234521576, | |
| "grad_norm": 0.10424143821001053, | |
| "learning_rate": 1.8488511363536138e-06, | |
| "loss": 0.6394, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.6012090890139671, | |
| "grad_norm": 0.09213969111442566, | |
| "learning_rate": 1.8483758878792236e-06, | |
| "loss": 0.6554, | |
| "step": 721 | |
| }, | |
| { | |
| "epoch": 0.6020429435063581, | |
| "grad_norm": 0.0998111367225647, | |
| "learning_rate": 1.8478999547104452e-06, | |
| "loss": 0.6717, | |
| "step": 722 | |
| }, | |
| { | |
| "epoch": 0.6028767979987493, | |
| "grad_norm": 0.09211792796850204, | |
| "learning_rate": 1.8474233372313877e-06, | |
| "loss": 0.5694, | |
| "step": 723 | |
| }, | |
| { | |
| "epoch": 0.6037106524911403, | |
| "grad_norm": 0.10054522007703781, | |
| "learning_rate": 1.8469460358267127e-06, | |
| "loss": 0.6501, | |
| "step": 724 | |
| }, | |
| { | |
| "epoch": 0.6045445069835313, | |
| "grad_norm": 0.09979543089866638, | |
| "learning_rate": 1.8464680508816323e-06, | |
| "loss": 0.6056, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 0.6053783614759225, | |
| "grad_norm": 0.1000729575753212, | |
| "learning_rate": 1.8459893827819126e-06, | |
| "loss": 0.6302, | |
| "step": 726 | |
| }, | |
| { | |
| "epoch": 0.6062122159683135, | |
| "grad_norm": 0.09924089163541794, | |
| "learning_rate": 1.8455100319138692e-06, | |
| "loss": 0.6432, | |
| "step": 727 | |
| }, | |
| { | |
| "epoch": 0.6070460704607046, | |
| "grad_norm": 0.0920461118221283, | |
| "learning_rate": 1.8450299986643695e-06, | |
| "loss": 0.6081, | |
| "step": 728 | |
| }, | |
| { | |
| "epoch": 0.6078799249530957, | |
| "grad_norm": 0.10430511087179184, | |
| "learning_rate": 1.8445492834208316e-06, | |
| "loss": 0.6586, | |
| "step": 729 | |
| }, | |
| { | |
| "epoch": 0.6087137794454868, | |
| "grad_norm": 0.09923563152551651, | |
| "learning_rate": 1.8440678865712236e-06, | |
| "loss": 0.6166, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.6095476339378778, | |
| "grad_norm": 0.1012277603149414, | |
| "learning_rate": 1.843585808504064e-06, | |
| "loss": 0.6843, | |
| "step": 731 | |
| }, | |
| { | |
| "epoch": 0.610381488430269, | |
| "grad_norm": 0.09866297990083694, | |
| "learning_rate": 1.8431030496084214e-06, | |
| "loss": 0.6556, | |
| "step": 732 | |
| }, | |
| { | |
| "epoch": 0.61121534292266, | |
| "grad_norm": 0.09874875843524933, | |
| "learning_rate": 1.8426196102739132e-06, | |
| "loss": 0.5993, | |
| "step": 733 | |
| }, | |
| { | |
| "epoch": 0.612049197415051, | |
| "grad_norm": 0.09275837987661362, | |
| "learning_rate": 1.8421354908907067e-06, | |
| "loss": 0.6326, | |
| "step": 734 | |
| }, | |
| { | |
| "epoch": 0.6128830519074422, | |
| "grad_norm": 0.09502032399177551, | |
| "learning_rate": 1.8416506918495173e-06, | |
| "loss": 0.6077, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 0.6137169063998332, | |
| "grad_norm": 0.09185368567705154, | |
| "learning_rate": 1.8411652135416093e-06, | |
| "loss": 0.6064, | |
| "step": 736 | |
| }, | |
| { | |
| "epoch": 0.6145507608922243, | |
| "grad_norm": 0.09815992414951324, | |
| "learning_rate": 1.8406790563587956e-06, | |
| "loss": 0.6313, | |
| "step": 737 | |
| }, | |
| { | |
| "epoch": 0.6153846153846154, | |
| "grad_norm": 0.09629546850919724, | |
| "learning_rate": 1.8401922206934361e-06, | |
| "loss": 0.6106, | |
| "step": 738 | |
| }, | |
| { | |
| "epoch": 0.6162184698770065, | |
| "grad_norm": 0.09635171294212341, | |
| "learning_rate": 1.839704706938439e-06, | |
| "loss": 0.6523, | |
| "step": 739 | |
| }, | |
| { | |
| "epoch": 0.6170523243693975, | |
| "grad_norm": 0.10179682075977325, | |
| "learning_rate": 1.8392165154872595e-06, | |
| "loss": 0.5873, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.6178861788617886, | |
| "grad_norm": 0.10270754247903824, | |
| "learning_rate": 1.8387276467338996e-06, | |
| "loss": 0.673, | |
| "step": 741 | |
| }, | |
| { | |
| "epoch": 0.6187200333541797, | |
| "grad_norm": 0.09746810793876648, | |
| "learning_rate": 1.8382381010729086e-06, | |
| "loss": 0.6564, | |
| "step": 742 | |
| }, | |
| { | |
| "epoch": 0.6195538878465707, | |
| "grad_norm": 0.0962262824177742, | |
| "learning_rate": 1.8377478788993813e-06, | |
| "loss": 0.5738, | |
| "step": 743 | |
| }, | |
| { | |
| "epoch": 0.6203877423389619, | |
| "grad_norm": 0.09518938511610031, | |
| "learning_rate": 1.8372569806089586e-06, | |
| "loss": 0.6345, | |
| "step": 744 | |
| }, | |
| { | |
| "epoch": 0.6212215968313529, | |
| "grad_norm": 0.09188467264175415, | |
| "learning_rate": 1.8367654065978276e-06, | |
| "loss": 0.6112, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 0.622055451323744, | |
| "grad_norm": 0.09746599197387695, | |
| "learning_rate": 1.8362731572627202e-06, | |
| "loss": 0.6349, | |
| "step": 746 | |
| }, | |
| { | |
| "epoch": 0.6228893058161351, | |
| "grad_norm": 0.10009641200304031, | |
| "learning_rate": 1.8357802330009136e-06, | |
| "loss": 0.6502, | |
| "step": 747 | |
| }, | |
| { | |
| "epoch": 0.6237231603085261, | |
| "grad_norm": 0.10499203950166702, | |
| "learning_rate": 1.8352866342102296e-06, | |
| "loss": 0.6812, | |
| "step": 748 | |
| }, | |
| { | |
| "epoch": 0.6245570148009172, | |
| "grad_norm": 0.09757328033447266, | |
| "learning_rate": 1.8347923612890349e-06, | |
| "loss": 0.6025, | |
| "step": 749 | |
| }, | |
| { | |
| "epoch": 0.6253908692933083, | |
| "grad_norm": 0.10616346448659897, | |
| "learning_rate": 1.8342974146362394e-06, | |
| "loss": 0.6638, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.6262247237856994, | |
| "grad_norm": 0.10131774097681046, | |
| "learning_rate": 1.8338017946512968e-06, | |
| "loss": 0.5456, | |
| "step": 751 | |
| }, | |
| { | |
| "epoch": 0.6270585782780905, | |
| "grad_norm": 0.10244927555322647, | |
| "learning_rate": 1.8333055017342054e-06, | |
| "loss": 0.6751, | |
| "step": 752 | |
| }, | |
| { | |
| "epoch": 0.6278924327704816, | |
| "grad_norm": 0.10338141769170761, | |
| "learning_rate": 1.8328085362855055e-06, | |
| "loss": 0.6647, | |
| "step": 753 | |
| }, | |
| { | |
| "epoch": 0.6287262872628726, | |
| "grad_norm": 0.10026909410953522, | |
| "learning_rate": 1.8323108987062802e-06, | |
| "loss": 0.6543, | |
| "step": 754 | |
| }, | |
| { | |
| "epoch": 0.6295601417552638, | |
| "grad_norm": 0.10641954839229584, | |
| "learning_rate": 1.8318125893981553e-06, | |
| "loss": 0.6425, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 0.6303939962476548, | |
| "grad_norm": 0.09939148277044296, | |
| "learning_rate": 1.8313136087632995e-06, | |
| "loss": 0.6531, | |
| "step": 756 | |
| }, | |
| { | |
| "epoch": 0.6312278507400458, | |
| "grad_norm": 0.10437260568141937, | |
| "learning_rate": 1.8308139572044215e-06, | |
| "loss": 0.6783, | |
| "step": 757 | |
| }, | |
| { | |
| "epoch": 0.632061705232437, | |
| "grad_norm": 0.09760341793298721, | |
| "learning_rate": 1.830313635124773e-06, | |
| "loss": 0.6734, | |
| "step": 758 | |
| }, | |
| { | |
| "epoch": 0.632895559724828, | |
| "grad_norm": 0.1017068475484848, | |
| "learning_rate": 1.8298126429281467e-06, | |
| "loss": 0.6171, | |
| "step": 759 | |
| }, | |
| { | |
| "epoch": 0.6337294142172191, | |
| "grad_norm": 0.09881062060594559, | |
| "learning_rate": 1.8293109810188754e-06, | |
| "loss": 0.6309, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.6345632687096102, | |
| "grad_norm": 0.11102797091007233, | |
| "learning_rate": 1.8288086498018326e-06, | |
| "loss": 0.6361, | |
| "step": 761 | |
| }, | |
| { | |
| "epoch": 0.6353971232020013, | |
| "grad_norm": 0.09473263472318649, | |
| "learning_rate": 1.828305649682433e-06, | |
| "loss": 0.5787, | |
| "step": 762 | |
| }, | |
| { | |
| "epoch": 0.6362309776943923, | |
| "grad_norm": 0.10742990672588348, | |
| "learning_rate": 1.8278019810666293e-06, | |
| "loss": 0.6638, | |
| "step": 763 | |
| }, | |
| { | |
| "epoch": 0.6370648321867834, | |
| "grad_norm": 0.09987856447696686, | |
| "learning_rate": 1.8272976443609156e-06, | |
| "loss": 0.6054, | |
| "step": 764 | |
| }, | |
| { | |
| "epoch": 0.6378986866791745, | |
| "grad_norm": 0.09839458018541336, | |
| "learning_rate": 1.8267926399723238e-06, | |
| "loss": 0.6258, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 0.6387325411715655, | |
| "grad_norm": 0.09883236885070801, | |
| "learning_rate": 1.8262869683084258e-06, | |
| "loss": 0.6695, | |
| "step": 766 | |
| }, | |
| { | |
| "epoch": 0.6395663956639567, | |
| "grad_norm": 0.10263945162296295, | |
| "learning_rate": 1.825780629777331e-06, | |
| "loss": 0.6032, | |
| "step": 767 | |
| }, | |
| { | |
| "epoch": 0.6404002501563477, | |
| "grad_norm": 0.09623769670724869, | |
| "learning_rate": 1.8252736247876874e-06, | |
| "loss": 0.6794, | |
| "step": 768 | |
| }, | |
| { | |
| "epoch": 0.6412341046487388, | |
| "grad_norm": 0.10218116641044617, | |
| "learning_rate": 1.8247659537486811e-06, | |
| "loss": 0.5969, | |
| "step": 769 | |
| }, | |
| { | |
| "epoch": 0.6420679591411299, | |
| "grad_norm": 0.09405802190303802, | |
| "learning_rate": 1.8242576170700352e-06, | |
| "loss": 0.6493, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.642901813633521, | |
| "grad_norm": 0.09709781408309937, | |
| "learning_rate": 1.8237486151620112e-06, | |
| "loss": 0.7054, | |
| "step": 771 | |
| }, | |
| { | |
| "epoch": 0.643735668125912, | |
| "grad_norm": 0.09584866464138031, | |
| "learning_rate": 1.823238948435406e-06, | |
| "loss": 0.6299, | |
| "step": 772 | |
| }, | |
| { | |
| "epoch": 0.6445695226183031, | |
| "grad_norm": 0.09731820970773697, | |
| "learning_rate": 1.8227286173015538e-06, | |
| "loss": 0.6633, | |
| "step": 773 | |
| }, | |
| { | |
| "epoch": 0.6454033771106942, | |
| "grad_norm": 0.09664048999547958, | |
| "learning_rate": 1.822217622172325e-06, | |
| "loss": 0.6075, | |
| "step": 774 | |
| }, | |
| { | |
| "epoch": 0.6462372316030852, | |
| "grad_norm": 0.0990557000041008, | |
| "learning_rate": 1.8217059634601259e-06, | |
| "loss": 0.6755, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 0.6470710860954764, | |
| "grad_norm": 0.10071469098329544, | |
| "learning_rate": 1.8211936415778983e-06, | |
| "loss": 0.6197, | |
| "step": 776 | |
| }, | |
| { | |
| "epoch": 0.6479049405878674, | |
| "grad_norm": 0.10386556386947632, | |
| "learning_rate": 1.820680656939119e-06, | |
| "loss": 0.6195, | |
| "step": 777 | |
| }, | |
| { | |
| "epoch": 0.6487387950802584, | |
| "grad_norm": 0.09966234862804413, | |
| "learning_rate": 1.8201670099578001e-06, | |
| "loss": 0.6654, | |
| "step": 778 | |
| }, | |
| { | |
| "epoch": 0.6495726495726496, | |
| "grad_norm": 0.09882418066263199, | |
| "learning_rate": 1.819652701048488e-06, | |
| "loss": 0.6525, | |
| "step": 779 | |
| }, | |
| { | |
| "epoch": 0.6504065040650406, | |
| "grad_norm": 0.09504549950361252, | |
| "learning_rate": 1.8191377306262633e-06, | |
| "loss": 0.6326, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.6512403585574317, | |
| "grad_norm": 0.09834848344326019, | |
| "learning_rate": 1.8186220991067404e-06, | |
| "loss": 0.6417, | |
| "step": 781 | |
| }, | |
| { | |
| "epoch": 0.6520742130498228, | |
| "grad_norm": 0.09583116322755814, | |
| "learning_rate": 1.8181058069060676e-06, | |
| "loss": 0.6644, | |
| "step": 782 | |
| }, | |
| { | |
| "epoch": 0.6529080675422139, | |
| "grad_norm": 0.09287475794553757, | |
| "learning_rate": 1.8175888544409264e-06, | |
| "loss": 0.6385, | |
| "step": 783 | |
| }, | |
| { | |
| "epoch": 0.653741922034605, | |
| "grad_norm": 0.10364864766597748, | |
| "learning_rate": 1.8170712421285305e-06, | |
| "loss": 0.6996, | |
| "step": 784 | |
| }, | |
| { | |
| "epoch": 0.6545757765269961, | |
| "grad_norm": 0.09863609820604324, | |
| "learning_rate": 1.816552970386627e-06, | |
| "loss": 0.6272, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 0.6554096310193871, | |
| "grad_norm": 0.10557069629430771, | |
| "learning_rate": 1.8160340396334949e-06, | |
| "loss": 0.6318, | |
| "step": 786 | |
| }, | |
| { | |
| "epoch": 0.6562434855117782, | |
| "grad_norm": 0.10506842285394669, | |
| "learning_rate": 1.8155144502879443e-06, | |
| "loss": 0.6557, | |
| "step": 787 | |
| }, | |
| { | |
| "epoch": 0.6570773400041693, | |
| "grad_norm": 0.09928935766220093, | |
| "learning_rate": 1.8149942027693182e-06, | |
| "loss": 0.6832, | |
| "step": 788 | |
| }, | |
| { | |
| "epoch": 0.6579111944965603, | |
| "grad_norm": 0.09821849316358566, | |
| "learning_rate": 1.8144732974974902e-06, | |
| "loss": 0.6342, | |
| "step": 789 | |
| }, | |
| { | |
| "epoch": 0.6587450489889515, | |
| "grad_norm": 0.09966670721769333, | |
| "learning_rate": 1.8139517348928637e-06, | |
| "loss": 0.6321, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.6595789034813425, | |
| "grad_norm": 0.10439924150705338, | |
| "learning_rate": 1.8134295153763745e-06, | |
| "loss": 0.6552, | |
| "step": 791 | |
| }, | |
| { | |
| "epoch": 0.6604127579737336, | |
| "grad_norm": 0.09848225861787796, | |
| "learning_rate": 1.8129066393694877e-06, | |
| "loss": 0.593, | |
| "step": 792 | |
| }, | |
| { | |
| "epoch": 0.6612466124661247, | |
| "grad_norm": 0.09793344885110855, | |
| "learning_rate": 1.8123831072941974e-06, | |
| "loss": 0.6406, | |
| "step": 793 | |
| }, | |
| { | |
| "epoch": 0.6620804669585157, | |
| "grad_norm": 0.0945219025015831, | |
| "learning_rate": 1.8118589195730285e-06, | |
| "loss": 0.6118, | |
| "step": 794 | |
| }, | |
| { | |
| "epoch": 0.6629143214509068, | |
| "grad_norm": 0.10230748355388641, | |
| "learning_rate": 1.8113340766290343e-06, | |
| "loss": 0.65, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 0.6637481759432979, | |
| "grad_norm": 0.10141695290803909, | |
| "learning_rate": 1.8108085788857973e-06, | |
| "loss": 0.6109, | |
| "step": 796 | |
| }, | |
| { | |
| "epoch": 0.664582030435689, | |
| "grad_norm": 0.11011820286512375, | |
| "learning_rate": 1.8102824267674282e-06, | |
| "loss": 0.6531, | |
| "step": 797 | |
| }, | |
| { | |
| "epoch": 0.66541588492808, | |
| "grad_norm": 0.09687966108322144, | |
| "learning_rate": 1.8097556206985658e-06, | |
| "loss": 0.649, | |
| "step": 798 | |
| }, | |
| { | |
| "epoch": 0.6662497394204712, | |
| "grad_norm": 0.09653454273939133, | |
| "learning_rate": 1.8092281611043764e-06, | |
| "loss": 0.6487, | |
| "step": 799 | |
| }, | |
| { | |
| "epoch": 0.6670835939128622, | |
| "grad_norm": 0.10373052954673767, | |
| "learning_rate": 1.808700048410555e-06, | |
| "loss": 0.6221, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.6670835939128622, | |
| "eval_loss": 0.6415942907333374, | |
| "eval_runtime": 321.6008, | |
| "eval_samples_per_second": 16.738, | |
| "eval_steps_per_second": 2.792, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.6679174484052532, | |
| "grad_norm": 0.10485277324914932, | |
| "learning_rate": 1.8081712830433224e-06, | |
| "loss": 0.6176, | |
| "step": 801 | |
| }, | |
| { | |
| "epoch": 0.6687513028976444, | |
| "grad_norm": 0.10105489194393158, | |
| "learning_rate": 1.8076418654294266e-06, | |
| "loss": 0.6708, | |
| "step": 802 | |
| }, | |
| { | |
| "epoch": 0.6695851573900354, | |
| "grad_norm": 0.1096423864364624, | |
| "learning_rate": 1.8071117959961416e-06, | |
| "loss": 0.601, | |
| "step": 803 | |
| }, | |
| { | |
| "epoch": 0.6704190118824265, | |
| "grad_norm": 0.10762202739715576, | |
| "learning_rate": 1.8065810751712682e-06, | |
| "loss": 0.6567, | |
| "step": 804 | |
| }, | |
| { | |
| "epoch": 0.6712528663748176, | |
| "grad_norm": 0.1107725277543068, | |
| "learning_rate": 1.8060497033831324e-06, | |
| "loss": 0.5889, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 0.6720867208672087, | |
| "grad_norm": 0.10468114167451859, | |
| "learning_rate": 1.8055176810605857e-06, | |
| "loss": 0.5847, | |
| "step": 806 | |
| }, | |
| { | |
| "epoch": 0.6729205753595997, | |
| "grad_norm": 0.11013887077569962, | |
| "learning_rate": 1.8049850086330048e-06, | |
| "loss": 0.6008, | |
| "step": 807 | |
| }, | |
| { | |
| "epoch": 0.6737544298519909, | |
| "grad_norm": 0.1043848916888237, | |
| "learning_rate": 1.8044516865302908e-06, | |
| "loss": 0.6368, | |
| "step": 808 | |
| }, | |
| { | |
| "epoch": 0.6745882843443819, | |
| "grad_norm": 0.09962385892868042, | |
| "learning_rate": 1.8039177151828691e-06, | |
| "loss": 0.5842, | |
| "step": 809 | |
| }, | |
| { | |
| "epoch": 0.6754221388367729, | |
| "grad_norm": 0.1098175197839737, | |
| "learning_rate": 1.8033830950216897e-06, | |
| "loss": 0.6397, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.6762559933291641, | |
| "grad_norm": 0.10420270264148712, | |
| "learning_rate": 1.8028478264782252e-06, | |
| "loss": 0.6131, | |
| "step": 811 | |
| }, | |
| { | |
| "epoch": 0.6770898478215551, | |
| "grad_norm": 0.10396002978086472, | |
| "learning_rate": 1.8023119099844723e-06, | |
| "loss": 0.6499, | |
| "step": 812 | |
| }, | |
| { | |
| "epoch": 0.6779237023139462, | |
| "grad_norm": 0.09836006909608841, | |
| "learning_rate": 1.8017753459729504e-06, | |
| "loss": 0.5962, | |
| "step": 813 | |
| }, | |
| { | |
| "epoch": 0.6787575568063373, | |
| "grad_norm": 0.11005699634552002, | |
| "learning_rate": 1.8012381348767014e-06, | |
| "loss": 0.6202, | |
| "step": 814 | |
| }, | |
| { | |
| "epoch": 0.6795914112987284, | |
| "grad_norm": 0.10937905311584473, | |
| "learning_rate": 1.8007002771292896e-06, | |
| "loss": 0.6053, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 0.6804252657911195, | |
| "grad_norm": 0.10817151516675949, | |
| "learning_rate": 1.800161773164801e-06, | |
| "loss": 0.6713, | |
| "step": 816 | |
| }, | |
| { | |
| "epoch": 0.6812591202835105, | |
| "grad_norm": 0.10142417997121811, | |
| "learning_rate": 1.7996226234178434e-06, | |
| "loss": 0.6166, | |
| "step": 817 | |
| }, | |
| { | |
| "epoch": 0.6820929747759016, | |
| "grad_norm": 0.10415297746658325, | |
| "learning_rate": 1.799082828323545e-06, | |
| "loss": 0.6912, | |
| "step": 818 | |
| }, | |
| { | |
| "epoch": 0.6829268292682927, | |
| "grad_norm": 0.09985463321208954, | |
| "learning_rate": 1.7985423883175565e-06, | |
| "loss": 0.6694, | |
| "step": 819 | |
| }, | |
| { | |
| "epoch": 0.6837606837606838, | |
| "grad_norm": 0.1025400459766388, | |
| "learning_rate": 1.7980013038360475e-06, | |
| "loss": 0.6464, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.6845945382530748, | |
| "grad_norm": 0.09324987977743149, | |
| "learning_rate": 1.7974595753157082e-06, | |
| "loss": 0.6251, | |
| "step": 821 | |
| }, | |
| { | |
| "epoch": 0.685428392745466, | |
| "grad_norm": 0.1065581887960434, | |
| "learning_rate": 1.796917203193749e-06, | |
| "loss": 0.6427, | |
| "step": 822 | |
| }, | |
| { | |
| "epoch": 0.686262247237857, | |
| "grad_norm": 0.1113988533616066, | |
| "learning_rate": 1.7963741879078985e-06, | |
| "loss": 0.6652, | |
| "step": 823 | |
| }, | |
| { | |
| "epoch": 0.687096101730248, | |
| "grad_norm": 0.09690185636281967, | |
| "learning_rate": 1.7958305298964066e-06, | |
| "loss": 0.6294, | |
| "step": 824 | |
| }, | |
| { | |
| "epoch": 0.6879299562226392, | |
| "grad_norm": 0.10385391116142273, | |
| "learning_rate": 1.7952862295980397e-06, | |
| "loss": 0.611, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 0.6887638107150302, | |
| "grad_norm": 0.10155371576547623, | |
| "learning_rate": 1.794741287452084e-06, | |
| "loss": 0.6268, | |
| "step": 826 | |
| }, | |
| { | |
| "epoch": 0.6895976652074213, | |
| "grad_norm": 0.10731323808431625, | |
| "learning_rate": 1.7941957038983425e-06, | |
| "loss": 0.6294, | |
| "step": 827 | |
| }, | |
| { | |
| "epoch": 0.6904315196998124, | |
| "grad_norm": 0.0989241898059845, | |
| "learning_rate": 1.7936494793771369e-06, | |
| "loss": 0.6739, | |
| "step": 828 | |
| }, | |
| { | |
| "epoch": 0.6912653741922035, | |
| "grad_norm": 0.10223259776830673, | |
| "learning_rate": 1.7931026143293059e-06, | |
| "loss": 0.5857, | |
| "step": 829 | |
| }, | |
| { | |
| "epoch": 0.6920992286845945, | |
| "grad_norm": 0.10547658056020737, | |
| "learning_rate": 1.792555109196205e-06, | |
| "loss": 0.6864, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.6929330831769857, | |
| "grad_norm": 0.1027214452624321, | |
| "learning_rate": 1.7920069644197063e-06, | |
| "loss": 0.5857, | |
| "step": 831 | |
| }, | |
| { | |
| "epoch": 0.6937669376693767, | |
| "grad_norm": 0.10017300397157669, | |
| "learning_rate": 1.7914581804421983e-06, | |
| "loss": 0.6466, | |
| "step": 832 | |
| }, | |
| { | |
| "epoch": 0.6946007921617677, | |
| "grad_norm": 0.09962292015552521, | |
| "learning_rate": 1.7909087577065851e-06, | |
| "loss": 0.5917, | |
| "step": 833 | |
| }, | |
| { | |
| "epoch": 0.6954346466541589, | |
| "grad_norm": 0.10357961803674698, | |
| "learning_rate": 1.7903586966562868e-06, | |
| "loss": 0.6382, | |
| "step": 834 | |
| }, | |
| { | |
| "epoch": 0.6962685011465499, | |
| "grad_norm": 0.1033301129937172, | |
| "learning_rate": 1.789807997735238e-06, | |
| "loss": 0.624, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 0.697102355638941, | |
| "grad_norm": 0.09935612231492996, | |
| "learning_rate": 1.789256661387889e-06, | |
| "loss": 0.6475, | |
| "step": 836 | |
| }, | |
| { | |
| "epoch": 0.6979362101313321, | |
| "grad_norm": 0.09738706797361374, | |
| "learning_rate": 1.7887046880592033e-06, | |
| "loss": 0.652, | |
| "step": 837 | |
| }, | |
| { | |
| "epoch": 0.6987700646237232, | |
| "grad_norm": 0.10389488190412521, | |
| "learning_rate": 1.7881520781946597e-06, | |
| "loss": 0.6755, | |
| "step": 838 | |
| }, | |
| { | |
| "epoch": 0.6996039191161142, | |
| "grad_norm": 0.09787558019161224, | |
| "learning_rate": 1.7875988322402501e-06, | |
| "loss": 0.618, | |
| "step": 839 | |
| }, | |
| { | |
| "epoch": 0.7004377736085053, | |
| "grad_norm": 0.09852594137191772, | |
| "learning_rate": 1.78704495064248e-06, | |
| "loss": 0.618, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.7012716281008964, | |
| "grad_norm": 0.09819158911705017, | |
| "learning_rate": 1.7864904338483675e-06, | |
| "loss": 0.5994, | |
| "step": 841 | |
| }, | |
| { | |
| "epoch": 0.7021054825932874, | |
| "grad_norm": 0.10942944139242172, | |
| "learning_rate": 1.7859352823054437e-06, | |
| "loss": 0.6405, | |
| "step": 842 | |
| }, | |
| { | |
| "epoch": 0.7029393370856786, | |
| "grad_norm": 0.10699477791786194, | |
| "learning_rate": 1.7853794964617522e-06, | |
| "loss": 0.6444, | |
| "step": 843 | |
| }, | |
| { | |
| "epoch": 0.7037731915780696, | |
| "grad_norm": 0.10151588916778564, | |
| "learning_rate": 1.784823076765848e-06, | |
| "loss": 0.6233, | |
| "step": 844 | |
| }, | |
| { | |
| "epoch": 0.7046070460704607, | |
| "grad_norm": 0.10147272795438766, | |
| "learning_rate": 1.7842660236667974e-06, | |
| "loss": 0.6606, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 0.7054409005628518, | |
| "grad_norm": 0.1031784638762474, | |
| "learning_rate": 1.7837083376141796e-06, | |
| "loss": 0.6332, | |
| "step": 846 | |
| }, | |
| { | |
| "epoch": 0.7062747550552428, | |
| "grad_norm": 0.10318465530872345, | |
| "learning_rate": 1.7831500190580823e-06, | |
| "loss": 0.5944, | |
| "step": 847 | |
| }, | |
| { | |
| "epoch": 0.707108609547634, | |
| "grad_norm": 0.10107962787151337, | |
| "learning_rate": 1.7825910684491054e-06, | |
| "loss": 0.686, | |
| "step": 848 | |
| }, | |
| { | |
| "epoch": 0.707942464040025, | |
| "grad_norm": 0.1008550301194191, | |
| "learning_rate": 1.7820314862383584e-06, | |
| "loss": 0.6356, | |
| "step": 849 | |
| }, | |
| { | |
| "epoch": 0.7087763185324161, | |
| "grad_norm": 0.1049947738647461, | |
| "learning_rate": 1.7814712728774598e-06, | |
| "loss": 0.6436, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.7096101730248072, | |
| "grad_norm": 0.10559200495481491, | |
| "learning_rate": 1.7809104288185389e-06, | |
| "loss": 0.6522, | |
| "step": 851 | |
| }, | |
| { | |
| "epoch": 0.7104440275171983, | |
| "grad_norm": 0.09807837009429932, | |
| "learning_rate": 1.7803489545142325e-06, | |
| "loss": 0.6252, | |
| "step": 852 | |
| }, | |
| { | |
| "epoch": 0.7112778820095893, | |
| "grad_norm": 0.1072913408279419, | |
| "learning_rate": 1.7797868504176874e-06, | |
| "loss": 0.6483, | |
| "step": 853 | |
| }, | |
| { | |
| "epoch": 0.7121117365019805, | |
| "grad_norm": 0.09987325221300125, | |
| "learning_rate": 1.7792241169825579e-06, | |
| "loss": 0.5955, | |
| "step": 854 | |
| }, | |
| { | |
| "epoch": 0.7129455909943715, | |
| "grad_norm": 0.10160518437623978, | |
| "learning_rate": 1.778660754663006e-06, | |
| "loss": 0.6006, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 0.7137794454867625, | |
| "grad_norm": 0.10053478926420212, | |
| "learning_rate": 1.7780967639137025e-06, | |
| "loss": 0.6583, | |
| "step": 856 | |
| }, | |
| { | |
| "epoch": 0.7146132999791537, | |
| "grad_norm": 0.10520876199007034, | |
| "learning_rate": 1.777532145189824e-06, | |
| "loss": 0.6061, | |
| "step": 857 | |
| }, | |
| { | |
| "epoch": 0.7154471544715447, | |
| "grad_norm": 0.10797824710607529, | |
| "learning_rate": 1.776966898947054e-06, | |
| "loss": 0.6764, | |
| "step": 858 | |
| }, | |
| { | |
| "epoch": 0.7162810089639358, | |
| "grad_norm": 0.10396280139684677, | |
| "learning_rate": 1.7764010256415837e-06, | |
| "loss": 0.6302, | |
| "step": 859 | |
| }, | |
| { | |
| "epoch": 0.7171148634563269, | |
| "grad_norm": 0.10683812946081161, | |
| "learning_rate": 1.7758345257301094e-06, | |
| "loss": 0.6618, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.717948717948718, | |
| "grad_norm": 0.09601815789937973, | |
| "learning_rate": 1.7752673996698326e-06, | |
| "loss": 0.6167, | |
| "step": 861 | |
| }, | |
| { | |
| "epoch": 0.718782572441109, | |
| "grad_norm": 0.1079709529876709, | |
| "learning_rate": 1.774699647918462e-06, | |
| "loss": 0.6085, | |
| "step": 862 | |
| }, | |
| { | |
| "epoch": 0.7196164269335001, | |
| "grad_norm": 0.10923435539007187, | |
| "learning_rate": 1.774131270934209e-06, | |
| "loss": 0.6492, | |
| "step": 863 | |
| }, | |
| { | |
| "epoch": 0.7204502814258912, | |
| "grad_norm": 0.10494917631149292, | |
| "learning_rate": 1.7735622691757912e-06, | |
| "loss": 0.6281, | |
| "step": 864 | |
| }, | |
| { | |
| "epoch": 0.7212841359182822, | |
| "grad_norm": 0.10305362194776535, | |
| "learning_rate": 1.7729926431024301e-06, | |
| "loss": 0.6381, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 0.7221179904106734, | |
| "grad_norm": 0.1005750373005867, | |
| "learning_rate": 1.7724223931738505e-06, | |
| "loss": 0.667, | |
| "step": 866 | |
| }, | |
| { | |
| "epoch": 0.7229518449030644, | |
| "grad_norm": 0.10202402621507645, | |
| "learning_rate": 1.7718515198502812e-06, | |
| "loss": 0.6685, | |
| "step": 867 | |
| }, | |
| { | |
| "epoch": 0.7237856993954555, | |
| "grad_norm": 0.10256760567426682, | |
| "learning_rate": 1.7712800235924546e-06, | |
| "loss": 0.6728, | |
| "step": 868 | |
| }, | |
| { | |
| "epoch": 0.7246195538878466, | |
| "grad_norm": 0.1020650714635849, | |
| "learning_rate": 1.7707079048616046e-06, | |
| "loss": 0.6144, | |
| "step": 869 | |
| }, | |
| { | |
| "epoch": 0.7254534083802376, | |
| "grad_norm": 0.10101509839296341, | |
| "learning_rate": 1.770135164119468e-06, | |
| "loss": 0.6244, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.7262872628726287, | |
| "grad_norm": 0.09737613797187805, | |
| "learning_rate": 1.769561801828284e-06, | |
| "loss": 0.6507, | |
| "step": 871 | |
| }, | |
| { | |
| "epoch": 0.7271211173650198, | |
| "grad_norm": 0.09866712987422943, | |
| "learning_rate": 1.7689878184507937e-06, | |
| "loss": 0.5875, | |
| "step": 872 | |
| }, | |
| { | |
| "epoch": 0.7279549718574109, | |
| "grad_norm": 0.0963241383433342, | |
| "learning_rate": 1.7684132144502382e-06, | |
| "loss": 0.6237, | |
| "step": 873 | |
| }, | |
| { | |
| "epoch": 0.7287888263498019, | |
| "grad_norm": 0.09706535190343857, | |
| "learning_rate": 1.7678379902903603e-06, | |
| "loss": 0.5665, | |
| "step": 874 | |
| }, | |
| { | |
| "epoch": 0.7296226808421931, | |
| "grad_norm": 0.09960032999515533, | |
| "learning_rate": 1.7672621464354034e-06, | |
| "loss": 0.6288, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 0.7304565353345841, | |
| "grad_norm": 0.10458897799253464, | |
| "learning_rate": 1.766685683350111e-06, | |
| "loss": 0.6148, | |
| "step": 876 | |
| }, | |
| { | |
| "epoch": 0.7312903898269751, | |
| "grad_norm": 0.10443190485239029, | |
| "learning_rate": 1.766108601499726e-06, | |
| "loss": 0.6248, | |
| "step": 877 | |
| }, | |
| { | |
| "epoch": 0.7321242443193663, | |
| "grad_norm": 0.10145988315343857, | |
| "learning_rate": 1.765530901349991e-06, | |
| "loss": 0.6117, | |
| "step": 878 | |
| }, | |
| { | |
| "epoch": 0.7329580988117573, | |
| "grad_norm": 0.10847010463476181, | |
| "learning_rate": 1.7649525833671474e-06, | |
| "loss": 0.7046, | |
| "step": 879 | |
| }, | |
| { | |
| "epoch": 0.7337919533041485, | |
| "grad_norm": 0.10503443330526352, | |
| "learning_rate": 1.7643736480179352e-06, | |
| "loss": 0.6176, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.7346258077965395, | |
| "grad_norm": 0.10929647833108902, | |
| "learning_rate": 1.7637940957695934e-06, | |
| "loss": 0.6434, | |
| "step": 881 | |
| }, | |
| { | |
| "epoch": 0.7354596622889306, | |
| "grad_norm": 0.10153260827064514, | |
| "learning_rate": 1.7632139270898576e-06, | |
| "loss": 0.638, | |
| "step": 882 | |
| }, | |
| { | |
| "epoch": 0.7362935167813217, | |
| "grad_norm": 0.10493889451026917, | |
| "learning_rate": 1.7626331424469615e-06, | |
| "loss": 0.6517, | |
| "step": 883 | |
| }, | |
| { | |
| "epoch": 0.7371273712737128, | |
| "grad_norm": 0.09808988124132156, | |
| "learning_rate": 1.7620517423096368e-06, | |
| "loss": 0.6179, | |
| "step": 884 | |
| }, | |
| { | |
| "epoch": 0.7379612257661038, | |
| "grad_norm": 0.10479886084794998, | |
| "learning_rate": 1.7614697271471103e-06, | |
| "loss": 0.6419, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 0.7387950802584949, | |
| "grad_norm": 0.09941962361335754, | |
| "learning_rate": 1.7608870974291065e-06, | |
| "loss": 0.6128, | |
| "step": 886 | |
| }, | |
| { | |
| "epoch": 0.739628934750886, | |
| "grad_norm": 0.10169458389282227, | |
| "learning_rate": 1.7603038536258453e-06, | |
| "loss": 0.609, | |
| "step": 887 | |
| }, | |
| { | |
| "epoch": 0.740462789243277, | |
| "grad_norm": 0.09726240485906601, | |
| "learning_rate": 1.7597199962080423e-06, | |
| "loss": 0.6371, | |
| "step": 888 | |
| }, | |
| { | |
| "epoch": 0.7412966437356682, | |
| "grad_norm": 0.10218486189842224, | |
| "learning_rate": 1.759135525646908e-06, | |
| "loss": 0.5768, | |
| "step": 889 | |
| }, | |
| { | |
| "epoch": 0.7421304982280592, | |
| "grad_norm": 0.10321173816919327, | |
| "learning_rate": 1.7585504424141483e-06, | |
| "loss": 0.6007, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.7429643527204502, | |
| "grad_norm": 0.10717494040727615, | |
| "learning_rate": 1.7579647469819631e-06, | |
| "loss": 0.6353, | |
| "step": 891 | |
| }, | |
| { | |
| "epoch": 0.7437982072128414, | |
| "grad_norm": 0.1051764264702797, | |
| "learning_rate": 1.7573784398230474e-06, | |
| "loss": 0.6098, | |
| "step": 892 | |
| }, | |
| { | |
| "epoch": 0.7446320617052324, | |
| "grad_norm": 0.10321137309074402, | |
| "learning_rate": 1.7567915214105881e-06, | |
| "loss": 0.5986, | |
| "step": 893 | |
| }, | |
| { | |
| "epoch": 0.7454659161976235, | |
| "grad_norm": 0.10979589819908142, | |
| "learning_rate": 1.7562039922182671e-06, | |
| "loss": 0.6518, | |
| "step": 894 | |
| }, | |
| { | |
| "epoch": 0.7462997706900146, | |
| "grad_norm": 0.10153292864561081, | |
| "learning_rate": 1.7556158527202585e-06, | |
| "loss": 0.6358, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 0.7471336251824057, | |
| "grad_norm": 0.10513276606798172, | |
| "learning_rate": 1.7550271033912287e-06, | |
| "loss": 0.6085, | |
| "step": 896 | |
| }, | |
| { | |
| "epoch": 0.7479674796747967, | |
| "grad_norm": 0.10275143384933472, | |
| "learning_rate": 1.7544377447063372e-06, | |
| "loss": 0.5748, | |
| "step": 897 | |
| }, | |
| { | |
| "epoch": 0.7488013341671879, | |
| "grad_norm": 0.10438092797994614, | |
| "learning_rate": 1.7538477771412339e-06, | |
| "loss": 0.5943, | |
| "step": 898 | |
| }, | |
| { | |
| "epoch": 0.7496351886595789, | |
| "grad_norm": 0.1010277271270752, | |
| "learning_rate": 1.7532572011720617e-06, | |
| "loss": 0.6112, | |
| "step": 899 | |
| }, | |
| { | |
| "epoch": 0.7504690431519699, | |
| "grad_norm": 0.10229716449975967, | |
| "learning_rate": 1.752666017275453e-06, | |
| "loss": 0.5861, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.7513028976443611, | |
| "grad_norm": 0.10199355334043503, | |
| "learning_rate": 1.7520742259285323e-06, | |
| "loss": 0.6388, | |
| "step": 901 | |
| }, | |
| { | |
| "epoch": 0.7521367521367521, | |
| "grad_norm": 0.10268741846084595, | |
| "learning_rate": 1.7514818276089128e-06, | |
| "loss": 0.6875, | |
| "step": 902 | |
| }, | |
| { | |
| "epoch": 0.7529706066291432, | |
| "grad_norm": 0.0992840975522995, | |
| "learning_rate": 1.7508888227946992e-06, | |
| "loss": 0.6756, | |
| "step": 903 | |
| }, | |
| { | |
| "epoch": 0.7538044611215343, | |
| "grad_norm": 0.09615278989076614, | |
| "learning_rate": 1.7502952119644845e-06, | |
| "loss": 0.6474, | |
| "step": 904 | |
| }, | |
| { | |
| "epoch": 0.7546383156139254, | |
| "grad_norm": 0.10315293818712234, | |
| "learning_rate": 1.749700995597351e-06, | |
| "loss": 0.6127, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 0.7554721701063164, | |
| "grad_norm": 0.10551444441080093, | |
| "learning_rate": 1.7491061741728702e-06, | |
| "loss": 0.6357, | |
| "step": 906 | |
| }, | |
| { | |
| "epoch": 0.7563060245987075, | |
| "grad_norm": 0.10280577838420868, | |
| "learning_rate": 1.748510748171101e-06, | |
| "loss": 0.6045, | |
| "step": 907 | |
| }, | |
| { | |
| "epoch": 0.7571398790910986, | |
| "grad_norm": 0.09516436606645584, | |
| "learning_rate": 1.7479147180725912e-06, | |
| "loss": 0.6368, | |
| "step": 908 | |
| }, | |
| { | |
| "epoch": 0.7579737335834896, | |
| "grad_norm": 0.10799692571163177, | |
| "learning_rate": 1.7473180843583762e-06, | |
| "loss": 0.6669, | |
| "step": 909 | |
| }, | |
| { | |
| "epoch": 0.7588075880758808, | |
| "grad_norm": 0.1064273938536644, | |
| "learning_rate": 1.7467208475099774e-06, | |
| "loss": 0.6481, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.7596414425682718, | |
| "grad_norm": 0.1054486557841301, | |
| "learning_rate": 1.7461230080094043e-06, | |
| "loss": 0.6108, | |
| "step": 911 | |
| }, | |
| { | |
| "epoch": 0.760475297060663, | |
| "grad_norm": 0.10060249269008636, | |
| "learning_rate": 1.7455245663391516e-06, | |
| "loss": 0.5604, | |
| "step": 912 | |
| }, | |
| { | |
| "epoch": 0.761309151553054, | |
| "grad_norm": 0.1015784963965416, | |
| "learning_rate": 1.7449255229822008e-06, | |
| "loss": 0.6248, | |
| "step": 913 | |
| }, | |
| { | |
| "epoch": 0.762143006045445, | |
| "grad_norm": 0.10114108771085739, | |
| "learning_rate": 1.744325878422019e-06, | |
| "loss": 0.6683, | |
| "step": 914 | |
| }, | |
| { | |
| "epoch": 0.7629768605378362, | |
| "grad_norm": 0.10236384719610214, | |
| "learning_rate": 1.7437256331425576e-06, | |
| "loss": 0.657, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 0.7638107150302272, | |
| "grad_norm": 0.10892149806022644, | |
| "learning_rate": 1.743124787628254e-06, | |
| "loss": 0.6434, | |
| "step": 916 | |
| }, | |
| { | |
| "epoch": 0.7646445695226183, | |
| "grad_norm": 0.10639214515686035, | |
| "learning_rate": 1.7425233423640298e-06, | |
| "loss": 0.5859, | |
| "step": 917 | |
| }, | |
| { | |
| "epoch": 0.7654784240150094, | |
| "grad_norm": 0.10279032588005066, | |
| "learning_rate": 1.7419212978352898e-06, | |
| "loss": 0.6779, | |
| "step": 918 | |
| }, | |
| { | |
| "epoch": 0.7663122785074005, | |
| "grad_norm": 0.11178586632013321, | |
| "learning_rate": 1.741318654527923e-06, | |
| "loss": 0.6827, | |
| "step": 919 | |
| }, | |
| { | |
| "epoch": 0.7671461329997915, | |
| "grad_norm": 0.1018877848982811, | |
| "learning_rate": 1.7407154129283019e-06, | |
| "loss": 0.6423, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.7679799874921827, | |
| "grad_norm": 0.1052117571234703, | |
| "learning_rate": 1.7401115735232817e-06, | |
| "loss": 0.6871, | |
| "step": 921 | |
| }, | |
| { | |
| "epoch": 0.7688138419845737, | |
| "grad_norm": 0.10656667500734329, | |
| "learning_rate": 1.7395071368001997e-06, | |
| "loss": 0.6384, | |
| "step": 922 | |
| }, | |
| { | |
| "epoch": 0.7696476964769647, | |
| "grad_norm": 0.09973619878292084, | |
| "learning_rate": 1.738902103246876e-06, | |
| "loss": 0.5659, | |
| "step": 923 | |
| }, | |
| { | |
| "epoch": 0.7704815509693559, | |
| "grad_norm": 0.10072916746139526, | |
| "learning_rate": 1.7382964733516112e-06, | |
| "loss": 0.6315, | |
| "step": 924 | |
| }, | |
| { | |
| "epoch": 0.7713154054617469, | |
| "grad_norm": 0.11351827532052994, | |
| "learning_rate": 1.7376902476031882e-06, | |
| "loss": 0.7005, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 0.772149259954138, | |
| "grad_norm": 0.1084694042801857, | |
| "learning_rate": 1.7370834264908711e-06, | |
| "loss": 0.6077, | |
| "step": 926 | |
| }, | |
| { | |
| "epoch": 0.7729831144465291, | |
| "grad_norm": 0.10465063899755478, | |
| "learning_rate": 1.7364760105044033e-06, | |
| "loss": 0.5937, | |
| "step": 927 | |
| }, | |
| { | |
| "epoch": 0.7738169689389202, | |
| "grad_norm": 0.09937559068202972, | |
| "learning_rate": 1.7358680001340092e-06, | |
| "loss": 0.5854, | |
| "step": 928 | |
| }, | |
| { | |
| "epoch": 0.7746508234313112, | |
| "grad_norm": 0.11517240107059479, | |
| "learning_rate": 1.7352593958703921e-06, | |
| "loss": 0.6738, | |
| "step": 929 | |
| }, | |
| { | |
| "epoch": 0.7754846779237023, | |
| "grad_norm": 0.10875809192657471, | |
| "learning_rate": 1.734650198204736e-06, | |
| "loss": 0.6487, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.7763185324160934, | |
| "grad_norm": 0.10209451615810394, | |
| "learning_rate": 1.7340404076287021e-06, | |
| "loss": 0.6277, | |
| "step": 931 | |
| }, | |
| { | |
| "epoch": 0.7771523869084844, | |
| "grad_norm": 0.11190709471702576, | |
| "learning_rate": 1.7334300246344316e-06, | |
| "loss": 0.6393, | |
| "step": 932 | |
| }, | |
| { | |
| "epoch": 0.7779862414008756, | |
| "grad_norm": 0.10622856765985489, | |
| "learning_rate": 1.7328190497145427e-06, | |
| "loss": 0.6665, | |
| "step": 933 | |
| }, | |
| { | |
| "epoch": 0.7788200958932666, | |
| "grad_norm": 0.11038866639137268, | |
| "learning_rate": 1.7322074833621318e-06, | |
| "loss": 0.6092, | |
| "step": 934 | |
| }, | |
| { | |
| "epoch": 0.7796539503856577, | |
| "grad_norm": 0.1060342788696289, | |
| "learning_rate": 1.7315953260707735e-06, | |
| "loss": 0.6427, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 0.7804878048780488, | |
| "grad_norm": 0.10829740762710571, | |
| "learning_rate": 1.7309825783345174e-06, | |
| "loss": 0.6851, | |
| "step": 936 | |
| }, | |
| { | |
| "epoch": 0.7813216593704398, | |
| "grad_norm": 0.1001635491847992, | |
| "learning_rate": 1.7303692406478908e-06, | |
| "loss": 0.5645, | |
| "step": 937 | |
| }, | |
| { | |
| "epoch": 0.7821555138628309, | |
| "grad_norm": 0.09842494130134583, | |
| "learning_rate": 1.7297553135058978e-06, | |
| "loss": 0.613, | |
| "step": 938 | |
| }, | |
| { | |
| "epoch": 0.782989368355222, | |
| "grad_norm": 0.10335814207792282, | |
| "learning_rate": 1.7291407974040167e-06, | |
| "loss": 0.6583, | |
| "step": 939 | |
| }, | |
| { | |
| "epoch": 0.7838232228476131, | |
| "grad_norm": 0.10193807631731033, | |
| "learning_rate": 1.728525692838202e-06, | |
| "loss": 0.6242, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.7846570773400041, | |
| "grad_norm": 0.1064397543668747, | |
| "learning_rate": 1.7279100003048832e-06, | |
| "loss": 0.7043, | |
| "step": 941 | |
| }, | |
| { | |
| "epoch": 0.7854909318323953, | |
| "grad_norm": 0.09790313243865967, | |
| "learning_rate": 1.7272937203009642e-06, | |
| "loss": 0.5869, | |
| "step": 942 | |
| }, | |
| { | |
| "epoch": 0.7863247863247863, | |
| "grad_norm": 0.10425697267055511, | |
| "learning_rate": 1.7266768533238225e-06, | |
| "loss": 0.6282, | |
| "step": 943 | |
| }, | |
| { | |
| "epoch": 0.7871586408171775, | |
| "grad_norm": 0.10319948196411133, | |
| "learning_rate": 1.72605939987131e-06, | |
| "loss": 0.5899, | |
| "step": 944 | |
| }, | |
| { | |
| "epoch": 0.7879924953095685, | |
| "grad_norm": 0.10117416828870773, | |
| "learning_rate": 1.725441360441752e-06, | |
| "loss": 0.5734, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 0.7888263498019595, | |
| "grad_norm": 0.10920300334692001, | |
| "learning_rate": 1.7248227355339458e-06, | |
| "loss": 0.612, | |
| "step": 946 | |
| }, | |
| { | |
| "epoch": 0.7896602042943507, | |
| "grad_norm": 0.0993722602725029, | |
| "learning_rate": 1.7242035256471623e-06, | |
| "loss": 0.5778, | |
| "step": 947 | |
| }, | |
| { | |
| "epoch": 0.7904940587867417, | |
| "grad_norm": 0.12078487873077393, | |
| "learning_rate": 1.723583731281144e-06, | |
| "loss": 0.7115, | |
| "step": 948 | |
| }, | |
| { | |
| "epoch": 0.7913279132791328, | |
| "grad_norm": 0.10817496478557587, | |
| "learning_rate": 1.7229633529361051e-06, | |
| "loss": 0.6699, | |
| "step": 949 | |
| }, | |
| { | |
| "epoch": 0.7921617677715239, | |
| "grad_norm": 0.10841819643974304, | |
| "learning_rate": 1.7223423911127313e-06, | |
| "loss": 0.6089, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.792995622263915, | |
| "grad_norm": 0.10630793124437332, | |
| "learning_rate": 1.7217208463121788e-06, | |
| "loss": 0.6615, | |
| "step": 951 | |
| }, | |
| { | |
| "epoch": 0.793829476756306, | |
| "grad_norm": 0.1000799611210823, | |
| "learning_rate": 1.721098719036075e-06, | |
| "loss": 0.6231, | |
| "step": 952 | |
| }, | |
| { | |
| "epoch": 0.7946633312486971, | |
| "grad_norm": 0.11009353399276733, | |
| "learning_rate": 1.7204760097865167e-06, | |
| "loss": 0.6197, | |
| "step": 953 | |
| }, | |
| { | |
| "epoch": 0.7954971857410882, | |
| "grad_norm": 0.10809972882270813, | |
| "learning_rate": 1.7198527190660706e-06, | |
| "loss": 0.5975, | |
| "step": 954 | |
| }, | |
| { | |
| "epoch": 0.7963310402334792, | |
| "grad_norm": 0.10163717716932297, | |
| "learning_rate": 1.719228847377773e-06, | |
| "loss": 0.7125, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 0.7971648947258704, | |
| "grad_norm": 0.1121022030711174, | |
| "learning_rate": 1.7186043952251286e-06, | |
| "loss": 0.6821, | |
| "step": 956 | |
| }, | |
| { | |
| "epoch": 0.7979987492182614, | |
| "grad_norm": 0.11300718039274216, | |
| "learning_rate": 1.7179793631121106e-06, | |
| "loss": 0.6137, | |
| "step": 957 | |
| }, | |
| { | |
| "epoch": 0.7988326037106525, | |
| "grad_norm": 0.10873489826917648, | |
| "learning_rate": 1.717353751543161e-06, | |
| "loss": 0.5621, | |
| "step": 958 | |
| }, | |
| { | |
| "epoch": 0.7996664582030436, | |
| "grad_norm": 0.11031791567802429, | |
| "learning_rate": 1.7167275610231887e-06, | |
| "loss": 0.6363, | |
| "step": 959 | |
| }, | |
| { | |
| "epoch": 0.8005003126954346, | |
| "grad_norm": 0.10201506316661835, | |
| "learning_rate": 1.7161007920575704e-06, | |
| "loss": 0.5696, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.8013341671878257, | |
| "grad_norm": 0.09877403825521469, | |
| "learning_rate": 1.7154734451521486e-06, | |
| "loss": 0.6027, | |
| "step": 961 | |
| }, | |
| { | |
| "epoch": 0.8021680216802168, | |
| "grad_norm": 0.1055438295006752, | |
| "learning_rate": 1.7148455208132334e-06, | |
| "loss": 0.6204, | |
| "step": 962 | |
| }, | |
| { | |
| "epoch": 0.8030018761726079, | |
| "grad_norm": 0.11029175668954849, | |
| "learning_rate": 1.7142170195476005e-06, | |
| "loss": 0.5826, | |
| "step": 963 | |
| }, | |
| { | |
| "epoch": 0.8038357306649989, | |
| "grad_norm": 0.1105756014585495, | |
| "learning_rate": 1.7135879418624913e-06, | |
| "loss": 0.6526, | |
| "step": 964 | |
| }, | |
| { | |
| "epoch": 0.8046695851573901, | |
| "grad_norm": 0.1054423451423645, | |
| "learning_rate": 1.712958288265612e-06, | |
| "loss": 0.6254, | |
| "step": 965 | |
| }, | |
| { | |
| "epoch": 0.8055034396497811, | |
| "grad_norm": 0.10613211989402771, | |
| "learning_rate": 1.712328059265134e-06, | |
| "loss": 0.5974, | |
| "step": 966 | |
| }, | |
| { | |
| "epoch": 0.8063372941421721, | |
| "grad_norm": 0.09556613117456436, | |
| "learning_rate": 1.7116972553696932e-06, | |
| "loss": 0.6144, | |
| "step": 967 | |
| }, | |
| { | |
| "epoch": 0.8071711486345633, | |
| "grad_norm": 0.10485559701919556, | |
| "learning_rate": 1.711065877088389e-06, | |
| "loss": 0.5998, | |
| "step": 968 | |
| }, | |
| { | |
| "epoch": 0.8080050031269543, | |
| "grad_norm": 0.10885662585496902, | |
| "learning_rate": 1.7104339249307848e-06, | |
| "loss": 0.6298, | |
| "step": 969 | |
| }, | |
| { | |
| "epoch": 0.8088388576193454, | |
| "grad_norm": 0.10140252858400345, | |
| "learning_rate": 1.709801399406907e-06, | |
| "loss": 0.5747, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.8096727121117365, | |
| "grad_norm": 0.09874554723501205, | |
| "learning_rate": 1.7091683010272446e-06, | |
| "loss": 0.5676, | |
| "step": 971 | |
| }, | |
| { | |
| "epoch": 0.8105065666041276, | |
| "grad_norm": 0.10777262598276138, | |
| "learning_rate": 1.7085346303027493e-06, | |
| "loss": 0.6326, | |
| "step": 972 | |
| }, | |
| { | |
| "epoch": 0.8113404210965186, | |
| "grad_norm": 0.09974364936351776, | |
| "learning_rate": 1.7079003877448344e-06, | |
| "loss": 0.5638, | |
| "step": 973 | |
| }, | |
| { | |
| "epoch": 0.8121742755889098, | |
| "grad_norm": 0.10661293566226959, | |
| "learning_rate": 1.7072655738653745e-06, | |
| "loss": 0.6641, | |
| "step": 974 | |
| }, | |
| { | |
| "epoch": 0.8130081300813008, | |
| "grad_norm": 0.10669861733913422, | |
| "learning_rate": 1.7066301891767061e-06, | |
| "loss": 0.6478, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 0.813841984573692, | |
| "grad_norm": 0.10368253290653229, | |
| "learning_rate": 1.7059942341916256e-06, | |
| "loss": 0.6198, | |
| "step": 976 | |
| }, | |
| { | |
| "epoch": 0.814675839066083, | |
| "grad_norm": 0.10855797678232193, | |
| "learning_rate": 1.7053577094233897e-06, | |
| "loss": 0.6453, | |
| "step": 977 | |
| }, | |
| { | |
| "epoch": 0.815509693558474, | |
| "grad_norm": 0.11007635295391083, | |
| "learning_rate": 1.7047206153857156e-06, | |
| "loss": 0.5748, | |
| "step": 978 | |
| }, | |
| { | |
| "epoch": 0.8163435480508652, | |
| "grad_norm": 0.10489246994256973, | |
| "learning_rate": 1.7040829525927796e-06, | |
| "loss": 0.579, | |
| "step": 979 | |
| }, | |
| { | |
| "epoch": 0.8171774025432562, | |
| "grad_norm": 0.11224298179149628, | |
| "learning_rate": 1.7034447215592164e-06, | |
| "loss": 0.6484, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.8180112570356473, | |
| "grad_norm": 0.10526615381240845, | |
| "learning_rate": 1.7028059228001204e-06, | |
| "loss": 0.6746, | |
| "step": 981 | |
| }, | |
| { | |
| "epoch": 0.8188451115280384, | |
| "grad_norm": 0.10187442600727081, | |
| "learning_rate": 1.7021665568310435e-06, | |
| "loss": 0.6462, | |
| "step": 982 | |
| }, | |
| { | |
| "epoch": 0.8196789660204294, | |
| "grad_norm": 0.10846269875764847, | |
| "learning_rate": 1.7015266241679952e-06, | |
| "loss": 0.6329, | |
| "step": 983 | |
| }, | |
| { | |
| "epoch": 0.8205128205128205, | |
| "grad_norm": 0.110533706843853, | |
| "learning_rate": 1.7008861253274429e-06, | |
| "loss": 0.658, | |
| "step": 984 | |
| }, | |
| { | |
| "epoch": 0.8213466750052116, | |
| "grad_norm": 0.10688811540603638, | |
| "learning_rate": 1.7002450608263107e-06, | |
| "loss": 0.5996, | |
| "step": 985 | |
| }, | |
| { | |
| "epoch": 0.8221805294976027, | |
| "grad_norm": 0.10414768755435944, | |
| "learning_rate": 1.6996034311819796e-06, | |
| "loss": 0.6333, | |
| "step": 986 | |
| }, | |
| { | |
| "epoch": 0.8230143839899937, | |
| "grad_norm": 0.0999814048409462, | |
| "learning_rate": 1.698961236912286e-06, | |
| "loss": 0.6609, | |
| "step": 987 | |
| }, | |
| { | |
| "epoch": 0.8238482384823849, | |
| "grad_norm": 0.10203441977500916, | |
| "learning_rate": 1.6983184785355222e-06, | |
| "loss": 0.6095, | |
| "step": 988 | |
| }, | |
| { | |
| "epoch": 0.8246820929747759, | |
| "grad_norm": 0.10571661591529846, | |
| "learning_rate": 1.6976751565704362e-06, | |
| "loss": 0.6125, | |
| "step": 989 | |
| }, | |
| { | |
| "epoch": 0.8255159474671669, | |
| "grad_norm": 0.10715372860431671, | |
| "learning_rate": 1.6970312715362304e-06, | |
| "loss": 0.689, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.8263498019595581, | |
| "grad_norm": 0.10807687044143677, | |
| "learning_rate": 1.696386823952562e-06, | |
| "loss": 0.611, | |
| "step": 991 | |
| }, | |
| { | |
| "epoch": 0.8271836564519491, | |
| "grad_norm": 0.10625305026769638, | |
| "learning_rate": 1.6957418143395418e-06, | |
| "loss": 0.6616, | |
| "step": 992 | |
| }, | |
| { | |
| "epoch": 0.8280175109443402, | |
| "grad_norm": 0.10512302070856094, | |
| "learning_rate": 1.6950962432177348e-06, | |
| "loss": 0.6147, | |
| "step": 993 | |
| }, | |
| { | |
| "epoch": 0.8288513654367313, | |
| "grad_norm": 0.10524202138185501, | |
| "learning_rate": 1.6944501111081579e-06, | |
| "loss": 0.6094, | |
| "step": 994 | |
| }, | |
| { | |
| "epoch": 0.8296852199291224, | |
| "grad_norm": 0.10839590430259705, | |
| "learning_rate": 1.6938034185322828e-06, | |
| "loss": 0.5816, | |
| "step": 995 | |
| }, | |
| { | |
| "epoch": 0.8305190744215134, | |
| "grad_norm": 0.10930492728948593, | |
| "learning_rate": 1.6931561660120312e-06, | |
| "loss": 0.6621, | |
| "step": 996 | |
| }, | |
| { | |
| "epoch": 0.8313529289139046, | |
| "grad_norm": 0.10111601650714874, | |
| "learning_rate": 1.692508354069779e-06, | |
| "loss": 0.6083, | |
| "step": 997 | |
| }, | |
| { | |
| "epoch": 0.8321867834062956, | |
| "grad_norm": 0.11010827124118805, | |
| "learning_rate": 1.691859983228352e-06, | |
| "loss": 0.6689, | |
| "step": 998 | |
| }, | |
| { | |
| "epoch": 0.8330206378986866, | |
| "grad_norm": 0.10622293502092361, | |
| "learning_rate": 1.6912110540110272e-06, | |
| "loss": 0.6142, | |
| "step": 999 | |
| }, | |
| { | |
| "epoch": 0.8338544923910778, | |
| "grad_norm": 0.10322947055101395, | |
| "learning_rate": 1.6905615669415325e-06, | |
| "loss": 0.5797, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.8346883468834688, | |
| "grad_norm": 0.11376485228538513, | |
| "learning_rate": 1.689911522544047e-06, | |
| "loss": 0.6323, | |
| "step": 1001 | |
| }, | |
| { | |
| "epoch": 0.8355222013758599, | |
| "grad_norm": 0.10336299985647202, | |
| "learning_rate": 1.6892609213431981e-06, | |
| "loss": 0.5907, | |
| "step": 1002 | |
| }, | |
| { | |
| "epoch": 0.836356055868251, | |
| "grad_norm": 0.10549134016036987, | |
| "learning_rate": 1.6886097638640631e-06, | |
| "loss": 0.6483, | |
| "step": 1003 | |
| }, | |
| { | |
| "epoch": 0.837189910360642, | |
| "grad_norm": 0.10979744791984558, | |
| "learning_rate": 1.6879580506321687e-06, | |
| "loss": 0.5829, | |
| "step": 1004 | |
| }, | |
| { | |
| "epoch": 0.8380237648530331, | |
| "grad_norm": 0.10647737979888916, | |
| "learning_rate": 1.6873057821734894e-06, | |
| "loss": 0.5714, | |
| "step": 1005 | |
| }, | |
| { | |
| "epoch": 0.8388576193454242, | |
| "grad_norm": 0.11350049823522568, | |
| "learning_rate": 1.6866529590144485e-06, | |
| "loss": 0.6756, | |
| "step": 1006 | |
| }, | |
| { | |
| "epoch": 0.8396914738378153, | |
| "grad_norm": 0.10672125220298767, | |
| "learning_rate": 1.6859995816819167e-06, | |
| "loss": 0.6459, | |
| "step": 1007 | |
| }, | |
| { | |
| "epoch": 0.8405253283302064, | |
| "grad_norm": 0.11561840772628784, | |
| "learning_rate": 1.6853456507032112e-06, | |
| "loss": 0.6912, | |
| "step": 1008 | |
| }, | |
| { | |
| "epoch": 0.8413591828225975, | |
| "grad_norm": 0.11554259806871414, | |
| "learning_rate": 1.6846911666060973e-06, | |
| "loss": 0.6032, | |
| "step": 1009 | |
| }, | |
| { | |
| "epoch": 0.8421930373149885, | |
| "grad_norm": 0.10536365956068039, | |
| "learning_rate": 1.6840361299187857e-06, | |
| "loss": 0.6182, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.8430268918073797, | |
| "grad_norm": 0.1125800758600235, | |
| "learning_rate": 1.6833805411699344e-06, | |
| "loss": 0.6519, | |
| "step": 1011 | |
| }, | |
| { | |
| "epoch": 0.8438607462997707, | |
| "grad_norm": 0.10594160109758377, | |
| "learning_rate": 1.6827244008886453e-06, | |
| "loss": 0.6407, | |
| "step": 1012 | |
| }, | |
| { | |
| "epoch": 0.8446946007921617, | |
| "grad_norm": 0.10480530560016632, | |
| "learning_rate": 1.6820677096044667e-06, | |
| "loss": 0.6398, | |
| "step": 1013 | |
| }, | |
| { | |
| "epoch": 0.8455284552845529, | |
| "grad_norm": 0.10666308552026749, | |
| "learning_rate": 1.6814104678473905e-06, | |
| "loss": 0.612, | |
| "step": 1014 | |
| }, | |
| { | |
| "epoch": 0.8463623097769439, | |
| "grad_norm": 0.10556510835886002, | |
| "learning_rate": 1.6807526761478533e-06, | |
| "loss": 0.5729, | |
| "step": 1015 | |
| }, | |
| { | |
| "epoch": 0.847196164269335, | |
| "grad_norm": 0.11039448529481888, | |
| "learning_rate": 1.6800943350367368e-06, | |
| "loss": 0.6323, | |
| "step": 1016 | |
| }, | |
| { | |
| "epoch": 0.8480300187617261, | |
| "grad_norm": 0.10304141789674759, | |
| "learning_rate": 1.6794354450453638e-06, | |
| "loss": 0.6426, | |
| "step": 1017 | |
| }, | |
| { | |
| "epoch": 0.8488638732541172, | |
| "grad_norm": 0.10969959199428558, | |
| "learning_rate": 1.678776006705502e-06, | |
| "loss": 0.558, | |
| "step": 1018 | |
| }, | |
| { | |
| "epoch": 0.8496977277465082, | |
| "grad_norm": 0.10663289576768875, | |
| "learning_rate": 1.6781160205493605e-06, | |
| "loss": 0.593, | |
| "step": 1019 | |
| }, | |
| { | |
| "epoch": 0.8505315822388994, | |
| "grad_norm": 0.10291384905576706, | |
| "learning_rate": 1.6774554871095915e-06, | |
| "loss": 0.6281, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.8513654367312904, | |
| "grad_norm": 0.10907971113920212, | |
| "learning_rate": 1.6767944069192876e-06, | |
| "loss": 0.6158, | |
| "step": 1021 | |
| }, | |
| { | |
| "epoch": 0.8521992912236814, | |
| "grad_norm": 0.11564616858959198, | |
| "learning_rate": 1.6761327805119838e-06, | |
| "loss": 0.6662, | |
| "step": 1022 | |
| }, | |
| { | |
| "epoch": 0.8530331457160726, | |
| "grad_norm": 0.10590541362762451, | |
| "learning_rate": 1.6754706084216555e-06, | |
| "loss": 0.6323, | |
| "step": 1023 | |
| }, | |
| { | |
| "epoch": 0.8538670002084636, | |
| "grad_norm": 0.10117157548666, | |
| "learning_rate": 1.6748078911827187e-06, | |
| "loss": 0.6112, | |
| "step": 1024 | |
| }, | |
| { | |
| "epoch": 0.8547008547008547, | |
| "grad_norm": 0.11531368643045425, | |
| "learning_rate": 1.674144629330029e-06, | |
| "loss": 0.6413, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 0.8555347091932458, | |
| "grad_norm": 0.10621378570795059, | |
| "learning_rate": 1.673480823398882e-06, | |
| "loss": 0.6394, | |
| "step": 1026 | |
| }, | |
| { | |
| "epoch": 0.8563685636856369, | |
| "grad_norm": 0.11612808704376221, | |
| "learning_rate": 1.672816473925012e-06, | |
| "loss": 0.637, | |
| "step": 1027 | |
| }, | |
| { | |
| "epoch": 0.8572024181780279, | |
| "grad_norm": 0.10893040150403976, | |
| "learning_rate": 1.672151581444592e-06, | |
| "loss": 0.5923, | |
| "step": 1028 | |
| }, | |
| { | |
| "epoch": 0.858036272670419, | |
| "grad_norm": 0.10822536796331406, | |
| "learning_rate": 1.6714861464942333e-06, | |
| "loss": 0.6766, | |
| "step": 1029 | |
| }, | |
| { | |
| "epoch": 0.8588701271628101, | |
| "grad_norm": 0.10714786499738693, | |
| "learning_rate": 1.6708201696109856e-06, | |
| "loss": 0.5649, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.8597039816552011, | |
| "grad_norm": 0.10966484993696213, | |
| "learning_rate": 1.6701536513323349e-06, | |
| "loss": 0.6219, | |
| "step": 1031 | |
| }, | |
| { | |
| "epoch": 0.8605378361475923, | |
| "grad_norm": 0.1037144884467125, | |
| "learning_rate": 1.669486592196205e-06, | |
| "loss": 0.6352, | |
| "step": 1032 | |
| }, | |
| { | |
| "epoch": 0.8613716906399833, | |
| "grad_norm": 0.11299126595258713, | |
| "learning_rate": 1.668818992740956e-06, | |
| "loss": 0.6734, | |
| "step": 1033 | |
| }, | |
| { | |
| "epoch": 0.8622055451323744, | |
| "grad_norm": 0.11029834300279617, | |
| "learning_rate": 1.6681508535053834e-06, | |
| "loss": 0.6579, | |
| "step": 1034 | |
| }, | |
| { | |
| "epoch": 0.8630393996247655, | |
| "grad_norm": 0.10519642382860184, | |
| "learning_rate": 1.6674821750287197e-06, | |
| "loss": 0.6072, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 0.8638732541171565, | |
| "grad_norm": 0.11065202951431274, | |
| "learning_rate": 1.6668129578506313e-06, | |
| "loss": 0.6562, | |
| "step": 1036 | |
| }, | |
| { | |
| "epoch": 0.8647071086095476, | |
| "grad_norm": 0.11024871468544006, | |
| "learning_rate": 1.66614320251122e-06, | |
| "loss": 0.6, | |
| "step": 1037 | |
| }, | |
| { | |
| "epoch": 0.8655409631019387, | |
| "grad_norm": 0.10754359513521194, | |
| "learning_rate": 1.6654729095510219e-06, | |
| "loss": 0.6028, | |
| "step": 1038 | |
| }, | |
| { | |
| "epoch": 0.8663748175943298, | |
| "grad_norm": 0.11345981806516647, | |
| "learning_rate": 1.6648020795110069e-06, | |
| "loss": 0.7018, | |
| "step": 1039 | |
| }, | |
| { | |
| "epoch": 0.8672086720867209, | |
| "grad_norm": 0.10346484929323196, | |
| "learning_rate": 1.6641307129325783e-06, | |
| "loss": 0.6866, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.868042526579112, | |
| "grad_norm": 0.10640288889408112, | |
| "learning_rate": 1.6634588103575723e-06, | |
| "loss": 0.5851, | |
| "step": 1041 | |
| }, | |
| { | |
| "epoch": 0.868876381071503, | |
| "grad_norm": 0.11161711812019348, | |
| "learning_rate": 1.6627863723282584e-06, | |
| "loss": 0.6182, | |
| "step": 1042 | |
| }, | |
| { | |
| "epoch": 0.8697102355638942, | |
| "grad_norm": 0.10999694466590881, | |
| "learning_rate": 1.662113399387337e-06, | |
| "loss": 0.5887, | |
| "step": 1043 | |
| }, | |
| { | |
| "epoch": 0.8705440900562852, | |
| "grad_norm": 0.10471642017364502, | |
| "learning_rate": 1.6614398920779418e-06, | |
| "loss": 0.6854, | |
| "step": 1044 | |
| }, | |
| { | |
| "epoch": 0.8713779445486762, | |
| "grad_norm": 0.10635983943939209, | |
| "learning_rate": 1.660765850943636e-06, | |
| "loss": 0.5945, | |
| "step": 1045 | |
| }, | |
| { | |
| "epoch": 0.8722117990410674, | |
| "grad_norm": 0.10250985622406006, | |
| "learning_rate": 1.6600912765284153e-06, | |
| "loss": 0.6283, | |
| "step": 1046 | |
| }, | |
| { | |
| "epoch": 0.8730456535334584, | |
| "grad_norm": 0.10342993587255478, | |
| "learning_rate": 1.6594161693767046e-06, | |
| "loss": 0.5878, | |
| "step": 1047 | |
| }, | |
| { | |
| "epoch": 0.8738795080258495, | |
| "grad_norm": 0.10388782620429993, | |
| "learning_rate": 1.6587405300333593e-06, | |
| "loss": 0.6197, | |
| "step": 1048 | |
| }, | |
| { | |
| "epoch": 0.8747133625182406, | |
| "grad_norm": 0.10757216066122055, | |
| "learning_rate": 1.658064359043664e-06, | |
| "loss": 0.6643, | |
| "step": 1049 | |
| }, | |
| { | |
| "epoch": 0.8755472170106317, | |
| "grad_norm": 0.10185371339321136, | |
| "learning_rate": 1.657387656953333e-06, | |
| "loss": 0.6129, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.8763810715030227, | |
| "grad_norm": 0.10625848174095154, | |
| "learning_rate": 1.6567104243085081e-06, | |
| "loss": 0.6389, | |
| "step": 1051 | |
| }, | |
| { | |
| "epoch": 0.8772149259954138, | |
| "grad_norm": 0.1068265438079834, | |
| "learning_rate": 1.6560326616557605e-06, | |
| "loss": 0.6474, | |
| "step": 1052 | |
| }, | |
| { | |
| "epoch": 0.8780487804878049, | |
| "grad_norm": 0.10008926689624786, | |
| "learning_rate": 1.6553543695420888e-06, | |
| "loss": 0.592, | |
| "step": 1053 | |
| }, | |
| { | |
| "epoch": 0.8788826349801959, | |
| "grad_norm": 0.11250849813222885, | |
| "learning_rate": 1.6546755485149181e-06, | |
| "loss": 0.6612, | |
| "step": 1054 | |
| }, | |
| { | |
| "epoch": 0.8797164894725871, | |
| "grad_norm": 0.10780669748783112, | |
| "learning_rate": 1.6539961991221017e-06, | |
| "loss": 0.5717, | |
| "step": 1055 | |
| }, | |
| { | |
| "epoch": 0.8805503439649781, | |
| "grad_norm": 0.10522449761629105, | |
| "learning_rate": 1.6533163219119181e-06, | |
| "loss": 0.6377, | |
| "step": 1056 | |
| }, | |
| { | |
| "epoch": 0.8813841984573692, | |
| "grad_norm": 0.11719299852848053, | |
| "learning_rate": 1.6526359174330727e-06, | |
| "loss": 0.6148, | |
| "step": 1057 | |
| }, | |
| { | |
| "epoch": 0.8822180529497603, | |
| "grad_norm": 0.10418464988470078, | |
| "learning_rate": 1.6519549862346959e-06, | |
| "loss": 0.547, | |
| "step": 1058 | |
| }, | |
| { | |
| "epoch": 0.8830519074421513, | |
| "grad_norm": 0.1149667277932167, | |
| "learning_rate": 1.6512735288663433e-06, | |
| "loss": 0.5509, | |
| "step": 1059 | |
| }, | |
| { | |
| "epoch": 0.8838857619345424, | |
| "grad_norm": 0.11353281885385513, | |
| "learning_rate": 1.6505915458779954e-06, | |
| "loss": 0.6679, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.8847196164269335, | |
| "grad_norm": 0.10084139555692673, | |
| "learning_rate": 1.6499090378200564e-06, | |
| "loss": 0.6315, | |
| "step": 1061 | |
| }, | |
| { | |
| "epoch": 0.8855534709193246, | |
| "grad_norm": 0.10753504931926727, | |
| "learning_rate": 1.6492260052433551e-06, | |
| "loss": 0.6441, | |
| "step": 1062 | |
| }, | |
| { | |
| "epoch": 0.8863873254117156, | |
| "grad_norm": 0.10944227129220963, | |
| "learning_rate": 1.6485424486991427e-06, | |
| "loss": 0.6378, | |
| "step": 1063 | |
| }, | |
| { | |
| "epoch": 0.8872211799041068, | |
| "grad_norm": 0.10542403161525726, | |
| "learning_rate": 1.6478583687390937e-06, | |
| "loss": 0.5715, | |
| "step": 1064 | |
| }, | |
| { | |
| "epoch": 0.8880550343964978, | |
| "grad_norm": 0.10192226618528366, | |
| "learning_rate": 1.6471737659153054e-06, | |
| "loss": 0.6035, | |
| "step": 1065 | |
| }, | |
| { | |
| "epoch": 0.8888888888888888, | |
| "grad_norm": 0.11085714399814606, | |
| "learning_rate": 1.6464886407802958e-06, | |
| "loss": 0.6705, | |
| "step": 1066 | |
| }, | |
| { | |
| "epoch": 0.88972274338128, | |
| "grad_norm": 0.10521090775728226, | |
| "learning_rate": 1.645802993887006e-06, | |
| "loss": 0.6487, | |
| "step": 1067 | |
| }, | |
| { | |
| "epoch": 0.890556597873671, | |
| "grad_norm": 0.11182847619056702, | |
| "learning_rate": 1.645116825788798e-06, | |
| "loss": 0.5908, | |
| "step": 1068 | |
| }, | |
| { | |
| "epoch": 0.8913904523660621, | |
| "grad_norm": 0.10744032263755798, | |
| "learning_rate": 1.6444301370394533e-06, | |
| "loss": 0.6978, | |
| "step": 1069 | |
| }, | |
| { | |
| "epoch": 0.8922243068584532, | |
| "grad_norm": 0.11645323783159256, | |
| "learning_rate": 1.6437429281931742e-06, | |
| "loss": 0.6552, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.8930581613508443, | |
| "grad_norm": 0.10556191205978394, | |
| "learning_rate": 1.6430551998045833e-06, | |
| "loss": 0.6485, | |
| "step": 1071 | |
| }, | |
| { | |
| "epoch": 0.8938920158432354, | |
| "grad_norm": 0.11244357377290726, | |
| "learning_rate": 1.6423669524287216e-06, | |
| "loss": 0.6352, | |
| "step": 1072 | |
| }, | |
| { | |
| "epoch": 0.8947258703356264, | |
| "grad_norm": 0.11015337705612183, | |
| "learning_rate": 1.6416781866210494e-06, | |
| "loss": 0.5928, | |
| "step": 1073 | |
| }, | |
| { | |
| "epoch": 0.8955597248280175, | |
| "grad_norm": 0.11401670426130295, | |
| "learning_rate": 1.6409889029374457e-06, | |
| "loss": 0.591, | |
| "step": 1074 | |
| }, | |
| { | |
| "epoch": 0.8963935793204086, | |
| "grad_norm": 0.10178755223751068, | |
| "learning_rate": 1.6402991019342073e-06, | |
| "loss": 0.5772, | |
| "step": 1075 | |
| }, | |
| { | |
| "epoch": 0.8972274338127997, | |
| "grad_norm": 0.10722572356462479, | |
| "learning_rate": 1.6396087841680478e-06, | |
| "loss": 0.616, | |
| "step": 1076 | |
| }, | |
| { | |
| "epoch": 0.8980612883051907, | |
| "grad_norm": 0.10824041068553925, | |
| "learning_rate": 1.6389179501960987e-06, | |
| "loss": 0.6289, | |
| "step": 1077 | |
| }, | |
| { | |
| "epoch": 0.8988951427975819, | |
| "grad_norm": 0.10898349434137344, | |
| "learning_rate": 1.638226600575908e-06, | |
| "loss": 0.6197, | |
| "step": 1078 | |
| }, | |
| { | |
| "epoch": 0.8997289972899729, | |
| "grad_norm": 0.10828059166669846, | |
| "learning_rate": 1.6375347358654397e-06, | |
| "loss": 0.6335, | |
| "step": 1079 | |
| }, | |
| { | |
| "epoch": 0.900562851782364, | |
| "grad_norm": 0.11353094130754471, | |
| "learning_rate": 1.6368423566230728e-06, | |
| "loss": 0.6306, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.9013967062747551, | |
| "grad_norm": 0.1095949038863182, | |
| "learning_rate": 1.6361494634076033e-06, | |
| "loss": 0.6579, | |
| "step": 1081 | |
| }, | |
| { | |
| "epoch": 0.9022305607671461, | |
| "grad_norm": 0.10779423266649246, | |
| "learning_rate": 1.6354560567782406e-06, | |
| "loss": 0.6268, | |
| "step": 1082 | |
| }, | |
| { | |
| "epoch": 0.9030644152595372, | |
| "grad_norm": 0.10539865493774414, | |
| "learning_rate": 1.6347621372946088e-06, | |
| "loss": 0.6318, | |
| "step": 1083 | |
| }, | |
| { | |
| "epoch": 0.9038982697519283, | |
| "grad_norm": 0.11474636197090149, | |
| "learning_rate": 1.6340677055167458e-06, | |
| "loss": 0.6843, | |
| "step": 1084 | |
| }, | |
| { | |
| "epoch": 0.9047321242443194, | |
| "grad_norm": 0.11284147948026657, | |
| "learning_rate": 1.633372762005103e-06, | |
| "loss": 0.6167, | |
| "step": 1085 | |
| }, | |
| { | |
| "epoch": 0.9055659787367104, | |
| "grad_norm": 0.10842804610729218, | |
| "learning_rate": 1.632677307320545e-06, | |
| "loss": 0.5851, | |
| "step": 1086 | |
| }, | |
| { | |
| "epoch": 0.9063998332291016, | |
| "grad_norm": 0.10843487083911896, | |
| "learning_rate": 1.6319813420243495e-06, | |
| "loss": 0.6337, | |
| "step": 1087 | |
| }, | |
| { | |
| "epoch": 0.9072336877214926, | |
| "grad_norm": 0.10792160034179688, | |
| "learning_rate": 1.6312848666782048e-06, | |
| "loss": 0.6015, | |
| "step": 1088 | |
| }, | |
| { | |
| "epoch": 0.9080675422138836, | |
| "grad_norm": 0.10741087794303894, | |
| "learning_rate": 1.6305878818442122e-06, | |
| "loss": 0.6285, | |
| "step": 1089 | |
| }, | |
| { | |
| "epoch": 0.9089013967062748, | |
| "grad_norm": 0.10371655970811844, | |
| "learning_rate": 1.6298903880848834e-06, | |
| "loss": 0.5982, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.9097352511986658, | |
| "grad_norm": 0.10612435638904572, | |
| "learning_rate": 1.6291923859631415e-06, | |
| "loss": 0.5962, | |
| "step": 1091 | |
| }, | |
| { | |
| "epoch": 0.9105691056910569, | |
| "grad_norm": 0.10705637186765671, | |
| "learning_rate": 1.6284938760423188e-06, | |
| "loss": 0.5926, | |
| "step": 1092 | |
| }, | |
| { | |
| "epoch": 0.911402960183448, | |
| "grad_norm": 0.10797090083360672, | |
| "learning_rate": 1.627794858886159e-06, | |
| "loss": 0.632, | |
| "step": 1093 | |
| }, | |
| { | |
| "epoch": 0.9122368146758391, | |
| "grad_norm": 0.1132771223783493, | |
| "learning_rate": 1.6270953350588137e-06, | |
| "loss": 0.647, | |
| "step": 1094 | |
| }, | |
| { | |
| "epoch": 0.9130706691682301, | |
| "grad_norm": 0.10793612152338028, | |
| "learning_rate": 1.626395305124844e-06, | |
| "loss": 0.6367, | |
| "step": 1095 | |
| }, | |
| { | |
| "epoch": 0.9139045236606212, | |
| "grad_norm": 0.10706419497728348, | |
| "learning_rate": 1.6256947696492196e-06, | |
| "loss": 0.6472, | |
| "step": 1096 | |
| }, | |
| { | |
| "epoch": 0.9147383781530123, | |
| "grad_norm": 0.10842972993850708, | |
| "learning_rate": 1.6249937291973184e-06, | |
| "loss": 0.6185, | |
| "step": 1097 | |
| }, | |
| { | |
| "epoch": 0.9155722326454033, | |
| "grad_norm": 0.11361444741487503, | |
| "learning_rate": 1.6242921843349252e-06, | |
| "loss": 0.617, | |
| "step": 1098 | |
| }, | |
| { | |
| "epoch": 0.9164060871377945, | |
| "grad_norm": 0.11315753310918808, | |
| "learning_rate": 1.6235901356282322e-06, | |
| "loss": 0.6478, | |
| "step": 1099 | |
| }, | |
| { | |
| "epoch": 0.9172399416301855, | |
| "grad_norm": 0.12495766580104828, | |
| "learning_rate": 1.6228875836438385e-06, | |
| "loss": 0.6252, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.9180737961225766, | |
| "grad_norm": 0.11157601326704025, | |
| "learning_rate": 1.622184528948749e-06, | |
| "loss": 0.6214, | |
| "step": 1101 | |
| }, | |
| { | |
| "epoch": 0.9189076506149677, | |
| "grad_norm": 0.10694071650505066, | |
| "learning_rate": 1.6214809721103744e-06, | |
| "loss": 0.619, | |
| "step": 1102 | |
| }, | |
| { | |
| "epoch": 0.9197415051073587, | |
| "grad_norm": 0.1112818494439125, | |
| "learning_rate": 1.6207769136965307e-06, | |
| "loss": 0.6229, | |
| "step": 1103 | |
| }, | |
| { | |
| "epoch": 0.9205753595997499, | |
| "grad_norm": 0.10422814637422562, | |
| "learning_rate": 1.6200723542754389e-06, | |
| "loss": 0.5837, | |
| "step": 1104 | |
| }, | |
| { | |
| "epoch": 0.9214092140921409, | |
| "grad_norm": 0.11240324378013611, | |
| "learning_rate": 1.619367294415724e-06, | |
| "loss": 0.7071, | |
| "step": 1105 | |
| }, | |
| { | |
| "epoch": 0.922243068584532, | |
| "grad_norm": 0.10819468647241592, | |
| "learning_rate": 1.6186617346864151e-06, | |
| "loss": 0.5728, | |
| "step": 1106 | |
| }, | |
| { | |
| "epoch": 0.9230769230769231, | |
| "grad_norm": 0.1053207591176033, | |
| "learning_rate": 1.6179556756569448e-06, | |
| "loss": 0.5915, | |
| "step": 1107 | |
| }, | |
| { | |
| "epoch": 0.9239107775693142, | |
| "grad_norm": 0.1075674444437027, | |
| "learning_rate": 1.6172491178971482e-06, | |
| "loss": 0.5935, | |
| "step": 1108 | |
| }, | |
| { | |
| "epoch": 0.9247446320617052, | |
| "grad_norm": 0.11159633845090866, | |
| "learning_rate": 1.6165420619772635e-06, | |
| "loss": 0.6493, | |
| "step": 1109 | |
| }, | |
| { | |
| "epoch": 0.9255784865540964, | |
| "grad_norm": 0.11671009659767151, | |
| "learning_rate": 1.6158345084679307e-06, | |
| "loss": 0.6661, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.9264123410464874, | |
| "grad_norm": 0.11934785544872284, | |
| "learning_rate": 1.6151264579401917e-06, | |
| "loss": 0.6335, | |
| "step": 1111 | |
| }, | |
| { | |
| "epoch": 0.9272461955388784, | |
| "grad_norm": 0.11291314661502838, | |
| "learning_rate": 1.6144179109654887e-06, | |
| "loss": 0.6413, | |
| "step": 1112 | |
| }, | |
| { | |
| "epoch": 0.9280800500312696, | |
| "grad_norm": 0.10902924090623856, | |
| "learning_rate": 1.6137088681156654e-06, | |
| "loss": 0.6713, | |
| "step": 1113 | |
| }, | |
| { | |
| "epoch": 0.9289139045236606, | |
| "grad_norm": 0.10608460754156113, | |
| "learning_rate": 1.6129993299629651e-06, | |
| "loss": 0.6024, | |
| "step": 1114 | |
| }, | |
| { | |
| "epoch": 0.9297477590160517, | |
| "grad_norm": 0.10624940693378448, | |
| "learning_rate": 1.6122892970800317e-06, | |
| "loss": 0.6128, | |
| "step": 1115 | |
| }, | |
| { | |
| "epoch": 0.9305816135084428, | |
| "grad_norm": 0.11591742187738419, | |
| "learning_rate": 1.6115787700399071e-06, | |
| "loss": 0.6242, | |
| "step": 1116 | |
| }, | |
| { | |
| "epoch": 0.9314154680008339, | |
| "grad_norm": 0.11706274002790451, | |
| "learning_rate": 1.610867749416033e-06, | |
| "loss": 0.6356, | |
| "step": 1117 | |
| }, | |
| { | |
| "epoch": 0.9322493224932249, | |
| "grad_norm": 0.10509736090898514, | |
| "learning_rate": 1.6101562357822491e-06, | |
| "loss": 0.6337, | |
| "step": 1118 | |
| }, | |
| { | |
| "epoch": 0.933083176985616, | |
| "grad_norm": 0.10801702737808228, | |
| "learning_rate": 1.6094442297127935e-06, | |
| "loss": 0.615, | |
| "step": 1119 | |
| }, | |
| { | |
| "epoch": 0.9339170314780071, | |
| "grad_norm": 0.11627451330423355, | |
| "learning_rate": 1.6087317317823007e-06, | |
| "loss": 0.6059, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.9347508859703981, | |
| "grad_norm": 0.11316727101802826, | |
| "learning_rate": 1.6080187425658033e-06, | |
| "loss": 0.6092, | |
| "step": 1121 | |
| }, | |
| { | |
| "epoch": 0.9355847404627893, | |
| "grad_norm": 0.10899489372968674, | |
| "learning_rate": 1.6073052626387296e-06, | |
| "loss": 0.5787, | |
| "step": 1122 | |
| }, | |
| { | |
| "epoch": 0.9364185949551803, | |
| "grad_norm": 0.10977214574813843, | |
| "learning_rate": 1.606591292576904e-06, | |
| "loss": 0.6022, | |
| "step": 1123 | |
| }, | |
| { | |
| "epoch": 0.9372524494475714, | |
| "grad_norm": 0.10926050692796707, | |
| "learning_rate": 1.6058768329565469e-06, | |
| "loss": 0.6626, | |
| "step": 1124 | |
| }, | |
| { | |
| "epoch": 0.9380863039399625, | |
| "grad_norm": 0.10779014229774475, | |
| "learning_rate": 1.6051618843542736e-06, | |
| "loss": 0.6381, | |
| "step": 1125 | |
| }, | |
| { | |
| "epoch": 0.9389201584323535, | |
| "grad_norm": 0.11509402096271515, | |
| "learning_rate": 1.6044464473470939e-06, | |
| "loss": 0.5817, | |
| "step": 1126 | |
| }, | |
| { | |
| "epoch": 0.9397540129247446, | |
| "grad_norm": 0.10561665147542953, | |
| "learning_rate": 1.6037305225124121e-06, | |
| "loss": 0.5634, | |
| "step": 1127 | |
| }, | |
| { | |
| "epoch": 0.9405878674171357, | |
| "grad_norm": 0.1081278994679451, | |
| "learning_rate": 1.6030141104280253e-06, | |
| "loss": 0.6437, | |
| "step": 1128 | |
| }, | |
| { | |
| "epoch": 0.9414217219095268, | |
| "grad_norm": 0.10470031946897507, | |
| "learning_rate": 1.6022972116721256e-06, | |
| "loss": 0.5853, | |
| "step": 1129 | |
| }, | |
| { | |
| "epoch": 0.9422555764019178, | |
| "grad_norm": 0.1069476306438446, | |
| "learning_rate": 1.601579826823296e-06, | |
| "loss": 0.5653, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.943089430894309, | |
| "grad_norm": 0.1095786914229393, | |
| "learning_rate": 1.600861956460513e-06, | |
| "loss": 0.6191, | |
| "step": 1131 | |
| }, | |
| { | |
| "epoch": 0.9439232853867, | |
| "grad_norm": 0.10863108187913895, | |
| "learning_rate": 1.600143601163144e-06, | |
| "loss": 0.6166, | |
| "step": 1132 | |
| }, | |
| { | |
| "epoch": 0.944757139879091, | |
| "grad_norm": 0.110601507127285, | |
| "learning_rate": 1.5994247615109498e-06, | |
| "loss": 0.5716, | |
| "step": 1133 | |
| }, | |
| { | |
| "epoch": 0.9455909943714822, | |
| "grad_norm": 0.108455128967762, | |
| "learning_rate": 1.598705438084079e-06, | |
| "loss": 0.5888, | |
| "step": 1134 | |
| }, | |
| { | |
| "epoch": 0.9464248488638732, | |
| "grad_norm": 0.11156295984983444, | |
| "learning_rate": 1.5979856314630728e-06, | |
| "loss": 0.6723, | |
| "step": 1135 | |
| }, | |
| { | |
| "epoch": 0.9472587033562644, | |
| "grad_norm": 0.11098117381334305, | |
| "learning_rate": 1.5972653422288622e-06, | |
| "loss": 0.5671, | |
| "step": 1136 | |
| }, | |
| { | |
| "epoch": 0.9480925578486554, | |
| "grad_norm": 0.11285511404275894, | |
| "learning_rate": 1.5965445709627669e-06, | |
| "loss": 0.6437, | |
| "step": 1137 | |
| }, | |
| { | |
| "epoch": 0.9489264123410465, | |
| "grad_norm": 0.10554268956184387, | |
| "learning_rate": 1.5958233182464965e-06, | |
| "loss": 0.6289, | |
| "step": 1138 | |
| }, | |
| { | |
| "epoch": 0.9497602668334376, | |
| "grad_norm": 0.10124699026346207, | |
| "learning_rate": 1.5951015846621483e-06, | |
| "loss": 0.6068, | |
| "step": 1139 | |
| }, | |
| { | |
| "epoch": 0.9505941213258287, | |
| "grad_norm": 0.11128734052181244, | |
| "learning_rate": 1.5943793707922084e-06, | |
| "loss": 0.5907, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.9514279758182197, | |
| "grad_norm": 0.11676076054573059, | |
| "learning_rate": 1.5936566772195503e-06, | |
| "loss": 0.602, | |
| "step": 1141 | |
| }, | |
| { | |
| "epoch": 0.9522618303106108, | |
| "grad_norm": 0.11664719134569168, | |
| "learning_rate": 1.5929335045274343e-06, | |
| "loss": 0.6197, | |
| "step": 1142 | |
| }, | |
| { | |
| "epoch": 0.9530956848030019, | |
| "grad_norm": 0.1060841903090477, | |
| "learning_rate": 1.5922098532995083e-06, | |
| "loss": 0.553, | |
| "step": 1143 | |
| }, | |
| { | |
| "epoch": 0.9539295392953929, | |
| "grad_norm": 0.11080587655305862, | |
| "learning_rate": 1.591485724119805e-06, | |
| "loss": 0.6357, | |
| "step": 1144 | |
| }, | |
| { | |
| "epoch": 0.9547633937877841, | |
| "grad_norm": 0.10428017377853394, | |
| "learning_rate": 1.5907611175727442e-06, | |
| "loss": 0.5693, | |
| "step": 1145 | |
| }, | |
| { | |
| "epoch": 0.9555972482801751, | |
| "grad_norm": 0.10886859893798828, | |
| "learning_rate": 1.59003603424313e-06, | |
| "loss": 0.6142, | |
| "step": 1146 | |
| }, | |
| { | |
| "epoch": 0.9564311027725662, | |
| "grad_norm": 0.1091897264122963, | |
| "learning_rate": 1.5893104747161522e-06, | |
| "loss": 0.6234, | |
| "step": 1147 | |
| }, | |
| { | |
| "epoch": 0.9572649572649573, | |
| "grad_norm": 0.11705927550792694, | |
| "learning_rate": 1.5885844395773841e-06, | |
| "loss": 0.6322, | |
| "step": 1148 | |
| }, | |
| { | |
| "epoch": 0.9580988117573483, | |
| "grad_norm": 0.11114629358053207, | |
| "learning_rate": 1.5878579294127831e-06, | |
| "loss": 0.6233, | |
| "step": 1149 | |
| }, | |
| { | |
| "epoch": 0.9589326662497394, | |
| "grad_norm": 0.11006072908639908, | |
| "learning_rate": 1.5871309448086903e-06, | |
| "loss": 0.6271, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.9597665207421305, | |
| "grad_norm": 0.10859903693199158, | |
| "learning_rate": 1.5864034863518292e-06, | |
| "loss": 0.6101, | |
| "step": 1151 | |
| }, | |
| { | |
| "epoch": 0.9606003752345216, | |
| "grad_norm": 0.11398789286613464, | |
| "learning_rate": 1.585675554629306e-06, | |
| "loss": 0.6183, | |
| "step": 1152 | |
| }, | |
| { | |
| "epoch": 0.9614342297269126, | |
| "grad_norm": 0.10914972424507141, | |
| "learning_rate": 1.5849471502286087e-06, | |
| "loss": 0.6687, | |
| "step": 1153 | |
| }, | |
| { | |
| "epoch": 0.9622680842193038, | |
| "grad_norm": 0.11407602578401566, | |
| "learning_rate": 1.5842182737376074e-06, | |
| "loss": 0.6513, | |
| "step": 1154 | |
| }, | |
| { | |
| "epoch": 0.9631019387116948, | |
| "grad_norm": 0.11537064611911774, | |
| "learning_rate": 1.5834889257445525e-06, | |
| "loss": 0.6284, | |
| "step": 1155 | |
| }, | |
| { | |
| "epoch": 0.9639357932040858, | |
| "grad_norm": 0.10912200808525085, | |
| "learning_rate": 1.582759106838075e-06, | |
| "loss": 0.634, | |
| "step": 1156 | |
| }, | |
| { | |
| "epoch": 0.964769647696477, | |
| "grad_norm": 0.11783526092767715, | |
| "learning_rate": 1.5820288176071861e-06, | |
| "loss": 0.6435, | |
| "step": 1157 | |
| }, | |
| { | |
| "epoch": 0.965603502188868, | |
| "grad_norm": 0.11490530520677567, | |
| "learning_rate": 1.5812980586412765e-06, | |
| "loss": 0.5872, | |
| "step": 1158 | |
| }, | |
| { | |
| "epoch": 0.9664373566812591, | |
| "grad_norm": 0.10283337533473969, | |
| "learning_rate": 1.5805668305301167e-06, | |
| "loss": 0.565, | |
| "step": 1159 | |
| }, | |
| { | |
| "epoch": 0.9672712111736502, | |
| "grad_norm": 0.10747389495372772, | |
| "learning_rate": 1.5798351338638548e-06, | |
| "loss": 0.5878, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.9681050656660413, | |
| "grad_norm": 0.11634895950555801, | |
| "learning_rate": 1.5791029692330172e-06, | |
| "loss": 0.6547, | |
| "step": 1161 | |
| }, | |
| { | |
| "epoch": 0.9689389201584323, | |
| "grad_norm": 0.11201413720846176, | |
| "learning_rate": 1.5783703372285086e-06, | |
| "loss": 0.6039, | |
| "step": 1162 | |
| }, | |
| { | |
| "epoch": 0.9697727746508235, | |
| "grad_norm": 0.11117483675479889, | |
| "learning_rate": 1.5776372384416105e-06, | |
| "loss": 0.6447, | |
| "step": 1163 | |
| }, | |
| { | |
| "epoch": 0.9706066291432145, | |
| "grad_norm": 0.11274691671133041, | |
| "learning_rate": 1.5769036734639815e-06, | |
| "loss": 0.5847, | |
| "step": 1164 | |
| }, | |
| { | |
| "epoch": 0.9714404836356055, | |
| "grad_norm": 0.11471108347177505, | |
| "learning_rate": 1.5761696428876556e-06, | |
| "loss": 0.6523, | |
| "step": 1165 | |
| }, | |
| { | |
| "epoch": 0.9722743381279967, | |
| "grad_norm": 0.11089842766523361, | |
| "learning_rate": 1.5754351473050434e-06, | |
| "loss": 0.5949, | |
| "step": 1166 | |
| }, | |
| { | |
| "epoch": 0.9731081926203877, | |
| "grad_norm": 0.10404518991708755, | |
| "learning_rate": 1.5747001873089306e-06, | |
| "loss": 0.5496, | |
| "step": 1167 | |
| }, | |
| { | |
| "epoch": 0.9739420471127789, | |
| "grad_norm": 0.116578109562397, | |
| "learning_rate": 1.5739647634924773e-06, | |
| "loss": 0.6103, | |
| "step": 1168 | |
| }, | |
| { | |
| "epoch": 0.9747759016051699, | |
| "grad_norm": 0.1054786667227745, | |
| "learning_rate": 1.5732288764492184e-06, | |
| "loss": 0.623, | |
| "step": 1169 | |
| }, | |
| { | |
| "epoch": 0.975609756097561, | |
| "grad_norm": 0.11369701474905014, | |
| "learning_rate": 1.5724925267730624e-06, | |
| "loss": 0.567, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.9764436105899521, | |
| "grad_norm": 0.1198234111070633, | |
| "learning_rate": 1.5717557150582917e-06, | |
| "loss": 0.6188, | |
| "step": 1171 | |
| }, | |
| { | |
| "epoch": 0.9772774650823431, | |
| "grad_norm": 0.12100456655025482, | |
| "learning_rate": 1.5710184418995604e-06, | |
| "loss": 0.641, | |
| "step": 1172 | |
| }, | |
| { | |
| "epoch": 0.9781113195747342, | |
| "grad_norm": 0.10783829540014267, | |
| "learning_rate": 1.5702807078918965e-06, | |
| "loss": 0.5889, | |
| "step": 1173 | |
| }, | |
| { | |
| "epoch": 0.9789451740671253, | |
| "grad_norm": 0.11204763501882553, | |
| "learning_rate": 1.5695425136306987e-06, | |
| "loss": 0.6271, | |
| "step": 1174 | |
| }, | |
| { | |
| "epoch": 0.9797790285595164, | |
| "grad_norm": 0.11363273113965988, | |
| "learning_rate": 1.568803859711738e-06, | |
| "loss": 0.6099, | |
| "step": 1175 | |
| }, | |
| { | |
| "epoch": 0.9806128830519074, | |
| "grad_norm": 0.10527540743350983, | |
| "learning_rate": 1.5680647467311555e-06, | |
| "loss": 0.5782, | |
| "step": 1176 | |
| }, | |
| { | |
| "epoch": 0.9814467375442986, | |
| "grad_norm": 0.11106487363576889, | |
| "learning_rate": 1.5673251752854645e-06, | |
| "loss": 0.645, | |
| "step": 1177 | |
| }, | |
| { | |
| "epoch": 0.9822805920366896, | |
| "grad_norm": 0.11001728475093842, | |
| "learning_rate": 1.5665851459715457e-06, | |
| "loss": 0.611, | |
| "step": 1178 | |
| }, | |
| { | |
| "epoch": 0.9831144465290806, | |
| "grad_norm": 0.10988467186689377, | |
| "learning_rate": 1.5658446593866517e-06, | |
| "loss": 0.6229, | |
| "step": 1179 | |
| }, | |
| { | |
| "epoch": 0.9839483010214718, | |
| "grad_norm": 0.10910095274448395, | |
| "learning_rate": 1.565103716128403e-06, | |
| "loss": 0.5964, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.9847821555138628, | |
| "grad_norm": 0.11653497070074081, | |
| "learning_rate": 1.5643623167947891e-06, | |
| "loss": 0.6451, | |
| "step": 1181 | |
| }, | |
| { | |
| "epoch": 0.9856160100062539, | |
| "grad_norm": 0.11314839869737625, | |
| "learning_rate": 1.5636204619841667e-06, | |
| "loss": 0.6419, | |
| "step": 1182 | |
| }, | |
| { | |
| "epoch": 0.986449864498645, | |
| "grad_norm": 0.10827460885047913, | |
| "learning_rate": 1.5628781522952611e-06, | |
| "loss": 0.6129, | |
| "step": 1183 | |
| }, | |
| { | |
| "epoch": 0.9872837189910361, | |
| "grad_norm": 0.1053231805562973, | |
| "learning_rate": 1.5621353883271649e-06, | |
| "loss": 0.6455, | |
| "step": 1184 | |
| }, | |
| { | |
| "epoch": 0.9881175734834271, | |
| "grad_norm": 0.11166463792324066, | |
| "learning_rate": 1.561392170679336e-06, | |
| "loss": 0.594, | |
| "step": 1185 | |
| }, | |
| { | |
| "epoch": 0.9889514279758183, | |
| "grad_norm": 0.11850026249885559, | |
| "learning_rate": 1.5606484999516e-06, | |
| "loss": 0.575, | |
| "step": 1186 | |
| }, | |
| { | |
| "epoch": 0.9897852824682093, | |
| "grad_norm": 0.11536535620689392, | |
| "learning_rate": 1.5599043767441471e-06, | |
| "loss": 0.645, | |
| "step": 1187 | |
| }, | |
| { | |
| "epoch": 0.9906191369606003, | |
| "grad_norm": 0.1191059872508049, | |
| "learning_rate": 1.5591598016575328e-06, | |
| "loss": 0.6007, | |
| "step": 1188 | |
| }, | |
| { | |
| "epoch": 0.9914529914529915, | |
| "grad_norm": 0.1131712794303894, | |
| "learning_rate": 1.558414775292678e-06, | |
| "loss": 0.5926, | |
| "step": 1189 | |
| }, | |
| { | |
| "epoch": 0.9922868459453825, | |
| "grad_norm": 0.11502361297607422, | |
| "learning_rate": 1.5576692982508663e-06, | |
| "loss": 0.6223, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.9931207004377736, | |
| "grad_norm": 0.10861647129058838, | |
| "learning_rate": 1.5569233711337474e-06, | |
| "loss": 0.5948, | |
| "step": 1191 | |
| }, | |
| { | |
| "epoch": 0.9939545549301647, | |
| "grad_norm": 0.11109884828329086, | |
| "learning_rate": 1.5561769945433323e-06, | |
| "loss": 0.5778, | |
| "step": 1192 | |
| }, | |
| { | |
| "epoch": 0.9947884094225558, | |
| "grad_norm": 0.1124456375837326, | |
| "learning_rate": 1.5554301690819951e-06, | |
| "loss": 0.6732, | |
| "step": 1193 | |
| }, | |
| { | |
| "epoch": 0.9956222639149468, | |
| "grad_norm": 0.10982295870780945, | |
| "learning_rate": 1.5546828953524727e-06, | |
| "loss": 0.6221, | |
| "step": 1194 | |
| }, | |
| { | |
| "epoch": 0.9964561184073379, | |
| "grad_norm": 0.10954145342111588, | |
| "learning_rate": 1.553935173957863e-06, | |
| "loss": 0.6045, | |
| "step": 1195 | |
| }, | |
| { | |
| "epoch": 0.997289972899729, | |
| "grad_norm": 0.10329638421535492, | |
| "learning_rate": 1.5531870055016265e-06, | |
| "loss": 0.5827, | |
| "step": 1196 | |
| }, | |
| { | |
| "epoch": 0.99812382739212, | |
| "grad_norm": 0.1147105023264885, | |
| "learning_rate": 1.5524383905875828e-06, | |
| "loss": 0.6676, | |
| "step": 1197 | |
| }, | |
| { | |
| "epoch": 0.9989576818845112, | |
| "grad_norm": 0.10970206558704376, | |
| "learning_rate": 1.5516893298199133e-06, | |
| "loss": 0.6159, | |
| "step": 1198 | |
| }, | |
| { | |
| "epoch": 0.9997915363769022, | |
| "grad_norm": 0.10848627239465714, | |
| "learning_rate": 1.5509398238031585e-06, | |
| "loss": 0.6106, | |
| "step": 1199 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 3597, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 1199, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 5.45611918040105e+18, | |
| "train_batch_size": 6, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |