| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.0, | |
| "eval_steps": 500, | |
| "global_step": 396, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.050505050505050504, | |
| "grad_norm": 2.394207715988159, | |
| "learning_rate": 1.9545454545454546e-05, | |
| "loss": 5.1582, | |
| "mean_token_accuracy": 0.2001233294606209, | |
| "num_tokens": 870609.0, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.10101010101010101, | |
| "grad_norm": 1.912199854850769, | |
| "learning_rate": 1.904040404040404e-05, | |
| "loss": 4.9452, | |
| "mean_token_accuracy": 0.2211832121014595, | |
| "num_tokens": 1740025.0, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.15151515151515152, | |
| "grad_norm": 1.7625455856323242, | |
| "learning_rate": 1.8535353535353537e-05, | |
| "loss": 4.793, | |
| "mean_token_accuracy": 0.23438227660953997, | |
| "num_tokens": 2612799.0, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.20202020202020202, | |
| "grad_norm": 1.8361155986785889, | |
| "learning_rate": 1.803030303030303e-05, | |
| "loss": 4.6586, | |
| "mean_token_accuracy": 0.2492609653621912, | |
| "num_tokens": 3481685.0, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.25252525252525254, | |
| "grad_norm": 1.7993879318237305, | |
| "learning_rate": 1.752525252525253e-05, | |
| "loss": 4.5569, | |
| "mean_token_accuracy": 0.2619336560368538, | |
| "num_tokens": 4345672.0, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.30303030303030304, | |
| "grad_norm": 1.399275541305542, | |
| "learning_rate": 1.7020202020202023e-05, | |
| "loss": 4.4388, | |
| "mean_token_accuracy": 0.27335850447416304, | |
| "num_tokens": 5211312.0, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.35353535353535354, | |
| "grad_norm": 1.0864430665969849, | |
| "learning_rate": 1.6515151515151517e-05, | |
| "loss": 4.3731, | |
| "mean_token_accuracy": 0.2805166557431221, | |
| "num_tokens": 6083451.0, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.40404040404040403, | |
| "grad_norm": 1.1086381673812866, | |
| "learning_rate": 1.601010101010101e-05, | |
| "loss": 4.2975, | |
| "mean_token_accuracy": 0.28774467632174494, | |
| "num_tokens": 6945294.0, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.45454545454545453, | |
| "grad_norm": 3.123107671737671, | |
| "learning_rate": 1.5505050505050508e-05, | |
| "loss": 4.2324, | |
| "mean_token_accuracy": 0.2966753713786602, | |
| "num_tokens": 7817939.0, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.5050505050505051, | |
| "grad_norm": 0.9251999855041504, | |
| "learning_rate": 1.5000000000000002e-05, | |
| "loss": 4.1625, | |
| "mean_token_accuracy": 0.3053535230457783, | |
| "num_tokens": 8691417.0, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.5555555555555556, | |
| "grad_norm": 0.9401581287384033, | |
| "learning_rate": 1.4494949494949494e-05, | |
| "loss": 4.1047, | |
| "mean_token_accuracy": 0.31215064153075217, | |
| "num_tokens": 9562161.0, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.6060606060606061, | |
| "grad_norm": 1.956444263458252, | |
| "learning_rate": 1.3989898989898992e-05, | |
| "loss": 4.0634, | |
| "mean_token_accuracy": 0.31729411631822585, | |
| "num_tokens": 10428287.0, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.6565656565656566, | |
| "grad_norm": 1.4297130107879639, | |
| "learning_rate": 1.3484848484848486e-05, | |
| "loss": 4.0134, | |
| "mean_token_accuracy": 0.321382661908865, | |
| "num_tokens": 11297263.0, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.7070707070707071, | |
| "grad_norm": 0.9279031753540039, | |
| "learning_rate": 1.2979797979797981e-05, | |
| "loss": 3.9757, | |
| "mean_token_accuracy": 0.3249970510601997, | |
| "num_tokens": 12166798.0, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.7575757575757576, | |
| "grad_norm": 0.7721190452575684, | |
| "learning_rate": 1.2474747474747475e-05, | |
| "loss": 3.9472, | |
| "mean_token_accuracy": 0.32724879309535027, | |
| "num_tokens": 13039570.0, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.8080808080808081, | |
| "grad_norm": 0.8442203402519226, | |
| "learning_rate": 1.1969696969696971e-05, | |
| "loss": 3.9148, | |
| "mean_token_accuracy": 0.3321701854467392, | |
| "num_tokens": 13914152.0, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.8585858585858586, | |
| "grad_norm": 2.19596266746521, | |
| "learning_rate": 1.1464646464646465e-05, | |
| "loss": 3.8928, | |
| "mean_token_accuracy": 0.3358906783163548, | |
| "num_tokens": 14784924.0, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.9090909090909091, | |
| "grad_norm": 0.7308172583580017, | |
| "learning_rate": 1.0959595959595961e-05, | |
| "loss": 3.8529, | |
| "mean_token_accuracy": 0.33917868658900263, | |
| "num_tokens": 15653435.0, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.9595959595959596, | |
| "grad_norm": 0.6991971135139465, | |
| "learning_rate": 1.0454545454545455e-05, | |
| "loss": 3.847, | |
| "mean_token_accuracy": 0.3425107732415199, | |
| "num_tokens": 16520037.0, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 1.0101010101010102, | |
| "grad_norm": 0.6827145218849182, | |
| "learning_rate": 9.94949494949495e-06, | |
| "loss": 3.832, | |
| "mean_token_accuracy": 0.34413780048489573, | |
| "num_tokens": 17380162.0, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 1.0606060606060606, | |
| "grad_norm": 0.7619866132736206, | |
| "learning_rate": 9.444444444444445e-06, | |
| "loss": 3.7904, | |
| "mean_token_accuracy": 0.3476664200425148, | |
| "num_tokens": 18252280.0, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 1.1111111111111112, | |
| "grad_norm": 0.6569390892982483, | |
| "learning_rate": 8.93939393939394e-06, | |
| "loss": 3.7948, | |
| "mean_token_accuracy": 0.3474963568150997, | |
| "num_tokens": 19118807.0, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 1.1616161616161615, | |
| "grad_norm": 1.1422395706176758, | |
| "learning_rate": 8.434343434343434e-06, | |
| "loss": 3.7545, | |
| "mean_token_accuracy": 0.350558465719223, | |
| "num_tokens": 19987183.0, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 1.2121212121212122, | |
| "grad_norm": 0.9118938446044922, | |
| "learning_rate": 7.92929292929293e-06, | |
| "loss": 3.7356, | |
| "mean_token_accuracy": 0.3532609537243843, | |
| "num_tokens": 20864146.0, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 1.2626262626262625, | |
| "grad_norm": 0.830457329750061, | |
| "learning_rate": 7.424242424242425e-06, | |
| "loss": 3.7153, | |
| "mean_token_accuracy": 0.3551630362868309, | |
| "num_tokens": 21719950.0, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 1.3131313131313131, | |
| "grad_norm": 0.735416054725647, | |
| "learning_rate": 6.91919191919192e-06, | |
| "loss": 3.719, | |
| "mean_token_accuracy": 0.35566510185599326, | |
| "num_tokens": 22598888.0, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 1.3636363636363638, | |
| "grad_norm": 0.7242897152900696, | |
| "learning_rate": 6.4141414141414145e-06, | |
| "loss": 3.699, | |
| "mean_token_accuracy": 0.35784361362457273, | |
| "num_tokens": 23475457.0, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 1.4141414141414141, | |
| "grad_norm": 0.876636803150177, | |
| "learning_rate": 5.90909090909091e-06, | |
| "loss": 3.6957, | |
| "mean_token_accuracy": 0.35799761712551115, | |
| "num_tokens": 24343304.0, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 1.4646464646464645, | |
| "grad_norm": 0.7917715907096863, | |
| "learning_rate": 5.404040404040405e-06, | |
| "loss": 3.6889, | |
| "mean_token_accuracy": 0.3586253359913826, | |
| "num_tokens": 25207489.0, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 1.5151515151515151, | |
| "grad_norm": 0.7085571885108948, | |
| "learning_rate": 4.898989898989899e-06, | |
| "loss": 3.668, | |
| "mean_token_accuracy": 0.36179669201374054, | |
| "num_tokens": 26072804.0, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 1.5656565656565657, | |
| "grad_norm": 0.5749636292457581, | |
| "learning_rate": 4.393939393939394e-06, | |
| "loss": 3.657, | |
| "mean_token_accuracy": 0.3627041429281235, | |
| "num_tokens": 26939192.0, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 1.6161616161616161, | |
| "grad_norm": 0.8387623429298401, | |
| "learning_rate": 3.88888888888889e-06, | |
| "loss": 3.668, | |
| "mean_token_accuracy": 0.3618996299803257, | |
| "num_tokens": 27805084.0, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 1.6666666666666665, | |
| "grad_norm": 0.6559562683105469, | |
| "learning_rate": 3.3838383838383844e-06, | |
| "loss": 3.6492, | |
| "mean_token_accuracy": 0.36368392035365105, | |
| "num_tokens": 28676488.0, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 1.7171717171717171, | |
| "grad_norm": 0.6847353577613831, | |
| "learning_rate": 2.8787878787878793e-06, | |
| "loss": 3.6622, | |
| "mean_token_accuracy": 0.3614747479557991, | |
| "num_tokens": 29545994.0, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 1.7676767676767677, | |
| "grad_norm": 0.8682920336723328, | |
| "learning_rate": 2.373737373737374e-06, | |
| "loss": 3.648, | |
| "mean_token_accuracy": 0.36393540278077124, | |
| "num_tokens": 30421682.0, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 1.8181818181818183, | |
| "grad_norm": 0.7822595834732056, | |
| "learning_rate": 1.868686868686869e-06, | |
| "loss": 3.6468, | |
| "mean_token_accuracy": 0.36377949416637423, | |
| "num_tokens": 31290360.0, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 1.8686868686868687, | |
| "grad_norm": 0.6070267558097839, | |
| "learning_rate": 1.3636363636363636e-06, | |
| "loss": 3.6441, | |
| "mean_token_accuracy": 0.3641804397106171, | |
| "num_tokens": 32154120.0, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 1.9191919191919191, | |
| "grad_norm": 0.7802349328994751, | |
| "learning_rate": 8.585858585858587e-07, | |
| "loss": 3.6413, | |
| "mean_token_accuracy": 0.3650804080069065, | |
| "num_tokens": 33021167.0, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 1.9696969696969697, | |
| "grad_norm": 0.5713841319084167, | |
| "learning_rate": 3.535353535353536e-07, | |
| "loss": 3.6402, | |
| "mean_token_accuracy": 0.3654905065894127, | |
| "num_tokens": 33894053.0, | |
| "step": 390 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 396, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 2, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.6464949686042624e+16, | |
| "train_batch_size": 64, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |