| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 5.0, | |
| "eval_steps": 20, | |
| "global_step": 695, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.07194244604316546, | |
| "grad_norm": 0.7139045596122742, | |
| "learning_rate": 0.0002, | |
| "loss": 2.0811, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.14388489208633093, | |
| "grad_norm": 0.5728554129600525, | |
| "learning_rate": 0.0002, | |
| "loss": 1.567, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.2158273381294964, | |
| "grad_norm": 0.28863629698753357, | |
| "learning_rate": 0.0002, | |
| "loss": 1.357, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.28776978417266186, | |
| "grad_norm": 0.3229255974292755, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2521, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.3597122302158273, | |
| "grad_norm": 0.33988726139068604, | |
| "learning_rate": 0.0002, | |
| "loss": 1.2164, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.4316546762589928, | |
| "grad_norm": 0.40238669514656067, | |
| "learning_rate": 0.0002, | |
| "loss": 1.1385, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.5035971223021583, | |
| "grad_norm": 0.38314664363861084, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0973, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.5755395683453237, | |
| "grad_norm": 0.3884238302707672, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0836, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.6474820143884892, | |
| "grad_norm": 0.3774580657482147, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0409, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.7194244604316546, | |
| "grad_norm": 0.43280744552612305, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0462, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.7913669064748201, | |
| "grad_norm": 0.3758571445941925, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0206, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.8633093525179856, | |
| "grad_norm": 0.3923165798187256, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0219, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.935251798561151, | |
| "grad_norm": 0.4213840663433075, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9971, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_loss": 0.9775540828704834, | |
| "eval_runtime": 21.5909, | |
| "eval_samples_per_second": 4.771, | |
| "eval_steps_per_second": 0.602, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 1.0071942446043165, | |
| "grad_norm": 0.40076276659965515, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9885, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 1.079136690647482, | |
| "grad_norm": 0.3987070918083191, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9555, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 1.1510791366906474, | |
| "grad_norm": 0.41527315974235535, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9391, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 1.223021582733813, | |
| "grad_norm": 0.42107248306274414, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9332, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 1.2949640287769784, | |
| "grad_norm": 0.4587080776691437, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9163, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 1.3669064748201438, | |
| "grad_norm": 0.3977225124835968, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9142, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 1.4388489208633093, | |
| "grad_norm": 0.41111239790916443, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9088, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 1.5107913669064748, | |
| "grad_norm": 0.4326966404914856, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9082, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 1.5827338129496402, | |
| "grad_norm": 0.3831544816493988, | |
| "learning_rate": 0.0002, | |
| "loss": 0.908, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 1.6546762589928057, | |
| "grad_norm": 0.39992555975914, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9076, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 1.7266187050359711, | |
| "grad_norm": 0.39961057901382446, | |
| "learning_rate": 0.0002, | |
| "loss": 0.9059, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 1.7985611510791366, | |
| "grad_norm": 0.3854500949382782, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8903, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 1.870503597122302, | |
| "grad_norm": 0.4092749357223511, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8904, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 1.9424460431654675, | |
| "grad_norm": 0.40900877118110657, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8679, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_loss": 0.8940790891647339, | |
| "eval_runtime": 21.5978, | |
| "eval_samples_per_second": 4.769, | |
| "eval_steps_per_second": 0.602, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 2.014388489208633, | |
| "grad_norm": 0.3746669888496399, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8669, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 2.0863309352517985, | |
| "grad_norm": 0.4724111258983612, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8304, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 2.158273381294964, | |
| "grad_norm": 0.39437365531921387, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8318, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 2.2302158273381294, | |
| "grad_norm": 0.4238971471786499, | |
| "learning_rate": 0.0002, | |
| "loss": 0.829, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 2.302158273381295, | |
| "grad_norm": 0.37740206718444824, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8235, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 2.3741007194244603, | |
| "grad_norm": 0.40223780274391174, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8328, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 2.446043165467626, | |
| "grad_norm": 0.4160473048686981, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8252, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 2.5179856115107913, | |
| "grad_norm": 0.4427769184112549, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8215, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 2.5899280575539567, | |
| "grad_norm": 0.47273996472358704, | |
| "learning_rate": 0.0002, | |
| "loss": 0.819, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 2.661870503597122, | |
| "grad_norm": 0.38981807231903076, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8225, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 2.7338129496402876, | |
| "grad_norm": 0.385079562664032, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8045, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 2.805755395683453, | |
| "grad_norm": 0.3796544075012207, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8234, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 2.8776978417266186, | |
| "grad_norm": 0.3638385236263275, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8195, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 2.949640287769784, | |
| "grad_norm": 0.37602856755256653, | |
| "learning_rate": 0.0002, | |
| "loss": 0.831, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_loss": 0.8596920967102051, | |
| "eval_runtime": 21.5962, | |
| "eval_samples_per_second": 4.769, | |
| "eval_steps_per_second": 0.602, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 3.0215827338129495, | |
| "grad_norm": 0.3858413100242615, | |
| "learning_rate": 0.0002, | |
| "loss": 0.8076, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 3.093525179856115, | |
| "grad_norm": 0.39952272176742554, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7777, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 3.1654676258992804, | |
| "grad_norm": 0.42352095246315, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7704, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 3.237410071942446, | |
| "grad_norm": 0.4132436513900757, | |
| "learning_rate": 0.0002, | |
| "loss": 0.768, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 3.3093525179856114, | |
| "grad_norm": 0.414110392332077, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7631, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 3.381294964028777, | |
| "grad_norm": 0.43551069498062134, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7673, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 3.4532374100719423, | |
| "grad_norm": 0.4042975604534149, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7552, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 3.5251798561151078, | |
| "grad_norm": 0.4289880394935608, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7559, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 3.597122302158273, | |
| "grad_norm": 0.39530348777770996, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7661, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 3.6690647482014387, | |
| "grad_norm": 0.4326631724834442, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7605, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 3.741007194244604, | |
| "grad_norm": 0.3833424150943756, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7549, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 3.81294964028777, | |
| "grad_norm": 0.40698808431625366, | |
| "learning_rate": 0.0002, | |
| "loss": 0.764, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 3.884892086330935, | |
| "grad_norm": 0.39018431305885315, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7653, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 3.956834532374101, | |
| "grad_norm": 0.37253016233444214, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7677, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "eval_loss": 0.8444326519966125, | |
| "eval_runtime": 21.5712, | |
| "eval_samples_per_second": 4.775, | |
| "eval_steps_per_second": 0.603, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 4.028776978417266, | |
| "grad_norm": 0.4029178023338318, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7467, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 4.100719424460432, | |
| "grad_norm": 0.382028192281723, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7116, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 4.172661870503597, | |
| "grad_norm": 0.42502614855766296, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7298, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 4.244604316546763, | |
| "grad_norm": 0.4241486191749573, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7073, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 4.316546762589928, | |
| "grad_norm": 0.4571862816810608, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7106, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 4.388489208633094, | |
| "grad_norm": 0.43121734261512756, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7031, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 4.460431654676259, | |
| "grad_norm": 0.40107443928718567, | |
| "learning_rate": 0.0002, | |
| "loss": 0.693, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 4.532374100719425, | |
| "grad_norm": 0.4040583670139313, | |
| "learning_rate": 0.0002, | |
| "loss": 0.707, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 4.60431654676259, | |
| "grad_norm": 0.4380245506763458, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7066, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 4.676258992805756, | |
| "grad_norm": 0.4263726472854614, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7216, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 4.748201438848921, | |
| "grad_norm": 0.4532300531864166, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7337, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 4.820143884892087, | |
| "grad_norm": 0.4051191210746765, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7133, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 4.892086330935252, | |
| "grad_norm": 0.434962660074234, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7144, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 4.9640287769784175, | |
| "grad_norm": 0.4131017029285431, | |
| "learning_rate": 0.0002, | |
| "loss": 0.7167, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "eval_loss": 0.8359549641609192, | |
| "eval_runtime": 21.598, | |
| "eval_samples_per_second": 4.769, | |
| "eval_steps_per_second": 0.602, | |
| "step": 695 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 695, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 5, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.8287701530771456e+17, | |
| "train_batch_size": 3, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |