| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 3.0, | |
| "eval_steps": 500, | |
| "global_step": 678, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.04424778761061947, | |
| "grad_norm": 6.127150945779567, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5902, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.08849557522123894, | |
| "grad_norm": 5.74329347119261, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5179, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.13274336283185842, | |
| "grad_norm": 0.7209844282034923, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4956, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.17699115044247787, | |
| "grad_norm": 0.8838428515444021, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4693, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.22123893805309736, | |
| "grad_norm": 1.047241933895238, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4542, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.26548672566371684, | |
| "grad_norm": 0.6249873423196232, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4507, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.30973451327433627, | |
| "grad_norm": 0.6688340399513245, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4422, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.35398230088495575, | |
| "grad_norm": 0.5938577799260061, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4371, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.39823008849557523, | |
| "grad_norm": 0.39377946979563905, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4288, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.4424778761061947, | |
| "grad_norm": 0.3494222859720882, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4274, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.48672566371681414, | |
| "grad_norm": 0.37585161708391424, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4204, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.5309734513274337, | |
| "grad_norm": 0.41629482517793825, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4204, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.5752212389380531, | |
| "grad_norm": 0.476310089674328, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4205, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.6194690265486725, | |
| "grad_norm": 0.33466258836176344, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4189, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.6637168141592921, | |
| "grad_norm": 0.4017367648814896, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4153, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.7079646017699115, | |
| "grad_norm": 0.4216322560941302, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4127, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.7522123893805309, | |
| "grad_norm": 0.4974612613694754, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4171, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.7964601769911505, | |
| "grad_norm": 0.4002778638113851, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4136, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.8407079646017699, | |
| "grad_norm": 0.38244063424271957, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4141, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.8849557522123894, | |
| "grad_norm": 0.6422183389700243, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4047, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.9292035398230089, | |
| "grad_norm": 0.47781718318215916, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4114, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.9734513274336283, | |
| "grad_norm": 0.458646597184022, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4066, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_loss": 0.40742945671081543, | |
| "eval_runtime": 21.8393, | |
| "eval_samples_per_second": 278.306, | |
| "eval_steps_per_second": 1.099, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 1.0176991150442478, | |
| "grad_norm": 0.40450956449919656, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3954, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 1.0619469026548674, | |
| "grad_norm": 0.33840868352830317, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3815, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 1.1061946902654867, | |
| "grad_norm": 0.42957928697241526, | |
| "learning_rate": 5e-06, | |
| "loss": 0.376, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 1.1504424778761062, | |
| "grad_norm": 0.44727075308000447, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3824, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 1.1946902654867257, | |
| "grad_norm": 0.3832458382206874, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3799, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 1.238938053097345, | |
| "grad_norm": 0.46565842447548633, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3763, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 1.2831858407079646, | |
| "grad_norm": 0.3663739624439434, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3813, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 1.3274336283185841, | |
| "grad_norm": 0.38686016936047624, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3825, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 1.3716814159292037, | |
| "grad_norm": 0.4469971428096629, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3797, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 1.415929203539823, | |
| "grad_norm": 0.3943929709248344, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3771, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 1.4601769911504425, | |
| "grad_norm": 0.5005987203493157, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3797, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 1.504424778761062, | |
| "grad_norm": 0.4398991459645874, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3782, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 1.5486725663716814, | |
| "grad_norm": 0.34410767390147695, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3781, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 1.592920353982301, | |
| "grad_norm": 0.3960889669303223, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3749, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 1.6371681415929205, | |
| "grad_norm": 0.4269177410578348, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3774, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 1.6814159292035398, | |
| "grad_norm": 0.5155261299248002, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3804, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 1.7256637168141593, | |
| "grad_norm": 0.43969372154192843, | |
| "learning_rate": 5e-06, | |
| "loss": 0.378, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 1.7699115044247788, | |
| "grad_norm": 0.4476097491183775, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3755, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 1.8141592920353982, | |
| "grad_norm": 0.38686563229116155, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3801, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 1.8584070796460177, | |
| "grad_norm": 0.35475500773398083, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3737, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 1.9026548672566372, | |
| "grad_norm": 0.3300920550320411, | |
| "learning_rate": 5e-06, | |
| "loss": 0.377, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 1.9469026548672566, | |
| "grad_norm": 0.37302140942747203, | |
| "learning_rate": 5e-06, | |
| "loss": 0.372, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 1.991150442477876, | |
| "grad_norm": 0.3842303488060595, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3734, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_loss": 0.39825692772865295, | |
| "eval_runtime": 21.4661, | |
| "eval_samples_per_second": 283.144, | |
| "eval_steps_per_second": 1.118, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 2.0353982300884956, | |
| "grad_norm": 0.48990695812840507, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3513, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 2.079646017699115, | |
| "grad_norm": 0.45716992449403504, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3424, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 2.1238938053097347, | |
| "grad_norm": 0.4089130956180029, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3436, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 2.168141592920354, | |
| "grad_norm": 0.426162580374922, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3476, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 2.2123893805309733, | |
| "grad_norm": 0.3887768051888474, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3456, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 2.256637168141593, | |
| "grad_norm": 0.49818725927434065, | |
| "learning_rate": 5e-06, | |
| "loss": 0.346, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 2.3008849557522124, | |
| "grad_norm": 0.3869718840156669, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3458, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 2.3451327433628317, | |
| "grad_norm": 0.4420786488564716, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3409, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 2.3893805309734515, | |
| "grad_norm": 0.4328163120927778, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3487, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 2.433628318584071, | |
| "grad_norm": 0.39197907411119853, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3451, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 2.47787610619469, | |
| "grad_norm": 0.36416856007458076, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3503, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 2.52212389380531, | |
| "grad_norm": 0.3746017341827326, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3469, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 2.566371681415929, | |
| "grad_norm": 0.3618786916476107, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3464, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 2.6106194690265485, | |
| "grad_norm": 0.41598805205539596, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3435, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 2.6548672566371683, | |
| "grad_norm": 0.37961766765714905, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3486, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 2.6991150442477876, | |
| "grad_norm": 0.39738095608916224, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3488, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 2.7433628318584073, | |
| "grad_norm": 0.40094307810609703, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3459, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 2.7876106194690267, | |
| "grad_norm": 0.355834042921306, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3415, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 2.831858407079646, | |
| "grad_norm": 0.3745969159488807, | |
| "learning_rate": 5e-06, | |
| "loss": 0.35, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 2.8761061946902657, | |
| "grad_norm": 0.3789920652707106, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3477, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 2.920353982300885, | |
| "grad_norm": 0.3918177591119166, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3477, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 2.9646017699115044, | |
| "grad_norm": 0.33578037413287726, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3466, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_loss": 0.3991487920284271, | |
| "eval_runtime": 21.5542, | |
| "eval_samples_per_second": 281.987, | |
| "eval_steps_per_second": 1.113, | |
| "step": 678 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "step": 678, | |
| "total_flos": 1135675252408320.0, | |
| "train_loss": 0.38758731908151184, | |
| "train_runtime": 4370.1856, | |
| "train_samples_per_second": 79.274, | |
| "train_steps_per_second": 0.155 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 678, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1135675252408320.0, | |
| "train_batch_size": 16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |