| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.0, | |
| "eval_steps": 500, | |
| "global_step": 590, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.033927056827820185, | |
| "grad_norm": 0.5479967594146729, | |
| "learning_rate": 0.0001977324263038549, | |
| "loss": 1.8868, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.06785411365564037, | |
| "grad_norm": 0.224782794713974, | |
| "learning_rate": 0.00019546485260770976, | |
| "loss": 1.2591, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.10178117048346055, | |
| "grad_norm": 0.24415747821331024, | |
| "learning_rate": 0.00019319727891156462, | |
| "loss": 1.0807, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.13570822731128074, | |
| "grad_norm": 0.22518040239810944, | |
| "learning_rate": 0.0001909297052154195, | |
| "loss": 1.0525, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.16963528413910092, | |
| "grad_norm": 0.2071247398853302, | |
| "learning_rate": 0.0001886621315192744, | |
| "loss": 1.0399, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.2035623409669211, | |
| "grad_norm": 0.21717332303524017, | |
| "learning_rate": 0.00018639455782312926, | |
| "loss": 0.9676, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.23748939779474132, | |
| "grad_norm": 0.23291213810443878, | |
| "learning_rate": 0.00018412698412698412, | |
| "loss": 0.9593, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.2714164546225615, | |
| "grad_norm": 0.24413667619228363, | |
| "learning_rate": 0.000181859410430839, | |
| "loss": 0.9712, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.3053435114503817, | |
| "grad_norm": 0.24343140423297882, | |
| "learning_rate": 0.0001795918367346939, | |
| "loss": 0.9398, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.33927056827820185, | |
| "grad_norm": 0.2566365897655487, | |
| "learning_rate": 0.00017732426303854876, | |
| "loss": 0.9026, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.37319762510602206, | |
| "grad_norm": 0.2497139573097229, | |
| "learning_rate": 0.00017505668934240365, | |
| "loss": 0.8716, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.4071246819338422, | |
| "grad_norm": 0.2669306993484497, | |
| "learning_rate": 0.0001727891156462585, | |
| "loss": 0.8351, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.4410517387616624, | |
| "grad_norm": 0.27545222640037537, | |
| "learning_rate": 0.0001705215419501134, | |
| "loss": 0.8784, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.47497879558948264, | |
| "grad_norm": 0.2573504149913788, | |
| "learning_rate": 0.00016825396825396826, | |
| "loss": 0.8346, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.5089058524173028, | |
| "grad_norm": 0.2542634904384613, | |
| "learning_rate": 0.00016598639455782315, | |
| "loss": 0.8113, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.542832909245123, | |
| "grad_norm": 0.314864844083786, | |
| "learning_rate": 0.000163718820861678, | |
| "loss": 0.8093, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.5767599660729432, | |
| "grad_norm": 0.2655967175960541, | |
| "learning_rate": 0.00016145124716553287, | |
| "loss": 0.8426, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.6106870229007634, | |
| "grad_norm": 0.27555230259895325, | |
| "learning_rate": 0.00015918367346938776, | |
| "loss": 0.8129, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.6446140797285835, | |
| "grad_norm": 0.28799182176589966, | |
| "learning_rate": 0.00015691609977324265, | |
| "loss": 0.7833, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.6785411365564037, | |
| "grad_norm": 0.27187833189964294, | |
| "learning_rate": 0.00015464852607709753, | |
| "loss": 0.7795, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.712468193384224, | |
| "grad_norm": 0.28134599328041077, | |
| "learning_rate": 0.00015238095238095237, | |
| "loss": 0.8008, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.7463952502120441, | |
| "grad_norm": 0.29679593443870544, | |
| "learning_rate": 0.00015011337868480726, | |
| "loss": 0.7823, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.7803223070398643, | |
| "grad_norm": 0.30081549286842346, | |
| "learning_rate": 0.00014784580498866215, | |
| "loss": 0.7772, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.8142493638676844, | |
| "grad_norm": 0.28066059947013855, | |
| "learning_rate": 0.000145578231292517, | |
| "loss": 0.7776, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.8481764206955047, | |
| "grad_norm": 0.2732291519641876, | |
| "learning_rate": 0.0001433106575963719, | |
| "loss": 0.7719, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.8821034775233249, | |
| "grad_norm": 0.2929159104824066, | |
| "learning_rate": 0.00014104308390022676, | |
| "loss": 0.778, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.916030534351145, | |
| "grad_norm": 0.29763197898864746, | |
| "learning_rate": 0.00013877551020408165, | |
| "loss": 0.8044, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.9499575911789653, | |
| "grad_norm": 0.28522127866744995, | |
| "learning_rate": 0.0001365079365079365, | |
| "loss": 0.7877, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.9838846480067854, | |
| "grad_norm": 0.3052780032157898, | |
| "learning_rate": 0.0001342403628117914, | |
| "loss": 0.7463, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 1.01696352841391, | |
| "grad_norm": 0.323665976524353, | |
| "learning_rate": 0.00013197278911564626, | |
| "loss": 0.7519, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 1.0508905852417303, | |
| "grad_norm": 0.3049149513244629, | |
| "learning_rate": 0.00012970521541950114, | |
| "loss": 0.7254, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 1.0848176420695506, | |
| "grad_norm": 0.36289119720458984, | |
| "learning_rate": 0.000127437641723356, | |
| "loss": 0.7228, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 1.1187446988973706, | |
| "grad_norm": 0.32274919748306274, | |
| "learning_rate": 0.0001251700680272109, | |
| "loss": 0.7348, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 1.1526717557251909, | |
| "grad_norm": 0.3286229372024536, | |
| "learning_rate": 0.00012290249433106578, | |
| "loss": 0.736, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 1.1865988125530111, | |
| "grad_norm": 0.3304899036884308, | |
| "learning_rate": 0.00012063492063492063, | |
| "loss": 0.7248, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 1.2205258693808312, | |
| "grad_norm": 0.30971524119377136, | |
| "learning_rate": 0.00011836734693877552, | |
| "loss": 0.726, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 1.2544529262086515, | |
| "grad_norm": 0.3118181526660919, | |
| "learning_rate": 0.0001160997732426304, | |
| "loss": 0.715, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 1.2883799830364717, | |
| "grad_norm": 0.35270431637763977, | |
| "learning_rate": 0.00011383219954648527, | |
| "loss": 0.7117, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 1.3223070398642918, | |
| "grad_norm": 0.3350945711135864, | |
| "learning_rate": 0.00011156462585034013, | |
| "loss": 0.7161, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 1.356234096692112, | |
| "grad_norm": 0.3284754455089569, | |
| "learning_rate": 0.000109297052154195, | |
| "loss": 0.7131, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 1.390161153519932, | |
| "grad_norm": 0.31873300671577454, | |
| "learning_rate": 0.0001070294784580499, | |
| "loss": 0.6879, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 1.4240882103477523, | |
| "grad_norm": 0.32634538412094116, | |
| "learning_rate": 0.00010476190476190477, | |
| "loss": 0.7141, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 1.4580152671755724, | |
| "grad_norm": 0.3236243724822998, | |
| "learning_rate": 0.00010249433106575966, | |
| "loss": 0.7128, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 1.4919423240033927, | |
| "grad_norm": 0.34859582781791687, | |
| "learning_rate": 0.0001002267573696145, | |
| "loss": 0.7329, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 1.525869380831213, | |
| "grad_norm": 0.3041052222251892, | |
| "learning_rate": 9.79591836734694e-05, | |
| "loss": 0.7001, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 1.559796437659033, | |
| "grad_norm": 0.33919453620910645, | |
| "learning_rate": 9.569160997732427e-05, | |
| "loss": 0.6973, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 1.5937234944868532, | |
| "grad_norm": 0.333812952041626, | |
| "learning_rate": 9.342403628117914e-05, | |
| "loss": 0.7114, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 1.6276505513146735, | |
| "grad_norm": 0.32383838295936584, | |
| "learning_rate": 9.115646258503402e-05, | |
| "loss": 0.712, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 1.6615776081424936, | |
| "grad_norm": 0.29588553309440613, | |
| "learning_rate": 8.888888888888889e-05, | |
| "loss": 0.6894, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 1.6955046649703138, | |
| "grad_norm": 0.3401544392108917, | |
| "learning_rate": 8.662131519274377e-05, | |
| "loss": 0.7074, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 1.729431721798134, | |
| "grad_norm": 0.3124183714389801, | |
| "learning_rate": 8.435374149659864e-05, | |
| "loss": 0.7054, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 1.7633587786259541, | |
| "grad_norm": 0.31472259759902954, | |
| "learning_rate": 8.208616780045352e-05, | |
| "loss": 0.6761, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 1.7972858354537744, | |
| "grad_norm": 0.31297537684440613, | |
| "learning_rate": 7.981859410430839e-05, | |
| "loss": 0.6715, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 1.8312128922815947, | |
| "grad_norm": 0.33949580788612366, | |
| "learning_rate": 7.755102040816327e-05, | |
| "loss": 0.6883, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 1.8651399491094147, | |
| "grad_norm": 0.34998491406440735, | |
| "learning_rate": 7.528344671201814e-05, | |
| "loss": 0.6878, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 1.899067005937235, | |
| "grad_norm": 0.3444700241088867, | |
| "learning_rate": 7.301587301587302e-05, | |
| "loss": 0.7044, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 1.9329940627650553, | |
| "grad_norm": 0.34009718894958496, | |
| "learning_rate": 7.074829931972789e-05, | |
| "loss": 0.6753, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 1.9669211195928753, | |
| "grad_norm": 0.2923238277435303, | |
| "learning_rate": 6.848072562358277e-05, | |
| "loss": 0.6843, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "grad_norm": 0.43473541736602783, | |
| "learning_rate": 6.621315192743764e-05, | |
| "loss": 0.6944, | |
| "step": 590 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 882, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 8.726134597059871e+17, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |