| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.5486284289276808, | |
| "eval_steps": 500, | |
| "global_step": 550, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0113314447592068, | |
| "grad_norm": 4.46875, | |
| "learning_rate": 3.6000000000000003e-06, | |
| "loss": 0.5448, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.0226628895184136, | |
| "grad_norm": 0.74609375, | |
| "learning_rate": 7.600000000000001e-06, | |
| "loss": 0.4594, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.0339943342776204, | |
| "grad_norm": 0.380859375, | |
| "learning_rate": 1.16e-05, | |
| "loss": 0.3953, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.0453257790368272, | |
| "grad_norm": 0.31640625, | |
| "learning_rate": 1.5600000000000003e-05, | |
| "loss": 0.3695, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.056657223796033995, | |
| "grad_norm": 0.2314453125, | |
| "learning_rate": 1.9600000000000002e-05, | |
| "loss": 0.3366, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.0679886685552408, | |
| "grad_norm": 0.21484375, | |
| "learning_rate": 1.9998642592088543e-05, | |
| "loss": 0.3204, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.07932011331444759, | |
| "grad_norm": 0.2060546875, | |
| "learning_rate": 1.9993950790937545e-05, | |
| "loss": 0.3231, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.0906515580736544, | |
| "grad_norm": 0.2236328125, | |
| "learning_rate": 1.9985909410557404e-05, | |
| "loss": 0.3127, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.10198300283286119, | |
| "grad_norm": 0.21875, | |
| "learning_rate": 1.9974521146102535e-05, | |
| "loss": 0.3033, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.11331444759206799, | |
| "grad_norm": 0.267578125, | |
| "learning_rate": 1.9959789814471278e-05, | |
| "loss": 0.2907, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.12464589235127478, | |
| "grad_norm": 0.24609375, | |
| "learning_rate": 1.9941720353026582e-05, | |
| "loss": 0.2941, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.1359773371104816, | |
| "grad_norm": 0.2236328125, | |
| "learning_rate": 1.9920318817941234e-05, | |
| "loss": 0.2844, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.14730878186968838, | |
| "grad_norm": 0.23828125, | |
| "learning_rate": 1.9895592382168036e-05, | |
| "loss": 0.2801, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.15864022662889518, | |
| "grad_norm": 0.2294921875, | |
| "learning_rate": 1.986754933303574e-05, | |
| "loss": 0.2805, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.16997167138810199, | |
| "grad_norm": 0.2177734375, | |
| "learning_rate": 1.983619906947144e-05, | |
| "loss": 0.2706, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.1813031161473088, | |
| "grad_norm": 0.240234375, | |
| "learning_rate": 1.980155209885043e-05, | |
| "loss": 0.2756, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.19263456090651557, | |
| "grad_norm": 0.232421875, | |
| "learning_rate": 1.9763620033474552e-05, | |
| "loss": 0.2713, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.20396600566572237, | |
| "grad_norm": 0.2470703125, | |
| "learning_rate": 1.9722415586680204e-05, | |
| "loss": 0.2675, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.21529745042492918, | |
| "grad_norm": 0.236328125, | |
| "learning_rate": 1.9677952568577316e-05, | |
| "loss": 0.2574, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.22662889518413598, | |
| "grad_norm": 0.259765625, | |
| "learning_rate": 1.9630245881420764e-05, | |
| "loss": 0.2636, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.23796033994334279, | |
| "grad_norm": 0.26953125, | |
| "learning_rate": 1.957931151461572e-05, | |
| "loss": 0.2614, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.24929178470254956, | |
| "grad_norm": 0.271484375, | |
| "learning_rate": 1.9525166539358608e-05, | |
| "loss": 0.2548, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.26062322946175637, | |
| "grad_norm": 0.2412109375, | |
| "learning_rate": 1.946782910291554e-05, | |
| "loss": 0.2532, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.2719546742209632, | |
| "grad_norm": 0.271484375, | |
| "learning_rate": 1.9407318422540057e-05, | |
| "loss": 0.2545, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.28328611898017, | |
| "grad_norm": 0.267578125, | |
| "learning_rate": 1.9343654779032244e-05, | |
| "loss": 0.251, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.2593516209476309, | |
| "grad_norm": 0.29296875, | |
| "learning_rate": 1.944186059309318e-05, | |
| "loss": 0.2362, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.26932668329177056, | |
| "grad_norm": 0.31640625, | |
| "learning_rate": 1.93877370638343e-05, | |
| "loss": 0.2377, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.2793017456359102, | |
| "grad_norm": 0.3125, | |
| "learning_rate": 1.9331191872025963e-05, | |
| "loss": 0.2346, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.2892768079800499, | |
| "grad_norm": 0.400390625, | |
| "learning_rate": 1.927223960407727e-05, | |
| "loss": 0.233, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.29925187032418954, | |
| "grad_norm": 0.337890625, | |
| "learning_rate": 1.921089546732717e-05, | |
| "loss": 0.23, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.3092269326683292, | |
| "grad_norm": 0.31640625, | |
| "learning_rate": 1.9147175286121577e-05, | |
| "loss": 0.2345, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.3192019950124688, | |
| "grad_norm": 0.298828125, | |
| "learning_rate": 1.90810954977313e-05, | |
| "loss": 0.2325, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.32917705735660846, | |
| "grad_norm": 0.3359375, | |
| "learning_rate": 1.9012673148111908e-05, | |
| "loss": 0.2322, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.33915211970074816, | |
| "grad_norm": 0.298828125, | |
| "learning_rate": 1.8941925887506527e-05, | |
| "loss": 0.2257, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.3491271820448878, | |
| "grad_norm": 0.31640625, | |
| "learning_rate": 1.8868871965892794e-05, | |
| "loss": 0.2273, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.35910224438902744, | |
| "grad_norm": 0.400390625, | |
| "learning_rate": 1.879353022827509e-05, | |
| "loss": 0.2272, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.3690773067331671, | |
| "grad_norm": 0.3515625, | |
| "learning_rate": 1.8715920109823266e-05, | |
| "loss": 0.2223, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.3790523690773067, | |
| "grad_norm": 0.33203125, | |
| "learning_rate": 1.8636061630859157e-05, | |
| "loss": 0.2175, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.38902743142144636, | |
| "grad_norm": 0.345703125, | |
| "learning_rate": 1.855397539169214e-05, | |
| "loss": 0.2175, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.39900249376558605, | |
| "grad_norm": 0.4140625, | |
| "learning_rate": 1.846968256730504e-05, | |
| "loss": 0.2146, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.4089775561097257, | |
| "grad_norm": 0.330078125, | |
| "learning_rate": 1.838320490189187e-05, | |
| "loss": 0.2127, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.41895261845386533, | |
| "grad_norm": 0.35546875, | |
| "learning_rate": 1.8294564703248668e-05, | |
| "loss": 0.207, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.428927680798005, | |
| "grad_norm": 0.365234375, | |
| "learning_rate": 1.8203784837018985e-05, | |
| "loss": 0.2025, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.4389027431421446, | |
| "grad_norm": 0.390625, | |
| "learning_rate": 1.8110888720795453e-05, | |
| "loss": 0.203, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.4488778054862843, | |
| "grad_norm": 0.365234375, | |
| "learning_rate": 1.8015900318078976e-05, | |
| "loss": 0.2026, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.45885286783042395, | |
| "grad_norm": 0.345703125, | |
| "learning_rate": 1.791884413209711e-05, | |
| "loss": 0.194, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.4688279301745636, | |
| "grad_norm": 0.388671875, | |
| "learning_rate": 1.7819745199483196e-05, | |
| "loss": 0.1938, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.47880299251870323, | |
| "grad_norm": 0.380859375, | |
| "learning_rate": 1.7718629083817914e-05, | |
| "loss": 0.1981, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.48877805486284287, | |
| "grad_norm": 0.330078125, | |
| "learning_rate": 1.7615521869034887e-05, | |
| "loss": 0.1939, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.49875311720698257, | |
| "grad_norm": 0.365234375, | |
| "learning_rate": 1.751045015269207e-05, | |
| "loss": 0.1874, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.5087281795511222, | |
| "grad_norm": 0.38671875, | |
| "learning_rate": 1.740344103911063e-05, | |
| "loss": 0.1947, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.5187032418952618, | |
| "grad_norm": 0.400390625, | |
| "learning_rate": 1.7294522132383137e-05, | |
| "loss": 0.1927, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.5286783042394015, | |
| "grad_norm": 0.40234375, | |
| "learning_rate": 1.7183721529252762e-05, | |
| "loss": 0.1879, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.5386533665835411, | |
| "grad_norm": 0.43359375, | |
| "learning_rate": 1.7071067811865477e-05, | |
| "loss": 0.1899, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.5486284289276808, | |
| "grad_norm": 0.408203125, | |
| "learning_rate": 1.6956590040396975e-05, | |
| "loss": 0.1791, | |
| "step": 550 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 2006, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 2, | |
| "save_steps": 50, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 8.7337702045947e+18, | |
| "train_batch_size": 25, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |