{ "best_global_step": 705, "best_metric": 0.16024889051914215, "best_model_checkpoint": "./anime_out/checkpoint-705", "epoch": 5.0, "eval_steps": 500, "global_step": 705, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.07092198581560284, "grad_norm": 1956.7347412109375, "learning_rate": 1.9858156028368796e-05, "loss": 2.5378, "step": 10 }, { "epoch": 0.14184397163120568, "grad_norm": 398.7348327636719, "learning_rate": 1.971631205673759e-05, "loss": 0.7439, "step": 20 }, { "epoch": 0.2127659574468085, "grad_norm": 267.0840759277344, "learning_rate": 1.9574468085106384e-05, "loss": 1.8382, "step": 30 }, { "epoch": 0.28368794326241137, "grad_norm": 123.07373809814453, "learning_rate": 1.9432624113475178e-05, "loss": 2.3966, "step": 40 }, { "epoch": 0.3546099290780142, "grad_norm": 8.433001494267955e-05, "learning_rate": 1.929078014184397e-05, "loss": 1.4528, "step": 50 }, { "epoch": 0.425531914893617, "grad_norm": 310.42840576171875, "learning_rate": 1.914893617021277e-05, "loss": 1.5857, "step": 60 }, { "epoch": 0.49645390070921985, "grad_norm": 14.845270156860352, "learning_rate": 1.9007092198581563e-05, "loss": 0.6746, "step": 70 }, { "epoch": 0.5673758865248227, "grad_norm": 355.8478698730469, "learning_rate": 1.8865248226950357e-05, "loss": 1.2875, "step": 80 }, { "epoch": 0.6382978723404256, "grad_norm": 571.3264770507812, "learning_rate": 1.872340425531915e-05, "loss": 0.4645, "step": 90 }, { "epoch": 0.7092198581560284, "grad_norm": 270.5877380371094, "learning_rate": 1.8581560283687945e-05, "loss": 2.0277, "step": 100 }, { "epoch": 0.7801418439716312, "grad_norm": 14.781210899353027, "learning_rate": 1.843971631205674e-05, "loss": 1.1098, "step": 110 }, { "epoch": 0.851063829787234, "grad_norm": 323.6094665527344, "learning_rate": 1.8297872340425533e-05, "loss": 1.0614, "step": 120 }, { "epoch": 0.9219858156028369, "grad_norm": 469.43707275390625, "learning_rate": 1.8156028368794327e-05, "loss": 0.8355, "step": 130 }, { "epoch": 0.9929078014184397, "grad_norm": 0.007668596692383289, "learning_rate": 1.801418439716312e-05, "loss": 0.5349, "step": 140 }, { "epoch": 1.0, "eval_accuracy": 0.9197860962566845, "eval_loss": 0.36517199873924255, "eval_runtime": 19.748, "eval_samples_per_second": 9.469, "eval_steps_per_second": 1.215, "step": 141 }, { "epoch": 1.0638297872340425, "grad_norm": 1.1018470525741577, "learning_rate": 1.7872340425531915e-05, "loss": 0.4347, "step": 150 }, { "epoch": 1.1347517730496455, "grad_norm": 2.6226277351379395, "learning_rate": 1.773049645390071e-05, "loss": 0.6987, "step": 160 }, { "epoch": 1.2056737588652482, "grad_norm": 0.04894453287124634, "learning_rate": 1.7588652482269506e-05, "loss": 0.7401, "step": 170 }, { "epoch": 1.2765957446808511, "grad_norm": 0.19339050352573395, "learning_rate": 1.74468085106383e-05, "loss": 0.5764, "step": 180 }, { "epoch": 1.3475177304964538, "grad_norm": 176.8428192138672, "learning_rate": 1.7304964539007094e-05, "loss": 0.4433, "step": 190 }, { "epoch": 1.4184397163120568, "grad_norm": 253.8905029296875, "learning_rate": 1.716312056737589e-05, "loss": 0.5627, "step": 200 }, { "epoch": 1.4893617021276595, "grad_norm": 191.34634399414062, "learning_rate": 1.7021276595744682e-05, "loss": 0.757, "step": 210 }, { "epoch": 1.5602836879432624, "grad_norm": 436.973876953125, "learning_rate": 1.6879432624113476e-05, "loss": 1.9875, "step": 220 }, { "epoch": 1.6312056737588652, "grad_norm": 412.5374755859375, "learning_rate": 1.673758865248227e-05, "loss": 0.8893, "step": 230 }, { "epoch": 1.702127659574468, "grad_norm": 153.39613342285156, "learning_rate": 1.6595744680851064e-05, "loss": 1.1219, "step": 240 }, { "epoch": 1.773049645390071, "grad_norm": 303.0024108886719, "learning_rate": 1.645390070921986e-05, "loss": 0.8197, "step": 250 }, { "epoch": 1.8439716312056738, "grad_norm": 4.597931385040283, "learning_rate": 1.6312056737588652e-05, "loss": 0.3096, "step": 260 }, { "epoch": 1.9148936170212765, "grad_norm": 7.166381359100342, "learning_rate": 1.6170212765957446e-05, "loss": 0.304, "step": 270 }, { "epoch": 1.9858156028368794, "grad_norm": 301.7471618652344, "learning_rate": 1.6028368794326244e-05, "loss": 0.6178, "step": 280 }, { "epoch": 2.0, "eval_accuracy": 0.9251336898395722, "eval_loss": 0.28867608308792114, "eval_runtime": 19.4186, "eval_samples_per_second": 9.63, "eval_steps_per_second": 1.236, "step": 282 }, { "epoch": 2.0567375886524824, "grad_norm": 2.6156604290008545, "learning_rate": 1.5886524822695038e-05, "loss": 0.2282, "step": 290 }, { "epoch": 2.127659574468085, "grad_norm": 0.07026529312133789, "learning_rate": 1.5744680851063832e-05, "loss": 0.2304, "step": 300 }, { "epoch": 2.198581560283688, "grad_norm": 0.2403770089149475, "learning_rate": 1.5602836879432626e-05, "loss": 0.1619, "step": 310 }, { "epoch": 2.269503546099291, "grad_norm": 1001.3410034179688, "learning_rate": 1.546099290780142e-05, "loss": 0.4448, "step": 320 }, { "epoch": 2.3404255319148937, "grad_norm": 0.13391649723052979, "learning_rate": 1.5319148936170214e-05, "loss": 0.2728, "step": 330 }, { "epoch": 2.4113475177304964, "grad_norm": 189.5388946533203, "learning_rate": 1.5177304964539008e-05, "loss": 0.7779, "step": 340 }, { "epoch": 2.482269503546099, "grad_norm": 1.5827876040930278e-06, "learning_rate": 1.5035460992907802e-05, "loss": 0.8039, "step": 350 }, { "epoch": 2.5531914893617023, "grad_norm": 127.49388885498047, "learning_rate": 1.4893617021276596e-05, "loss": 0.6087, "step": 360 }, { "epoch": 2.624113475177305, "grad_norm": 0.00041575109935365617, "learning_rate": 1.475177304964539e-05, "loss": 0.468, "step": 370 }, { "epoch": 2.6950354609929077, "grad_norm": 637.6605224609375, "learning_rate": 1.4609929078014187e-05, "loss": 0.4797, "step": 380 }, { "epoch": 2.7659574468085104, "grad_norm": 0.000833332072943449, "learning_rate": 1.4468085106382981e-05, "loss": 0.7188, "step": 390 }, { "epoch": 2.8368794326241136, "grad_norm": 0.004562568850815296, "learning_rate": 1.4326241134751775e-05, "loss": 0.1451, "step": 400 }, { "epoch": 2.9078014184397163, "grad_norm": 360.48321533203125, "learning_rate": 1.418439716312057e-05, "loss": 0.7191, "step": 410 }, { "epoch": 2.978723404255319, "grad_norm": 236.08912658691406, "learning_rate": 1.4042553191489363e-05, "loss": 0.3556, "step": 420 }, { "epoch": 3.0, "eval_accuracy": 0.9197860962566845, "eval_loss": 0.297281414270401, "eval_runtime": 19.4814, "eval_samples_per_second": 9.599, "eval_steps_per_second": 1.232, "step": 423 }, { "epoch": 3.049645390070922, "grad_norm": 222.08946228027344, "learning_rate": 1.3900709219858157e-05, "loss": 0.6342, "step": 430 }, { "epoch": 3.120567375886525, "grad_norm": 0.12243126332759857, "learning_rate": 1.3758865248226951e-05, "loss": 0.9937, "step": 440 }, { "epoch": 3.1914893617021276, "grad_norm": 24.167377471923828, "learning_rate": 1.3617021276595745e-05, "loss": 0.122, "step": 450 }, { "epoch": 3.2624113475177303, "grad_norm": 0.006283226888626814, "learning_rate": 1.347517730496454e-05, "loss": 0.2216, "step": 460 }, { "epoch": 3.3333333333333335, "grad_norm": 41.654052734375, "learning_rate": 1.3333333333333333e-05, "loss": 0.0438, "step": 470 }, { "epoch": 3.404255319148936, "grad_norm": 317.71917724609375, "learning_rate": 1.3191489361702127e-05, "loss": 0.7924, "step": 480 }, { "epoch": 3.475177304964539, "grad_norm": 1.5460463762283325, "learning_rate": 1.3049645390070925e-05, "loss": 0.0088, "step": 490 }, { "epoch": 3.546099290780142, "grad_norm": 121.88572692871094, "learning_rate": 1.2907801418439719e-05, "loss": 0.2176, "step": 500 }, { "epoch": 3.617021276595745, "grad_norm": 0.06358540803194046, "learning_rate": 1.2765957446808513e-05, "loss": 0.1413, "step": 510 }, { "epoch": 3.6879432624113475, "grad_norm": 147.4714813232422, "learning_rate": 1.2624113475177307e-05, "loss": 0.1081, "step": 520 }, { "epoch": 3.7588652482269502, "grad_norm": 0.18462225794792175, "learning_rate": 1.24822695035461e-05, "loss": 0.1447, "step": 530 }, { "epoch": 3.829787234042553, "grad_norm": 0.011159995570778847, "learning_rate": 1.2340425531914895e-05, "loss": 0.3039, "step": 540 }, { "epoch": 3.900709219858156, "grad_norm": 5.591824054718018, "learning_rate": 1.2198581560283689e-05, "loss": 0.3814, "step": 550 }, { "epoch": 3.971631205673759, "grad_norm": 0.0026896377094089985, "learning_rate": 1.2056737588652483e-05, "loss": 0.2332, "step": 560 }, { "epoch": 4.0, "eval_accuracy": 0.9518716577540107, "eval_loss": 0.20226924121379852, "eval_runtime": 19.4299, "eval_samples_per_second": 9.624, "eval_steps_per_second": 1.235, "step": 564 }, { "epoch": 4.042553191489362, "grad_norm": 49.80411911010742, "learning_rate": 1.1914893617021277e-05, "loss": 1.1912, "step": 570 }, { "epoch": 4.113475177304965, "grad_norm": 485.7699279785156, "learning_rate": 1.177304964539007e-05, "loss": 0.5817, "step": 580 }, { "epoch": 4.184397163120567, "grad_norm": 19.611454010009766, "learning_rate": 1.1631205673758865e-05, "loss": 0.1125, "step": 590 }, { "epoch": 4.25531914893617, "grad_norm": 36.56155014038086, "learning_rate": 1.1489361702127662e-05, "loss": 0.6281, "step": 600 }, { "epoch": 4.326241134751773, "grad_norm": 279.4352722167969, "learning_rate": 1.1347517730496456e-05, "loss": 0.9411, "step": 610 }, { "epoch": 4.397163120567376, "grad_norm": 8.183698654174805, "learning_rate": 1.120567375886525e-05, "loss": 0.0053, "step": 620 }, { "epoch": 4.468085106382979, "grad_norm": 0.005647369660437107, "learning_rate": 1.1063829787234044e-05, "loss": 0.6476, "step": 630 }, { "epoch": 4.539007092198582, "grad_norm": 76.17412567138672, "learning_rate": 1.0921985815602838e-05, "loss": 0.2671, "step": 640 }, { "epoch": 4.609929078014185, "grad_norm": 177.41612243652344, "learning_rate": 1.0780141843971632e-05, "loss": 0.3544, "step": 650 }, { "epoch": 4.680851063829787, "grad_norm": 304.32763671875, "learning_rate": 1.0638297872340426e-05, "loss": 0.6, "step": 660 }, { "epoch": 4.75177304964539, "grad_norm": 1.9934466763515957e-05, "learning_rate": 1.049645390070922e-05, "loss": 0.1934, "step": 670 }, { "epoch": 4.822695035460993, "grad_norm": 284.1419677734375, "learning_rate": 1.0354609929078014e-05, "loss": 0.5667, "step": 680 }, { "epoch": 4.8936170212765955, "grad_norm": 12.019274711608887, "learning_rate": 1.0212765957446808e-05, "loss": 0.4064, "step": 690 }, { "epoch": 4.964539007092198, "grad_norm": 0.014889250509440899, "learning_rate": 1.0070921985815602e-05, "loss": 0.4825, "step": 700 }, { "epoch": 5.0, "eval_accuracy": 0.9625668449197861, "eval_loss": 0.16024889051914215, "eval_runtime": 19.4874, "eval_samples_per_second": 9.596, "eval_steps_per_second": 1.232, "step": 705 } ], "logging_steps": 10, "max_steps": 1410, "num_input_tokens_seen": 0, "num_train_epochs": 10, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 3.0811885244459827e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }