paperstack_document_data_retrieval
/
summarizer
/tmp_multilingual_results
/checkpoint-3000
/trainer_state.json
| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 3.278688524590164, | |
| "eval_steps": 500, | |
| "global_step": 3000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0546448087431694, | |
| "grad_norm": 2.1467645168304443, | |
| "learning_rate": 0.0004973224043715847, | |
| "loss": 5.4078, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.1092896174863388, | |
| "grad_norm": 0.8721267580986023, | |
| "learning_rate": 0.0004945901639344262, | |
| "loss": 1.2654, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.16393442622950818, | |
| "grad_norm": 0.8103071451187134, | |
| "learning_rate": 0.0004918579234972678, | |
| "loss": 1.0538, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.2185792349726776, | |
| "grad_norm": 1.3929318189620972, | |
| "learning_rate": 0.0004891256830601093, | |
| "loss": 1.0236, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.273224043715847, | |
| "grad_norm": 0.970960259437561, | |
| "learning_rate": 0.00048639344262295083, | |
| "loss": 1.075, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.32786885245901637, | |
| "grad_norm": 0.9727946519851685, | |
| "learning_rate": 0.00048366120218579234, | |
| "loss": 1.0434, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.3825136612021858, | |
| "grad_norm": 0.6411604285240173, | |
| "learning_rate": 0.0004809289617486339, | |
| "loss": 0.9285, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.4371584699453552, | |
| "grad_norm": 0.8864787817001343, | |
| "learning_rate": 0.0004781967213114754, | |
| "loss": 0.9315, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.4918032786885246, | |
| "grad_norm": 0.8886504173278809, | |
| "learning_rate": 0.0004754644808743169, | |
| "loss": 0.8818, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.546448087431694, | |
| "grad_norm": 0.7289233803749084, | |
| "learning_rate": 0.0004727322404371585, | |
| "loss": 0.92, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.6010928961748634, | |
| "grad_norm": 1.0301029682159424, | |
| "learning_rate": 0.00047, | |
| "loss": 0.8222, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.6557377049180327, | |
| "grad_norm": 0.808475136756897, | |
| "learning_rate": 0.00046726775956284155, | |
| "loss": 0.8767, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.7103825136612022, | |
| "grad_norm": 0.8232861161231995, | |
| "learning_rate": 0.0004645355191256831, | |
| "loss": 0.835, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.7650273224043715, | |
| "grad_norm": 0.9552908539772034, | |
| "learning_rate": 0.0004618032786885246, | |
| "loss": 0.8203, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.819672131147541, | |
| "grad_norm": 0.7888686656951904, | |
| "learning_rate": 0.00045907103825136613, | |
| "loss": 0.8296, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.8743169398907104, | |
| "grad_norm": 0.7473599910736084, | |
| "learning_rate": 0.0004563387978142077, | |
| "loss": 0.7801, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.9289617486338798, | |
| "grad_norm": 0.8543986678123474, | |
| "learning_rate": 0.0004536065573770492, | |
| "loss": 0.9248, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.9836065573770492, | |
| "grad_norm": 0.7910193204879761, | |
| "learning_rate": 0.0004508743169398907, | |
| "loss": 0.8386, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_loss": 0.6989373564720154, | |
| "eval_runtime": 35.1236, | |
| "eval_samples_per_second": 23.175, | |
| "eval_steps_per_second": 2.904, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 1.0382513661202186, | |
| "grad_norm": 1.1451815366744995, | |
| "learning_rate": 0.00044814207650273227, | |
| "loss": 0.7198, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 1.092896174863388, | |
| "grad_norm": 0.7096536755561829, | |
| "learning_rate": 0.0004454098360655738, | |
| "loss": 0.7374, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 1.1475409836065573, | |
| "grad_norm": 0.8436596393585205, | |
| "learning_rate": 0.0004426775956284153, | |
| "loss": 0.7471, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 1.2021857923497268, | |
| "grad_norm": 0.8844044208526611, | |
| "learning_rate": 0.00043994535519125685, | |
| "loss": 0.7072, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 1.2568306010928962, | |
| "grad_norm": 0.9547187089920044, | |
| "learning_rate": 0.00043721311475409836, | |
| "loss": 0.7618, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 1.3114754098360657, | |
| "grad_norm": 0.8470160961151123, | |
| "learning_rate": 0.00043448087431693987, | |
| "loss": 0.7719, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 1.366120218579235, | |
| "grad_norm": 0.8797179460525513, | |
| "learning_rate": 0.00043174863387978143, | |
| "loss": 0.7172, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 1.4207650273224044, | |
| "grad_norm": 0.6831291317939758, | |
| "learning_rate": 0.00042901639344262294, | |
| "loss": 0.7254, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 1.4754098360655736, | |
| "grad_norm": 0.7458399534225464, | |
| "learning_rate": 0.00042628415300546445, | |
| "loss": 0.7219, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 1.530054644808743, | |
| "grad_norm": 0.7829724550247192, | |
| "learning_rate": 0.000423551912568306, | |
| "loss": 0.6969, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 1.5846994535519126, | |
| "grad_norm": 0.914277970790863, | |
| "learning_rate": 0.0004208196721311475, | |
| "loss": 0.7088, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 1.639344262295082, | |
| "grad_norm": 0.9189252853393555, | |
| "learning_rate": 0.00041808743169398913, | |
| "loss": 0.7574, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 1.6939890710382515, | |
| "grad_norm": 0.8233757019042969, | |
| "learning_rate": 0.00041535519125683064, | |
| "loss": 0.7248, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 1.748633879781421, | |
| "grad_norm": 0.5294966101646423, | |
| "learning_rate": 0.00041262295081967215, | |
| "loss": 0.7173, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 1.8032786885245902, | |
| "grad_norm": 0.7432788014411926, | |
| "learning_rate": 0.0004098907103825137, | |
| "loss": 0.731, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 1.8579234972677594, | |
| "grad_norm": 0.9326347708702087, | |
| "learning_rate": 0.0004071584699453552, | |
| "loss": 0.6559, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 1.9125683060109289, | |
| "grad_norm": 0.6357129216194153, | |
| "learning_rate": 0.00040442622950819673, | |
| "loss": 0.7079, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 1.9672131147540983, | |
| "grad_norm": 0.6454396843910217, | |
| "learning_rate": 0.0004016939890710383, | |
| "loss": 0.6284, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_loss": 0.6172024607658386, | |
| "eval_runtime": 34.4665, | |
| "eval_samples_per_second": 23.617, | |
| "eval_steps_per_second": 2.959, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 2.021857923497268, | |
| "grad_norm": 0.9404008388519287, | |
| "learning_rate": 0.0003989617486338798, | |
| "loss": 0.6996, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 2.0765027322404372, | |
| "grad_norm": 0.7585016489028931, | |
| "learning_rate": 0.0003962295081967213, | |
| "loss": 0.6634, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 2.1311475409836067, | |
| "grad_norm": 0.9319397807121277, | |
| "learning_rate": 0.00039349726775956287, | |
| "loss": 0.6439, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 2.185792349726776, | |
| "grad_norm": 0.7812663912773132, | |
| "learning_rate": 0.0003907650273224044, | |
| "loss": 0.6807, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 2.240437158469945, | |
| "grad_norm": 0.8016160726547241, | |
| "learning_rate": 0.0003880327868852459, | |
| "loss": 0.6121, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 2.2950819672131146, | |
| "grad_norm": 0.5846936106681824, | |
| "learning_rate": 0.00038530054644808745, | |
| "loss": 0.6501, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 2.349726775956284, | |
| "grad_norm": 0.6993207931518555, | |
| "learning_rate": 0.00038256830601092896, | |
| "loss": 0.6242, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 2.4043715846994536, | |
| "grad_norm": 0.5866222381591797, | |
| "learning_rate": 0.00037983606557377047, | |
| "loss": 0.5766, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 2.459016393442623, | |
| "grad_norm": 0.8375122547149658, | |
| "learning_rate": 0.00037710382513661203, | |
| "loss": 0.6395, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 2.5136612021857925, | |
| "grad_norm": 0.9567583799362183, | |
| "learning_rate": 0.00037437158469945354, | |
| "loss": 0.6829, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 2.5683060109289615, | |
| "grad_norm": 0.829088032245636, | |
| "learning_rate": 0.00037163934426229505, | |
| "loss": 0.6138, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 2.6229508196721314, | |
| "grad_norm": 0.7738655805587769, | |
| "learning_rate": 0.00036890710382513666, | |
| "loss": 0.5961, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 2.6775956284153004, | |
| "grad_norm": 0.6849051117897034, | |
| "learning_rate": 0.00036617486338797817, | |
| "loss": 0.6158, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 2.73224043715847, | |
| "grad_norm": 0.6353682279586792, | |
| "learning_rate": 0.0003634426229508197, | |
| "loss": 0.6077, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 2.7868852459016393, | |
| "grad_norm": 0.6507243514060974, | |
| "learning_rate": 0.00036071038251366124, | |
| "loss": 0.5881, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 2.841530054644809, | |
| "grad_norm": 0.7680765390396118, | |
| "learning_rate": 0.00035797814207650275, | |
| "loss": 0.64, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 2.8961748633879782, | |
| "grad_norm": 0.8768549561500549, | |
| "learning_rate": 0.00035524590163934426, | |
| "loss": 0.5946, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 2.9508196721311473, | |
| "grad_norm": 0.9345018267631531, | |
| "learning_rate": 0.0003525136612021858, | |
| "loss": 0.5986, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_loss": 0.5772798657417297, | |
| "eval_runtime": 34.1537, | |
| "eval_samples_per_second": 23.833, | |
| "eval_steps_per_second": 2.986, | |
| "step": 2745 | |
| }, | |
| { | |
| "epoch": 3.0054644808743167, | |
| "grad_norm": 0.8109590411186218, | |
| "learning_rate": 0.00034978142076502733, | |
| "loss": 0.6056, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 3.060109289617486, | |
| "grad_norm": 1.1651564836502075, | |
| "learning_rate": 0.00034704918032786884, | |
| "loss": 0.5321, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 3.1147540983606556, | |
| "grad_norm": 0.9508790969848633, | |
| "learning_rate": 0.0003443169398907104, | |
| "loss": 0.5405, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 3.169398907103825, | |
| "grad_norm": 0.7765358686447144, | |
| "learning_rate": 0.0003415846994535519, | |
| "loss": 0.6053, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 3.2240437158469946, | |
| "grad_norm": 1.0186572074890137, | |
| "learning_rate": 0.0003388524590163934, | |
| "loss": 0.6171, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 3.278688524590164, | |
| "grad_norm": 0.9198819398880005, | |
| "learning_rate": 0.000336120218579235, | |
| "loss": 0.5755, | |
| "step": 3000 | |
| } | |
| ], | |
| "logging_steps": 50, | |
| "max_steps": 9150, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 10, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 4461372112896000.0, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |