|
|
{ |
|
|
"best_global_step": 400, |
|
|
"best_metric": 0.16972463, |
|
|
"best_model_checkpoint": "/Qwen/Qwen-VL-master/ckp/checkpoints_IO_3B/v0-20250504-043938/checkpoint-400", |
|
|
"epoch": 0.9996103390050656, |
|
|
"eval_steps": 100, |
|
|
"global_step": 481, |
|
|
"is_hyper_param_search": false, |
|
|
"is_local_process_zero": true, |
|
|
"is_world_process_zero": true, |
|
|
"log_history": [ |
|
|
{ |
|
|
"epoch": 0.002078191972983504, |
|
|
"grad_norm": 49.5, |
|
|
"learning_rate": 2.0000000000000002e-07, |
|
|
"loss": 2.024962902069092, |
|
|
"memory(GiB)": 30.5, |
|
|
"step": 1, |
|
|
"token_acc": 0.5925925925925926, |
|
|
"train_speed(iter/s)": 0.096714 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.010390959864917521, |
|
|
"grad_norm": 48.25, |
|
|
"learning_rate": 1.0000000000000002e-06, |
|
|
"loss": 1.9107239246368408, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 5, |
|
|
"token_acc": 0.6030968247745983, |
|
|
"train_speed(iter/s)": 0.104169 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.020781919729835042, |
|
|
"grad_norm": 46.0, |
|
|
"learning_rate": 2.0000000000000003e-06, |
|
|
"loss": 1.8856124877929688, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 10, |
|
|
"token_acc": 0.6060725241878392, |
|
|
"train_speed(iter/s)": 0.10431 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.031172879594752564, |
|
|
"grad_norm": 41.25, |
|
|
"learning_rate": 3e-06, |
|
|
"loss": 1.6671241760253905, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 15, |
|
|
"token_acc": 0.6276759587247805, |
|
|
"train_speed(iter/s)": 0.104357 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.041563839459670085, |
|
|
"grad_norm": 35.75, |
|
|
"learning_rate": 4.000000000000001e-06, |
|
|
"loss": 1.4283610343933106, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 20, |
|
|
"token_acc": 0.644587791063562, |
|
|
"train_speed(iter/s)": 0.104096 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.05195479932458761, |
|
|
"grad_norm": 24.125, |
|
|
"learning_rate": 5e-06, |
|
|
"loss": 0.8581354141235351, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 25, |
|
|
"token_acc": 0.7361944076641483, |
|
|
"train_speed(iter/s)": 0.103919 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.06234575918950513, |
|
|
"grad_norm": 10.4375, |
|
|
"learning_rate": 4.998516877114918e-06, |
|
|
"loss": 0.3941912889480591, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 30, |
|
|
"token_acc": 0.8509957048028114, |
|
|
"train_speed(iter/s)": 0.103913 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.07273671905442265, |
|
|
"grad_norm": 3.6875, |
|
|
"learning_rate": 4.994069268182461e-06, |
|
|
"loss": 0.24166102409362794, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 35, |
|
|
"token_acc": 0.9107004193197702, |
|
|
"train_speed(iter/s)": 0.103983 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.08312767891934017, |
|
|
"grad_norm": 1.828125, |
|
|
"learning_rate": 4.986662450283107e-06, |
|
|
"loss": 0.2008065700531006, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 40, |
|
|
"token_acc": 0.9205524090309566, |
|
|
"train_speed(iter/s)": 0.103895 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.0935186387842577, |
|
|
"grad_norm": 2.265625, |
|
|
"learning_rate": 4.976305211593758e-06, |
|
|
"loss": 0.19015437364578247, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 45, |
|
|
"token_acc": 0.9180647678501755, |
|
|
"train_speed(iter/s)": 0.103752 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.10390959864917522, |
|
|
"grad_norm": 1.9921875, |
|
|
"learning_rate": 4.963009840960598e-06, |
|
|
"loss": 0.19012608528137206, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 50, |
|
|
"token_acc": 0.9183419043166589, |
|
|
"train_speed(iter/s)": 0.103736 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.11430055851409274, |
|
|
"grad_norm": 1.9921875, |
|
|
"learning_rate": 4.9467921133183864e-06, |
|
|
"loss": 0.1876815915107727, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 55, |
|
|
"token_acc": 0.9174203696421549, |
|
|
"train_speed(iter/s)": 0.103693 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.12469151837901025, |
|
|
"grad_norm": 1.640625, |
|
|
"learning_rate": 4.927671270973533e-06, |
|
|
"loss": 0.18475005626678467, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 60, |
|
|
"token_acc": 0.9194869417400711, |
|
|
"train_speed(iter/s)": 0.103749 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.13508247824392777, |
|
|
"grad_norm": 2.03125, |
|
|
"learning_rate": 4.905670000773126e-06, |
|
|
"loss": 0.18700193166732787, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 65, |
|
|
"token_acc": 0.9183423809159722, |
|
|
"train_speed(iter/s)": 0.103789 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.1454734381088453, |
|
|
"grad_norm": 1.8984375, |
|
|
"learning_rate": 4.880814407187037e-06, |
|
|
"loss": 0.1833273410797119, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 70, |
|
|
"token_acc": 0.9198803956145059, |
|
|
"train_speed(iter/s)": 0.103777 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.15586439797376284, |
|
|
"grad_norm": 1.8046875, |
|
|
"learning_rate": 4.8531339813349995e-06, |
|
|
"loss": 0.17667144536972046, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 75, |
|
|
"token_acc": 0.9221237549996079, |
|
|
"train_speed(iter/s)": 0.101758 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.16625535783868034, |
|
|
"grad_norm": 1.8203125, |
|
|
"learning_rate": 4.822661565995454e-06, |
|
|
"loss": 0.17785824537277223, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 80, |
|
|
"token_acc": 0.9208750290900628, |
|
|
"train_speed(iter/s)": 0.096873 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.17664631770359787, |
|
|
"grad_norm": 1.5390625, |
|
|
"learning_rate": 4.789433316637644e-06, |
|
|
"loss": 0.18061867952346802, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 85, |
|
|
"token_acc": 0.9204729886389984, |
|
|
"train_speed(iter/s)": 0.092995 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.1870372775685154, |
|
|
"grad_norm": 1.8828125, |
|
|
"learning_rate": 4.753488658523212e-06, |
|
|
"loss": 0.17701296806335448, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 90, |
|
|
"token_acc": 0.9225134926754048, |
|
|
"train_speed(iter/s)": 0.089657 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.1974282374334329, |
|
|
"grad_norm": 3.21875, |
|
|
"learning_rate": 4.714870239928195e-06, |
|
|
"loss": 0.1849340558052063, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 95, |
|
|
"token_acc": 0.9169326586220319, |
|
|
"train_speed(iter/s)": 0.086893 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.20781919729835044, |
|
|
"grad_norm": 3.3125, |
|
|
"learning_rate": 4.673623881540917e-06, |
|
|
"loss": 0.1811493992805481, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 100, |
|
|
"token_acc": 0.9194093158432174, |
|
|
"train_speed(iter/s)": 0.084511 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.20781919729835044, |
|
|
"eval_loss": 0.17230889201164246, |
|
|
"eval_runtime": 1320.3184, |
|
|
"eval_samples_per_second": 60.881, |
|
|
"eval_steps_per_second": 3.805, |
|
|
"eval_token_acc": 0.9252880661892655, |
|
|
"step": 100 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.21821015716326797, |
|
|
"grad_norm": 1.9375, |
|
|
"learning_rate": 4.629798522095818e-06, |
|
|
"loss": 0.1755787253379822, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 105, |
|
|
"token_acc": 0.925254316998113, |
|
|
"train_speed(iter/s)": 0.03748 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.22860111702818547, |
|
|
"grad_norm": 1.6484375, |
|
|
"learning_rate": 4.583446160307729e-06, |
|
|
"loss": 0.16902718544006348, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 110, |
|
|
"token_acc": 0.9260815437286025, |
|
|
"train_speed(iter/s)": 0.0381 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.238992076893103, |
|
|
"grad_norm": 3.0, |
|
|
"learning_rate": 4.534621793175488e-06, |
|
|
"loss": 0.17990721464157106, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 115, |
|
|
"token_acc": 0.9196615695102073, |
|
|
"train_speed(iter/s)": 0.038616 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.2493830367580205, |
|
|
"grad_norm": 1.984375, |
|
|
"learning_rate": 4.4833833507280884e-06, |
|
|
"loss": 0.17740323543548583, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 120, |
|
|
"token_acc": 0.9212245219170536, |
|
|
"train_speed(iter/s)": 0.039106 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.25977399662293804, |
|
|
"grad_norm": 2.03125, |
|
|
"learning_rate": 4.429791627290802e-06, |
|
|
"loss": 0.17831547260284425, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 125, |
|
|
"token_acc": 0.9183422870299309, |
|
|
"train_speed(iter/s)": 0.039569 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.27016495648785555, |
|
|
"grad_norm": 1.5546875, |
|
|
"learning_rate": 4.373910209352816e-06, |
|
|
"loss": 0.1763285517692566, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 130, |
|
|
"token_acc": 0.9201423817998917, |
|
|
"train_speed(iter/s)": 0.040384 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.2805559163527731, |
|
|
"grad_norm": 1.625, |
|
|
"learning_rate": 4.315805400121963e-06, |
|
|
"loss": 0.17515565156936647, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 135, |
|
|
"token_acc": 0.9202636948673678, |
|
|
"train_speed(iter/s)": 0.041134 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.2909468762176906, |
|
|
"grad_norm": 2.09375, |
|
|
"learning_rate": 4.255546140856087e-06, |
|
|
"loss": 0.18066967725753785, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 140, |
|
|
"token_acc": 0.9194688203774171, |
|
|
"train_speed(iter/s)": 0.041899 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.3013378360826081, |
|
|
"grad_norm": 1.140625, |
|
|
"learning_rate": 4.1932039290643534e-06, |
|
|
"loss": 0.18014354705810548, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 145, |
|
|
"token_acc": 0.9178993503952414, |
|
|
"train_speed(iter/s)": 0.042363 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.31172879594752567, |
|
|
"grad_norm": 1.3515625, |
|
|
"learning_rate": 4.128852733675572e-06, |
|
|
"loss": 0.17886133193969728, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 150, |
|
|
"token_acc": 0.9210793106137409, |
|
|
"train_speed(iter/s)": 0.042968 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.3221197558124432, |
|
|
"grad_norm": 1.7734375, |
|
|
"learning_rate": 4.062568907274197e-06, |
|
|
"loss": 0.18089871406555175, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 155, |
|
|
"token_acc": 0.9191895486192599, |
|
|
"train_speed(iter/s)": 0.043361 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.3325107156773607, |
|
|
"grad_norm": 1.7421875, |
|
|
"learning_rate": 3.994431095508102e-06, |
|
|
"loss": 0.17732292413711548, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 160, |
|
|
"token_acc": 0.9188111067657411, |
|
|
"train_speed(iter/s)": 0.043692 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.34290167554227824, |
|
|
"grad_norm": 1.78125, |
|
|
"learning_rate": 3.9245201437756655e-06, |
|
|
"loss": 0.1741779327392578, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 165, |
|
|
"token_acc": 0.9216403392732083, |
|
|
"train_speed(iter/s)": 0.043984 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.35329263540719574, |
|
|
"grad_norm": 1.875, |
|
|
"learning_rate": 3.852919001302833e-06, |
|
|
"loss": 0.17619532346725464, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 170, |
|
|
"token_acc": 0.9227966824277187, |
|
|
"train_speed(iter/s)": 0.044279 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.36368359527211325, |
|
|
"grad_norm": 3.71875, |
|
|
"learning_rate": 3.779712622724003e-06, |
|
|
"loss": 0.17437649965286256, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 175, |
|
|
"token_acc": 0.9230951628201075, |
|
|
"train_speed(iter/s)": 0.044551 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.3740745551370308, |
|
|
"grad_norm": 1.84375, |
|
|
"learning_rate": 3.704987867283499e-06, |
|
|
"loss": 0.17724401950836183, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 180, |
|
|
"token_acc": 0.9202978056426332, |
|
|
"train_speed(iter/s)": 0.044808 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.3844655150019483, |
|
|
"grad_norm": 2.1875, |
|
|
"learning_rate": 3.628833395777224e-06, |
|
|
"loss": 0.18116596937179566, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 185, |
|
|
"token_acc": 0.9170886075949367, |
|
|
"train_speed(iter/s)": 0.045056 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.3948564748668658, |
|
|
"grad_norm": 2.046875, |
|
|
"learning_rate": 3.551339565356769e-06, |
|
|
"loss": 0.17593677043914796, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 190, |
|
|
"token_acc": 0.9202586206896551, |
|
|
"train_speed(iter/s)": 0.045325 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.4052474347317834, |
|
|
"grad_norm": 2.0, |
|
|
"learning_rate": 3.4725983223208155e-06, |
|
|
"loss": 0.1823140025138855, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 195, |
|
|
"token_acc": 0.917639683410391, |
|
|
"train_speed(iter/s)": 0.04554 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.4156383945967009, |
|
|
"grad_norm": 1.7109375, |
|
|
"learning_rate": 3.392703093021e-06, |
|
|
"loss": 0.18079570531845093, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 200, |
|
|
"token_acc": 0.9180621543733936, |
|
|
"train_speed(iter/s)": 0.045874 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.4156383945967009, |
|
|
"eval_loss": 0.17064262926578522, |
|
|
"eval_runtime": 1310.4171, |
|
|
"eval_samples_per_second": 61.341, |
|
|
"eval_steps_per_second": 3.834, |
|
|
"eval_token_acc": 0.9238987477359346, |
|
|
"step": 200 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.4260293544616184, |
|
|
"grad_norm": 1.7890625, |
|
|
"learning_rate": 3.3117486730117092e-06, |
|
|
"loss": 0.181607449054718, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 205, |
|
|
"token_acc": 0.9238299582974754, |
|
|
"train_speed(iter/s)": 0.034354 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.43642031432653594, |
|
|
"grad_norm": 2.0625, |
|
|
"learning_rate": 3.229831114575315e-06, |
|
|
"loss": 0.17884130477905275, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 210, |
|
|
"token_acc": 0.9180160352145889, |
|
|
"train_speed(iter/s)": 0.034671 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.44681127419145344, |
|
|
"grad_norm": 1.484375, |
|
|
"learning_rate": 3.147047612756302e-06, |
|
|
"loss": 0.174602210521698, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 215, |
|
|
"token_acc": 0.920204921214003, |
|
|
"train_speed(iter/s)": 0.034976 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.45720223405637095, |
|
|
"grad_norm": 1.234375, |
|
|
"learning_rate": 3.063496390039516e-06, |
|
|
"loss": 0.1783522129058838, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 220, |
|
|
"token_acc": 0.9198404629702042, |
|
|
"train_speed(iter/s)": 0.035377 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.46759319392128845, |
|
|
"grad_norm": 1.78125, |
|
|
"learning_rate": 2.9792765798093466e-06, |
|
|
"loss": 0.179638671875, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 225, |
|
|
"token_acc": 0.9207520783155932, |
|
|
"train_speed(iter/s)": 0.035838 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.477984153786206, |
|
|
"grad_norm": 1.921875, |
|
|
"learning_rate": 2.8944881087281375e-06, |
|
|
"loss": 0.18421536684036255, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 230, |
|
|
"token_acc": 0.916609076249712, |
|
|
"train_speed(iter/s)": 0.036335 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.4883751136511235, |
|
|
"grad_norm": 1.7890625, |
|
|
"learning_rate": 2.80923157817337e-06, |
|
|
"loss": 0.17683053016662598, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 235, |
|
|
"token_acc": 0.9213091922005571, |
|
|
"train_speed(iter/s)": 0.036825 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.498766073516041, |
|
|
"grad_norm": 1.84375, |
|
|
"learning_rate": 2.723608144874298e-06, |
|
|
"loss": 0.17713736295700072, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 240, |
|
|
"token_acc": 0.92035536159601, |
|
|
"train_speed(iter/s)": 0.037143 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.5091570333809585, |
|
|
"grad_norm": 1.5234375, |
|
|
"learning_rate": 2.637719400889664e-06, |
|
|
"loss": 0.17735506296157838, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 245, |
|
|
"token_acc": 0.9191973969631236, |
|
|
"train_speed(iter/s)": 0.037405 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.5195479932458761, |
|
|
"grad_norm": 1.3828125, |
|
|
"learning_rate": 2.5516672530688864e-06, |
|
|
"loss": 0.18297756910324098, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 250, |
|
|
"token_acc": 0.9193918758413177, |
|
|
"train_speed(iter/s)": 0.037652 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.5299389531107936, |
|
|
"grad_norm": 2.203125, |
|
|
"learning_rate": 2.4655538021397592e-06, |
|
|
"loss": 0.17588781118392943, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 255, |
|
|
"token_acc": 0.9216273752266814, |
|
|
"train_speed(iter/s)": 0.037896 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.5403299129757111, |
|
|
"grad_norm": 1.6796875, |
|
|
"learning_rate": 2.3794812215661134e-06, |
|
|
"loss": 0.18080203533172606, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 260, |
|
|
"token_acc": 0.921203216826477, |
|
|
"train_speed(iter/s)": 0.038125 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.5507208728406287, |
|
|
"grad_norm": 2.6875, |
|
|
"learning_rate": 2.2935516363191695e-06, |
|
|
"loss": 0.18104695081710814, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 265, |
|
|
"token_acc": 0.9180170831005029, |
|
|
"train_speed(iter/s)": 0.038349 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.5611118327055462, |
|
|
"grad_norm": 1.7578125, |
|
|
"learning_rate": 2.2078670017064366e-06, |
|
|
"loss": 0.1768229365348816, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 270, |
|
|
"token_acc": 0.9203719621786356, |
|
|
"train_speed(iter/s)": 0.038567 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.5715027925704637, |
|
|
"grad_norm": 1.34375, |
|
|
"learning_rate": 2.1225289824019077e-06, |
|
|
"loss": 0.17726024389266967, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 275, |
|
|
"token_acc": 0.9191293295201779, |
|
|
"train_speed(iter/s)": 0.038795 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.5818937524353812, |
|
|
"grad_norm": 1.71875, |
|
|
"learning_rate": 2.037638831821104e-06, |
|
|
"loss": 0.18150538206100464, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 280, |
|
|
"token_acc": 0.9174561267277528, |
|
|
"train_speed(iter/s)": 0.03901 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.5922847123002988, |
|
|
"grad_norm": 1.546875, |
|
|
"learning_rate": 1.953297271984061e-06, |
|
|
"loss": 0.1750028371810913, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 285, |
|
|
"token_acc": 0.9228211275230067, |
|
|
"train_speed(iter/s)": 0.039295 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.6026756721652162, |
|
|
"grad_norm": 1.234375, |
|
|
"learning_rate": 1.8696043740088236e-06, |
|
|
"loss": 0.17536247968673707, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 290, |
|
|
"token_acc": 0.9208355815925383, |
|
|
"train_speed(iter/s)": 0.039587 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.6130666320301338, |
|
|
"grad_norm": 1.5390625, |
|
|
"learning_rate": 1.7866594393772375e-06, |
|
|
"loss": 0.1767476439476013, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 295, |
|
|
"token_acc": 0.9220417633410672, |
|
|
"train_speed(iter/s)": 0.039803 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.6234575918950513, |
|
|
"grad_norm": 1.9140625, |
|
|
"learning_rate": 1.7045608821139045e-06, |
|
|
"loss": 0.17351619005203248, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 300, |
|
|
"token_acc": 0.9228345246330029, |
|
|
"train_speed(iter/s)": 0.040008 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.6234575918950513, |
|
|
"eval_loss": 0.16996929049491882, |
|
|
"eval_runtime": 1338.6589, |
|
|
"eval_samples_per_second": 60.047, |
|
|
"eval_steps_per_second": 3.753, |
|
|
"eval_token_acc": 0.9252277899263542, |
|
|
"step": 300 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.6338485517599688, |
|
|
"grad_norm": 1.7109375, |
|
|
"learning_rate": 1.6234061120181144e-06, |
|
|
"loss": 0.1729782223701477, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 305, |
|
|
"token_acc": 0.9252146572948351, |
|
|
"train_speed(iter/s)": 0.033422 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.6442395116248864, |
|
|
"grad_norm": 2.296875, |
|
|
"learning_rate": 1.5432914190872757e-06, |
|
|
"loss": 0.1758588194847107, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 310, |
|
|
"token_acc": 0.9202410211686355, |
|
|
"train_speed(iter/s)": 0.033685 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.6546304714898039, |
|
|
"grad_norm": 2.03125, |
|
|
"learning_rate": 1.464311859269003e-06, |
|
|
"loss": 0.17637253999710084, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 315, |
|
|
"token_acc": 0.9202949713658115, |
|
|
"train_speed(iter/s)": 0.034017 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.6650214313547214, |
|
|
"grad_norm": 2.296875, |
|
|
"learning_rate": 1.3865611416773921e-06, |
|
|
"loss": 0.1803775191307068, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 320, |
|
|
"token_acc": 0.9197953561589925, |
|
|
"train_speed(iter/s)": 0.034317 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.6754123912196389, |
|
|
"grad_norm": 2.046875, |
|
|
"learning_rate": 1.3101315174073162e-06, |
|
|
"loss": 0.17441234588623047, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 325, |
|
|
"token_acc": 0.9233814625058221, |
|
|
"train_speed(iter/s)": 0.034642 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.6858033510845565, |
|
|
"grad_norm": 1.640625, |
|
|
"learning_rate": 1.235113670078658e-06, |
|
|
"loss": 0.17859526872634887, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 330, |
|
|
"token_acc": 0.9203665665981988, |
|
|
"train_speed(iter/s)": 0.034906 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.6961943109494739, |
|
|
"grad_norm": 1.375, |
|
|
"learning_rate": 1.161596608240349e-06, |
|
|
"loss": 0.17650686502456664, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 335, |
|
|
"token_acc": 0.9200734956361966, |
|
|
"train_speed(iter/s)": 0.035109 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.7065852708143915, |
|
|
"grad_norm": 1.7734375, |
|
|
"learning_rate": 1.0896675597618725e-06, |
|
|
"loss": 0.18125290870666505, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 340, |
|
|
"token_acc": 0.9172004012655297, |
|
|
"train_speed(iter/s)": 0.035306 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.716976230679309, |
|
|
"grad_norm": 1.046875, |
|
|
"learning_rate": 1.0194118683375502e-06, |
|
|
"loss": 0.1810152769088745, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 345, |
|
|
"token_acc": 0.9158885657431501, |
|
|
"train_speed(iter/s)": 0.035495 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.7273671905442265, |
|
|
"grad_norm": 2.203125, |
|
|
"learning_rate": 9.509128922263886e-07, |
|
|
"loss": 0.17332950830459595, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 350, |
|
|
"token_acc": 0.9222935923546921, |
|
|
"train_speed(iter/s)": 0.035686 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.737758150409144, |
|
|
"grad_norm": 1.28125, |
|
|
"learning_rate": 8.842519053476476e-07, |
|
|
"loss": 0.17291887998580932, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 355, |
|
|
"token_acc": 0.9233901951744276, |
|
|
"train_speed(iter/s)": 0.035869 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.7481491102740616, |
|
|
"grad_norm": 1.5859375, |
|
|
"learning_rate": 8.195080008494744e-07, |
|
|
"loss": 0.17527254819869995, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 360, |
|
|
"token_acc": 0.9194406927782632, |
|
|
"train_speed(iter/s)": 0.03605 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.7585400701389791, |
|
|
"grad_norm": 1.8984375, |
|
|
"learning_rate": 7.567579972650116e-07, |
|
|
"loss": 0.1758326292037964, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 365, |
|
|
"token_acc": 0.9222837796244998, |
|
|
"train_speed(iter/s)": 0.036228 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.7689310300038966, |
|
|
"grad_norm": 1.4453125, |
|
|
"learning_rate": 6.960763473673451e-07, |
|
|
"loss": 0.17296208143234254, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 370, |
|
|
"token_acc": 0.9199661069172701, |
|
|
"train_speed(iter/s)": 0.036459 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.7793219898688142, |
|
|
"grad_norm": 2.15625, |
|
|
"learning_rate": 6.375350498314075e-07, |
|
|
"loss": 0.17531843185424806, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 375, |
|
|
"token_acc": 0.921526620645878, |
|
|
"train_speed(iter/s)": 0.036681 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.7897129497337316, |
|
|
"grad_norm": 1.453125, |
|
|
"learning_rate": 5.812035638076785e-07, |
|
|
"loss": 0.17355836629867555, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 380, |
|
|
"token_acc": 0.9211911683821796, |
|
|
"train_speed(iter/s)": 0.036859 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.8001039095986492, |
|
|
"grad_norm": 2.03125, |
|
|
"learning_rate": 5.271487265090163e-07, |
|
|
"loss": 0.1834742784500122, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 385, |
|
|
"token_acc": 0.9166990442147797, |
|
|
"train_speed(iter/s)": 0.037022 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.8104948694635667, |
|
|
"grad_norm": 1.78125, |
|
|
"learning_rate": 4.754346739084173e-07, |
|
|
"loss": 0.17866382598876954, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 390, |
|
|
"token_acc": 0.9190055507778907, |
|
|
"train_speed(iter/s)": 0.037179 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.8208858293284842, |
|
|
"grad_norm": 1.7265625, |
|
|
"learning_rate": 4.2612276464179673e-07, |
|
|
"loss": 0.16837071180343627, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 395, |
|
|
"token_acc": 0.925252047889099, |
|
|
"train_speed(iter/s)": 0.037337 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.8312767891934018, |
|
|
"grad_norm": 1.625, |
|
|
"learning_rate": 3.7927150720606596e-07, |
|
|
"loss": 0.17911137342453004, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 400, |
|
|
"token_acc": 0.920388196307401, |
|
|
"train_speed(iter/s)": 0.037494 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.8312767891934018, |
|
|
"eval_loss": 0.16972462832927704, |
|
|
"eval_runtime": 1325.8316, |
|
|
"eval_samples_per_second": 60.628, |
|
|
"eval_steps_per_second": 3.789, |
|
|
"eval_token_acc": 0.9239007240068497, |
|
|
"step": 400 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.8416677490583193, |
|
|
"grad_norm": 1.546875, |
|
|
"learning_rate": 3.3493649053890325e-07, |
|
|
"loss": 0.17124630212783815, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 405, |
|
|
"token_acc": 0.9238681922697054, |
|
|
"train_speed(iter/s)": 0.032973 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.8520587089232368, |
|
|
"grad_norm": 2.25, |
|
|
"learning_rate": 2.931703180625736e-07, |
|
|
"loss": 0.17123721837997435, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 410, |
|
|
"token_acc": 0.9216517506672947, |
|
|
"train_speed(iter/s)": 0.033214 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.8624496687881543, |
|
|
"grad_norm": 1.5859375, |
|
|
"learning_rate": 2.5402254527005286e-07, |
|
|
"loss": 0.17746515274047853, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 415, |
|
|
"token_acc": 0.9180823830579474, |
|
|
"train_speed(iter/s)": 0.033458 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.8728406286530719, |
|
|
"grad_norm": 1.3984375, |
|
|
"learning_rate": 2.1753962092752545e-07, |
|
|
"loss": 0.1777236580848694, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 420, |
|
|
"token_acc": 0.9190801349972529, |
|
|
"train_speed(iter/s)": 0.033705 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.8832315885179893, |
|
|
"grad_norm": 1.53125, |
|
|
"learning_rate": 1.837648319629956e-07, |
|
|
"loss": 0.17560485601425171, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 425, |
|
|
"token_acc": 0.9201813013751249, |
|
|
"train_speed(iter/s)": 0.033868 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.8936225483829069, |
|
|
"grad_norm": 1.7578125, |
|
|
"learning_rate": 1.5273825210642608e-07, |
|
|
"loss": 0.17915327548980714, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 430, |
|
|
"token_acc": 0.9182488623882002, |
|
|
"train_speed(iter/s)": 0.034029 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.9040135082478243, |
|
|
"grad_norm": 1.3671875, |
|
|
"learning_rate": 1.2449669434232202e-07, |
|
|
"loss": 0.17606736421585084, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 435, |
|
|
"token_acc": 0.9228688461241209, |
|
|
"train_speed(iter/s)": 0.034183 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.9144044681127419, |
|
|
"grad_norm": 1.546875, |
|
|
"learning_rate": 9.907366723118678e-08, |
|
|
"loss": 0.17585525512695313, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 440, |
|
|
"token_acc": 0.9183866393571766, |
|
|
"train_speed(iter/s)": 0.034333 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.9247954279776595, |
|
|
"grad_norm": 1.5, |
|
|
"learning_rate": 7.649933515167407e-08, |
|
|
"loss": 0.16930280923843383, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 445, |
|
|
"token_acc": 0.9238362821778419, |
|
|
"train_speed(iter/s)": 0.034485 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.9351863878425769, |
|
|
"grad_norm": 2.078125, |
|
|
"learning_rate": 5.6800482510601937e-08, |
|
|
"loss": 0.1705850124359131, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 450, |
|
|
"token_acc": 0.925625, |
|
|
"train_speed(iter/s)": 0.03464 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.9455773477074945, |
|
|
"grad_norm": 1.5703125, |
|
|
"learning_rate": 4.000048196330014e-08, |
|
|
"loss": 0.1817856550216675, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 455, |
|
|
"token_acc": 0.9208294062205467, |
|
|
"train_speed(iter/s)": 0.03485 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.955968307572412, |
|
|
"grad_norm": 1.234375, |
|
|
"learning_rate": 2.611926668199316e-08, |
|
|
"loss": 0.1783198595046997, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 460, |
|
|
"token_acc": 0.9211605415860735, |
|
|
"train_speed(iter/s)": 0.035001 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.9663592674373295, |
|
|
"grad_norm": 1.59375, |
|
|
"learning_rate": 1.517330670512629e-08, |
|
|
"loss": 0.17677730321884155, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 465, |
|
|
"token_acc": 0.9180868609125893, |
|
|
"train_speed(iter/s)": 0.03514 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.976750227302247, |
|
|
"grad_norm": 1.984375, |
|
|
"learning_rate": 7.175589395692351e-09, |
|
|
"loss": 0.17313094139099122, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 470, |
|
|
"token_acc": 0.9224636548382054, |
|
|
"train_speed(iter/s)": 0.03528 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.9871411871671646, |
|
|
"grad_norm": 2.0625, |
|
|
"learning_rate": 2.1356040317474512e-09, |
|
|
"loss": 0.1789810061454773, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 475, |
|
|
"token_acc": 0.918305714728202, |
|
|
"train_speed(iter/s)": 0.035421 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.997532147032082, |
|
|
"grad_norm": 1.296875, |
|
|
"learning_rate": 5.933054739837296e-11, |
|
|
"loss": 0.17718052864074707, |
|
|
"memory(GiB)": 36.99, |
|
|
"step": 480, |
|
|
"token_acc": 0.9196187450357427, |
|
|
"train_speed(iter/s)": 0.03556 |
|
|
}, |
|
|
{ |
|
|
"epoch": 0.9996103390050656, |
|
|
"eval_loss": 0.16983823478221893, |
|
|
"eval_runtime": 1324.7962, |
|
|
"eval_samples_per_second": 60.675, |
|
|
"eval_steps_per_second": 3.792, |
|
|
"eval_token_acc": 0.9239886680625727, |
|
|
"step": 481 |
|
|
} |
|
|
], |
|
|
"logging_steps": 5, |
|
|
"max_steps": 481, |
|
|
"num_input_tokens_seen": 0, |
|
|
"num_train_epochs": 1, |
|
|
"save_steps": 100, |
|
|
"stateful_callbacks": { |
|
|
"TrainerControl": { |
|
|
"args": { |
|
|
"should_epoch_stop": false, |
|
|
"should_evaluate": false, |
|
|
"should_log": false, |
|
|
"should_save": true, |
|
|
"should_training_stop": true |
|
|
}, |
|
|
"attributes": {} |
|
|
} |
|
|
}, |
|
|
"total_flos": 5.1002858766178714e+17, |
|
|
"train_batch_size": 16, |
|
|
"trial_name": null, |
|
|
"trial_params": null |
|
|
} |
|
|
|