| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.4926905132192845, | |
| "eval_steps": 500, | |
| "global_step": 1000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.024883359253499222, | |
| "grad_norm": 6.788532733917236, | |
| "learning_rate": 4.5e-06, | |
| "loss": 1.2189, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.049766718506998445, | |
| "grad_norm": 1.3378803730010986, | |
| "learning_rate": 9.5e-06, | |
| "loss": 0.6193, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.07465007776049767, | |
| "grad_norm": 0.7917852997779846, | |
| "learning_rate": 1.45e-05, | |
| "loss": 0.231, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.09953343701399689, | |
| "grad_norm": 0.7099640965461731, | |
| "learning_rate": 1.9500000000000003e-05, | |
| "loss": 0.1463, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.12441679626749612, | |
| "grad_norm": 0.686485230922699, | |
| "learning_rate": 2.45e-05, | |
| "loss": 0.108, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.14930015552099535, | |
| "grad_norm": 0.4254772961139679, | |
| "learning_rate": 2.95e-05, | |
| "loss": 0.0878, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.17418351477449456, | |
| "grad_norm": 0.41301316022872925, | |
| "learning_rate": 3.45e-05, | |
| "loss": 0.0733, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.19906687402799378, | |
| "grad_norm": 0.5910139679908752, | |
| "learning_rate": 3.9500000000000005e-05, | |
| "loss": 0.0654, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.223950233281493, | |
| "grad_norm": 0.4281249940395355, | |
| "learning_rate": 4.4500000000000004e-05, | |
| "loss": 0.0621, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.24883359253499224, | |
| "grad_norm": 0.33288511633872986, | |
| "learning_rate": 4.9500000000000004e-05, | |
| "loss": 0.0528, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.2737169517884914, | |
| "grad_norm": 0.3627478778362274, | |
| "learning_rate": 5.45e-05, | |
| "loss": 0.0511, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.2986003110419907, | |
| "grad_norm": 0.3620380461215973, | |
| "learning_rate": 5.95e-05, | |
| "loss": 0.0467, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.3234836702954899, | |
| "grad_norm": 0.37019455432891846, | |
| "learning_rate": 6.450000000000001e-05, | |
| "loss": 0.0444, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.3483670295489891, | |
| "grad_norm": 0.44425082206726074, | |
| "learning_rate": 6.95e-05, | |
| "loss": 0.0413, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.37325038880248834, | |
| "grad_norm": 0.3729994297027588, | |
| "learning_rate": 7.450000000000001e-05, | |
| "loss": 0.0395, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.39813374805598756, | |
| "grad_norm": 0.25299033522605896, | |
| "learning_rate": 7.950000000000001e-05, | |
| "loss": 0.0375, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.4230171073094868, | |
| "grad_norm": 0.24479390680789948, | |
| "learning_rate": 8.450000000000001e-05, | |
| "loss": 0.0336, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.447900466562986, | |
| "grad_norm": 0.4071142375469208, | |
| "learning_rate": 8.950000000000001e-05, | |
| "loss": 0.0316, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.4727838258164852, | |
| "grad_norm": 0.40290141105651855, | |
| "learning_rate": 9.449999999999999e-05, | |
| "loss": 0.0331, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.4976671850699845, | |
| "grad_norm": 0.38158151507377625, | |
| "learning_rate": 9.95e-05, | |
| "loss": 0.0311, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.5225505443234837, | |
| "grad_norm": 0.29570913314819336, | |
| "learning_rate": 9.999861593790126e-05, | |
| "loss": 0.0291, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.5474339035769828, | |
| "grad_norm": 0.37690913677215576, | |
| "learning_rate": 9.999383162408304e-05, | |
| "loss": 0.0277, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.5723172628304821, | |
| "grad_norm": 0.3893811106681824, | |
| "learning_rate": 9.998563029828259e-05, | |
| "loss": 0.028, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.5972006220839814, | |
| "grad_norm": 0.3241683542728424, | |
| "learning_rate": 9.997401252104962e-05, | |
| "loss": 0.0267, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.6220839813374806, | |
| "grad_norm": 0.31424349546432495, | |
| "learning_rate": 9.995897908644378e-05, | |
| "loss": 0.0268, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.6469673405909798, | |
| "grad_norm": 0.2265535593032837, | |
| "learning_rate": 9.994053102198034e-05, | |
| "loss": 0.0257, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.671850699844479, | |
| "grad_norm": 0.352764368057251, | |
| "learning_rate": 9.991866958856003e-05, | |
| "loss": 0.0244, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.6967340590979783, | |
| "grad_norm": 0.29578715562820435, | |
| "learning_rate": 9.989339628038276e-05, | |
| "loss": 0.0224, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.7216174183514774, | |
| "grad_norm": 0.36025872826576233, | |
| "learning_rate": 9.98647128248456e-05, | |
| "loss": 0.0231, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.7465007776049767, | |
| "grad_norm": 0.31710195541381836, | |
| "learning_rate": 9.98326211824246e-05, | |
| "loss": 0.0247, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.7713841368584758, | |
| "grad_norm": 0.28732505440711975, | |
| "learning_rate": 9.979712354654091e-05, | |
| "loss": 0.0224, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.7962674961119751, | |
| "grad_norm": 0.21972423791885376, | |
| "learning_rate": 9.975822234341079e-05, | |
| "loss": 0.0211, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.8211508553654744, | |
| "grad_norm": 0.22749976813793182, | |
| "learning_rate": 9.97159202318798e-05, | |
| "loss": 0.0201, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.8460342146189735, | |
| "grad_norm": 0.3257753551006317, | |
| "learning_rate": 9.967022010324105e-05, | |
| "loss": 0.0218, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.8709175738724728, | |
| "grad_norm": 0.2250833660364151, | |
| "learning_rate": 9.962112508103765e-05, | |
| "loss": 0.0192, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.895800933125972, | |
| "grad_norm": 0.3466697335243225, | |
| "learning_rate": 9.956863852084914e-05, | |
| "loss": 0.0194, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.9206842923794712, | |
| "grad_norm": 0.3127228319644928, | |
| "learning_rate": 9.951276401006221e-05, | |
| "loss": 0.0207, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.9455676516329704, | |
| "grad_norm": 0.2664047181606293, | |
| "learning_rate": 9.945350536762543e-05, | |
| "loss": 0.0194, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.9704510108864697, | |
| "grad_norm": 0.2534600794315338, | |
| "learning_rate": 9.939086664378829e-05, | |
| "loss": 0.0203, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.995334370139969, | |
| "grad_norm": 0.2339852750301361, | |
| "learning_rate": 9.932485211982437e-05, | |
| "loss": 0.0193, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 1.0223950233281494, | |
| "grad_norm": 0.2680792212486267, | |
| "learning_rate": 9.92554663077387e-05, | |
| "loss": 0.0188, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 1.0472783825816485, | |
| "grad_norm": 0.22340402007102966, | |
| "learning_rate": 9.918271394995935e-05, | |
| "loss": 0.0175, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 1.0721617418351477, | |
| "grad_norm": 0.22573307156562805, | |
| "learning_rate": 9.910660001901335e-05, | |
| "loss": 0.0191, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 1.097045101088647, | |
| "grad_norm": 0.16532698273658752, | |
| "learning_rate": 9.902712971718675e-05, | |
| "loss": 0.0189, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 1.1219284603421462, | |
| "grad_norm": 0.26129794120788574, | |
| "learning_rate": 9.894430847616915e-05, | |
| "loss": 0.0191, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 1.1468118195956454, | |
| "grad_norm": 0.2564263939857483, | |
| "learning_rate": 9.885814195668232e-05, | |
| "loss": 0.0184, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 1.1716951788491445, | |
| "grad_norm": 0.269545316696167, | |
| "learning_rate": 9.876863604809344e-05, | |
| "loss": 0.0202, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 1.196578538102644, | |
| "grad_norm": 0.25393763184547424, | |
| "learning_rate": 9.867579686801245e-05, | |
| "loss": 0.0193, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 1.221461897356143, | |
| "grad_norm": 0.26436880230903625, | |
| "learning_rate": 9.8579630761874e-05, | |
| "loss": 0.0184, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 1.2463452566096422, | |
| "grad_norm": 0.1819346696138382, | |
| "learning_rate": 9.848014430250367e-05, | |
| "loss": 0.0182, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 1.2712286158631416, | |
| "grad_norm": 0.26912328600883484, | |
| "learning_rate": 9.837734428966885e-05, | |
| "loss": 0.0177, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 1.2961119751166408, | |
| "grad_norm": 0.32153385877609253, | |
| "learning_rate": 9.827123774961383e-05, | |
| "loss": 0.0181, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 1.32099533437014, | |
| "grad_norm": 0.32203471660614014, | |
| "learning_rate": 9.816183193457968e-05, | |
| "loss": 0.0182, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 1.3458786936236393, | |
| "grad_norm": 0.34882357716560364, | |
| "learning_rate": 9.804913432230856e-05, | |
| "loss": 0.0182, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 1.3707620528771385, | |
| "grad_norm": 0.3095707297325134, | |
| "learning_rate": 9.793315261553252e-05, | |
| "loss": 0.0178, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 1.3956454121306376, | |
| "grad_norm": 0.31017982959747314, | |
| "learning_rate": 9.781389474144717e-05, | |
| "loss": 0.0182, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 1.4205287713841368, | |
| "grad_norm": 0.24658547341823578, | |
| "learning_rate": 9.76913688511698e-05, | |
| "loss": 0.0174, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 1.445412130637636, | |
| "grad_norm": 0.23288820683956146, | |
| "learning_rate": 9.756558331918227e-05, | |
| "loss": 0.0172, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 1.4702954898911353, | |
| "grad_norm": 0.16752177476882935, | |
| "learning_rate": 9.743654674275855e-05, | |
| "loss": 0.0185, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 1.4951788491446345, | |
| "grad_norm": 0.19120067358016968, | |
| "learning_rate": 9.730426794137727e-05, | |
| "loss": 0.0163, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 1.5200622083981337, | |
| "grad_norm": 0.2136230617761612, | |
| "learning_rate": 9.716875595611879e-05, | |
| "loss": 0.0171, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 1.544945567651633, | |
| "grad_norm": 0.18993347883224487, | |
| "learning_rate": 9.703002004904729e-05, | |
| "loss": 0.0164, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 1.5698289269051322, | |
| "grad_norm": 0.22324898838996887, | |
| "learning_rate": 9.688806970257773e-05, | |
| "loss": 0.0169, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 1.5947122861586314, | |
| "grad_norm": 0.18617911636829376, | |
| "learning_rate": 9.674291461882774e-05, | |
| "loss": 0.0153, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 1.6195956454121307, | |
| "grad_norm": 0.1991218626499176, | |
| "learning_rate": 9.659456471895445e-05, | |
| "loss": 0.0158, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 1.64447900466563, | |
| "grad_norm": 0.27028852701187134, | |
| "learning_rate": 9.644303014247648e-05, | |
| "loss": 0.0164, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 1.669362363919129, | |
| "grad_norm": 0.21468698978424072, | |
| "learning_rate": 9.628832124658085e-05, | |
| "loss": 0.0159, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 1.6942457231726284, | |
| "grad_norm": 0.25870266556739807, | |
| "learning_rate": 9.613044860541507e-05, | |
| "loss": 0.0159, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 1.7191290824261274, | |
| "grad_norm": 0.2562198042869568, | |
| "learning_rate": 9.596942300936445e-05, | |
| "loss": 0.0149, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 1.7440124416796268, | |
| "grad_norm": 0.22703538835048676, | |
| "learning_rate": 9.580525546431459e-05, | |
| "loss": 0.0165, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 1.768895800933126, | |
| "grad_norm": 0.27203115820884705, | |
| "learning_rate": 9.563795719089911e-05, | |
| "loss": 0.0153, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 1.793779160186625, | |
| "grad_norm": 0.337146133184433, | |
| "learning_rate": 9.546753962373281e-05, | |
| "loss": 0.0175, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 1.8186625194401245, | |
| "grad_norm": 0.3714126944541931, | |
| "learning_rate": 9.529401441062997e-05, | |
| "loss": 0.0162, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 1.8435458786936236, | |
| "grad_norm": 0.23839302361011505, | |
| "learning_rate": 9.511739341180842e-05, | |
| "loss": 0.0169, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 1.8684292379471228, | |
| "grad_norm": 0.2162984013557434, | |
| "learning_rate": 9.493768869907886e-05, | |
| "loss": 0.0153, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 1.8933125972006222, | |
| "grad_norm": 0.25327548384666443, | |
| "learning_rate": 9.475491255501968e-05, | |
| "loss": 0.0149, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 1.9181959564541213, | |
| "grad_norm": 0.17730680108070374, | |
| "learning_rate": 9.456907747213748e-05, | |
| "loss": 0.0147, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 1.9430793157076205, | |
| "grad_norm": 0.2287522703409195, | |
| "learning_rate": 9.438019615201336e-05, | |
| "loss": 0.0154, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 1.9679626749611199, | |
| "grad_norm": 0.2877958118915558, | |
| "learning_rate": 9.418828150443469e-05, | |
| "loss": 0.0157, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 1.9928460342146188, | |
| "grad_norm": 0.272636741399765, | |
| "learning_rate": 9.399334664651262e-05, | |
| "loss": 0.0153, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 2.0199066874027993, | |
| "grad_norm": 0.3082719147205353, | |
| "learning_rate": 9.379540490178581e-05, | |
| "loss": 0.0151, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 2.0447900466562987, | |
| "grad_norm": 0.3081252872943878, | |
| "learning_rate": 9.359446979930955e-05, | |
| "loss": 0.0156, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 2.0696734059097976, | |
| "grad_norm": 0.25696584582328796, | |
| "learning_rate": 9.33905550727312e-05, | |
| "loss": 0.0157, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 2.094556765163297, | |
| "grad_norm": 0.2680334150791168, | |
| "learning_rate": 9.318367465935142e-05, | |
| "loss": 0.0151, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 2.1194401244167964, | |
| "grad_norm": 0.20804259181022644, | |
| "learning_rate": 9.29738426991717e-05, | |
| "loss": 0.0144, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 2.1443234836702953, | |
| "grad_norm": 0.2176770567893982, | |
| "learning_rate": 9.276107353392774e-05, | |
| "loss": 0.0144, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 2.1692068429237947, | |
| "grad_norm": 0.21425634622573853, | |
| "learning_rate": 9.254538170610938e-05, | |
| "loss": 0.0145, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 2.194090202177294, | |
| "grad_norm": 0.20511746406555176, | |
| "learning_rate": 9.232678195796654e-05, | |
| "loss": 0.0146, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 2.218973561430793, | |
| "grad_norm": 0.2533751130104065, | |
| "learning_rate": 9.210528923050164e-05, | |
| "loss": 0.0134, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 2.2438569206842924, | |
| "grad_norm": 0.2530650794506073, | |
| "learning_rate": 9.188091866244834e-05, | |
| "loss": 0.0143, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 2.2687402799377914, | |
| "grad_norm": 0.25711414217948914, | |
| "learning_rate": 9.165368558923695e-05, | |
| "loss": 0.0139, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 2.2936236391912908, | |
| "grad_norm": 0.226607546210289, | |
| "learning_rate": 9.142360554194618e-05, | |
| "loss": 0.0146, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 2.31850699844479, | |
| "grad_norm": 0.15505652129650116, | |
| "learning_rate": 9.119069424624163e-05, | |
| "loss": 0.014, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 2.343390357698289, | |
| "grad_norm": 0.20164167881011963, | |
| "learning_rate": 9.0954967621301e-05, | |
| "loss": 0.0142, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 2.3682737169517885, | |
| "grad_norm": 0.2123536318540573, | |
| "learning_rate": 9.071644177872594e-05, | |
| "loss": 0.0136, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 2.393157076205288, | |
| "grad_norm": 0.218880295753479, | |
| "learning_rate": 9.047513302144095e-05, | |
| "loss": 0.0133, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 2.4180404354587868, | |
| "grad_norm": 0.20033881068229675, | |
| "learning_rate": 9.023105784257906e-05, | |
| "loss": 0.013, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 2.442923794712286, | |
| "grad_norm": 0.20657172799110413, | |
| "learning_rate": 8.998423292435454e-05, | |
| "loss": 0.0146, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 2.4678071539657855, | |
| "grad_norm": 0.14253973960876465, | |
| "learning_rate": 8.973467513692265e-05, | |
| "loss": 0.0133, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 2.4926905132192845, | |
| "grad_norm": 0.20540325343608856, | |
| "learning_rate": 8.94824015372267e-05, | |
| "loss": 0.0155, | |
| "step": 1000 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 4000, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 10, | |
| "save_steps": 1000, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 0.0, | |
| "train_batch_size": 32, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |