| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 9.91616766467066, | |
| "eval_steps": 500, | |
| "global_step": 1110, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.08982035928143713, | |
| "grad_norm": 49.68221664428711, | |
| "learning_rate": 1.4414414414414416e-06, | |
| "loss": 8.7467, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.17964071856287425, | |
| "grad_norm": 22.61308479309082, | |
| "learning_rate": 3.063063063063063e-06, | |
| "loss": 5.5972, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.2694610778443114, | |
| "grad_norm": 11.650931358337402, | |
| "learning_rate": 4.864864864864866e-06, | |
| "loss": 4.1239, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.3592814371257485, | |
| "grad_norm": 11.040465354919434, | |
| "learning_rate": 6.666666666666667e-06, | |
| "loss": 3.3473, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.4491017964071856, | |
| "grad_norm": 21.185564041137695, | |
| "learning_rate": 8.46846846846847e-06, | |
| "loss": 2.843, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.5389221556886228, | |
| "grad_norm": 6.222358226776123, | |
| "learning_rate": 1.027027027027027e-05, | |
| "loss": 2.5158, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.6287425149700598, | |
| "grad_norm": 12.189018249511719, | |
| "learning_rate": 1.2072072072072074e-05, | |
| "loss": 2.1371, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.718562874251497, | |
| "grad_norm": 17.56637191772461, | |
| "learning_rate": 1.3873873873873875e-05, | |
| "loss": 1.9438, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.8083832335329342, | |
| "grad_norm": 5.583638668060303, | |
| "learning_rate": 1.5675675675675676e-05, | |
| "loss": 1.7556, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.8982035928143712, | |
| "grad_norm": 3.642561197280884, | |
| "learning_rate": 1.7477477477477477e-05, | |
| "loss": 1.6005, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.9880239520958084, | |
| "grad_norm": 2.80586314201355, | |
| "learning_rate": 1.927927927927928e-05, | |
| "loss": 1.5588, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 1.0718562874251496, | |
| "grad_norm": 2.6551783084869385, | |
| "learning_rate": 1.9998219965624736e-05, | |
| "loss": 1.4294, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 1.1616766467065869, | |
| "grad_norm": 2.895826578140259, | |
| "learning_rate": 1.9987344272588007e-05, | |
| "loss": 1.4837, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 1.251497005988024, | |
| "grad_norm": 2.8669326305389404, | |
| "learning_rate": 1.9966592535953532e-05, | |
| "loss": 1.453, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 1.341317365269461, | |
| "grad_norm": 2.1909446716308594, | |
| "learning_rate": 1.9935985276197033e-05, | |
| "loss": 1.3837, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 1.4311377245508983, | |
| "grad_norm": 3.2299187183380127, | |
| "learning_rate": 1.989555275948572e-05, | |
| "loss": 1.3152, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 1.5209580838323353, | |
| "grad_norm": 1.9108384847640991, | |
| "learning_rate": 1.984533496774942e-05, | |
| "loss": 1.1961, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 1.6107784431137726, | |
| "grad_norm": 1.1921658515930176, | |
| "learning_rate": 1.9785381559144196e-05, | |
| "loss": 1.1209, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 1.7005988023952097, | |
| "grad_norm": 0.8914986848831177, | |
| "learning_rate": 1.9715751818947603e-05, | |
| "loss": 1.1056, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 1.7904191616766467, | |
| "grad_norm": 0.9408266544342041, | |
| "learning_rate": 1.963651460093409e-05, | |
| "loss": 1.0827, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 1.8802395209580838, | |
| "grad_norm": 0.8248458504676819, | |
| "learning_rate": 1.9547748259288536e-05, | |
| "loss": 1.0787, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 1.9700598802395208, | |
| "grad_norm": 0.9938483238220215, | |
| "learning_rate": 1.9449540571125284e-05, | |
| "loss": 1.0597, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 2.053892215568862, | |
| "grad_norm": 0.9142336249351501, | |
| "learning_rate": 1.93419886496892e-05, | |
| "loss": 0.9906, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 2.143712574850299, | |
| "grad_norm": 0.8745118379592896, | |
| "learning_rate": 1.9225198848324687e-05, | |
| "loss": 1.05, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 2.2335329341317367, | |
| "grad_norm": 0.7089764475822449, | |
| "learning_rate": 1.909928665530757e-05, | |
| "loss": 1.043, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 2.3233532934131738, | |
| "grad_norm": 0.4864867031574249, | |
| "learning_rate": 1.896437657964382e-05, | |
| "loss": 1.0463, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 2.413173652694611, | |
| "grad_norm": 0.971626341342926, | |
| "learning_rate": 1.8820602027948112e-05, | |
| "loss": 1.0418, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 2.502994011976048, | |
| "grad_norm": 0.699500322341919, | |
| "learning_rate": 1.866810517252393e-05, | |
| "loss": 1.038, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 2.592814371257485, | |
| "grad_norm": 0.9161490797996521, | |
| "learning_rate": 1.8507036810775617e-05, | |
| "loss": 1.0338, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 2.682634730538922, | |
| "grad_norm": 0.6857994198799133, | |
| "learning_rate": 1.833755621609152e-05, | |
| "loss": 1.0231, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 2.772455089820359, | |
| "grad_norm": 0.7041919827461243, | |
| "learning_rate": 1.815983098034555e-05, | |
| "loss": 1.0218, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 2.8622754491017965, | |
| "grad_norm": 0.9268773794174194, | |
| "learning_rate": 1.7974036848172992e-05, | |
| "loss": 1.0229, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 2.9520958083832336, | |
| "grad_norm": 0.7882820963859558, | |
| "learning_rate": 1.7780357543184396e-05, | |
| "loss": 1.0199, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 3.035928143712575, | |
| "grad_norm": 0.7048642635345459, | |
| "learning_rate": 1.757898458628941e-05, | |
| "loss": 0.9551, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 3.125748502994012, | |
| "grad_norm": 0.5380260348320007, | |
| "learning_rate": 1.7370117106310216e-05, | |
| "loss": 1.0078, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 3.215568862275449, | |
| "grad_norm": 0.9133287668228149, | |
| "learning_rate": 1.715396164307182e-05, | |
| "loss": 1.0026, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 3.305389221556886, | |
| "grad_norm": 0.8086357712745667, | |
| "learning_rate": 1.6930731943163975e-05, | |
| "loss": 1.0082, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 3.3952095808383236, | |
| "grad_norm": 0.7032626271247864, | |
| "learning_rate": 1.6700648748576577e-05, | |
| "loss": 1.0025, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 3.4850299401197606, | |
| "grad_norm": 0.7610228657722473, | |
| "learning_rate": 1.6463939578417695e-05, | |
| "loss": 0.9943, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 3.5748502994011977, | |
| "grad_norm": 0.7735270261764526, | |
| "learning_rate": 1.622083850392996e-05, | |
| "loss": 1.0014, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 3.6646706586826348, | |
| "grad_norm": 0.5382483005523682, | |
| "learning_rate": 1.5971585917027864e-05, | |
| "loss": 0.9961, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 3.754491017964072, | |
| "grad_norm": 0.6216614246368408, | |
| "learning_rate": 1.5716428292584788e-05, | |
| "loss": 0.992, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 3.844311377245509, | |
| "grad_norm": 0.5743616819381714, | |
| "learning_rate": 1.545561794470492e-05, | |
| "loss": 0.9861, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 3.934131736526946, | |
| "grad_norm": 0.7634334564208984, | |
| "learning_rate": 1.518941277722096e-05, | |
| "loss": 0.9881, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 4.017964071856287, | |
| "grad_norm": 0.7453029751777649, | |
| "learning_rate": 1.491807602866442e-05, | |
| "loss": 0.9172, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 4.107784431137724, | |
| "grad_norm": 1.0462350845336914, | |
| "learning_rate": 1.4641876011960661e-05, | |
| "loss": 0.9798, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 4.197604790419161, | |
| "grad_norm": 0.8113179206848145, | |
| "learning_rate": 1.436108584910611e-05, | |
| "loss": 0.9799, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 4.287425149700598, | |
| "grad_norm": 0.6361674070358276, | |
| "learning_rate": 1.4075983201089964e-05, | |
| "loss": 0.9654, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 4.3772455089820355, | |
| "grad_norm": 0.6291260123252869, | |
| "learning_rate": 1.3786849993327503e-05, | |
| "loss": 0.9631, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 4.467065868263473, | |
| "grad_norm": 0.988298773765564, | |
| "learning_rate": 1.349397213687651e-05, | |
| "loss": 0.9729, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 4.5568862275449105, | |
| "grad_norm": 1.0734843015670776, | |
| "learning_rate": 1.3197639245712454e-05, | |
| "loss": 0.963, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 4.6467065868263475, | |
| "grad_norm": 0.834682047367096, | |
| "learning_rate": 1.2898144350342015e-05, | |
| "loss": 0.9538, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 4.736526946107785, | |
| "grad_norm": 0.5939741134643555, | |
| "learning_rate": 1.2595783608038157e-05, | |
| "loss": 0.9623, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 4.826347305389222, | |
| "grad_norm": 0.8625423908233643, | |
| "learning_rate": 1.22908560099833e-05, | |
| "loss": 0.9557, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 4.916167664670659, | |
| "grad_norm": 0.8925888538360596, | |
| "learning_rate": 1.198366308561013e-05, | |
| "loss": 0.9419, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "grad_norm": 0.3238130807876587, | |
| "learning_rate": 1.1674508604432464e-05, | |
| "loss": 0.8782, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 5.089820359281437, | |
| "grad_norm": 0.5171638131141663, | |
| "learning_rate": 1.1363698275661002e-05, | |
| "loss": 0.9256, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 5.179640718562874, | |
| "grad_norm": 1.136551022529602, | |
| "learning_rate": 1.1051539445900982e-05, | |
| "loss": 0.9282, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 5.269461077844311, | |
| "grad_norm": 0.9350560307502747, | |
| "learning_rate": 1.0738340795230722e-05, | |
| "loss": 0.9316, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 5.359281437125748, | |
| "grad_norm": 1.0158970355987549, | |
| "learning_rate": 1.0424412031961485e-05, | |
| "loss": 0.9381, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 5.449101796407185, | |
| "grad_norm": 1.0037592649459839, | |
| "learning_rate": 1.0110063586380647e-05, | |
| "loss": 0.9288, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 5.538922155688622, | |
| "grad_norm": 0.6334489583969116, | |
| "learning_rate": 9.795606303780885e-06, | |
| "loss": 0.9272, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 5.62874251497006, | |
| "grad_norm": 0.7240671515464783, | |
| "learning_rate": 9.48135113707899e-06, | |
| "loss": 0.9318, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 5.718562874251497, | |
| "grad_norm": 0.9315741658210754, | |
| "learning_rate": 9.167608839328273e-06, | |
| "loss": 0.9297, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 5.808383233532934, | |
| "grad_norm": 0.7542333602905273, | |
| "learning_rate": 8.854689656428591e-06, | |
| "loss": 0.9119, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 5.8982035928143715, | |
| "grad_norm": 0.864987313747406, | |
| "learning_rate": 8.542903020337887e-06, | |
| "loss": 0.919, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 5.9880239520958085, | |
| "grad_norm": 0.5868083834648132, | |
| "learning_rate": 8.232557243088585e-06, | |
| "loss": 0.9136, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 6.07185628742515, | |
| "grad_norm": 0.8045146465301514, | |
| "learning_rate": 7.923959211911448e-06, | |
| "loss": 0.8452, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 6.161676646706587, | |
| "grad_norm": 0.9145955443382263, | |
| "learning_rate": 7.617414085768352e-06, | |
| "loss": 0.9, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 6.251497005988024, | |
| "grad_norm": 0.6515654921531677, | |
| "learning_rate": 7.313224993594057e-06, | |
| "loss": 0.8946, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 6.341317365269461, | |
| "grad_norm": 0.6843001246452332, | |
| "learning_rate": 7.011692734545403e-06, | |
| "loss": 0.8994, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 6.431137724550898, | |
| "grad_norm": 0.8299969434738159, | |
| "learning_rate": 6.713115480554313e-06, | |
| "loss": 0.894, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 6.520958083832335, | |
| "grad_norm": 1.121025800704956, | |
| "learning_rate": 6.4177884814787284e-06, | |
| "loss": 0.8982, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 6.610778443113772, | |
| "grad_norm": 1.1040045022964478, | |
| "learning_rate": 6.126003773143072e-06, | |
| "loss": 0.8985, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 6.700598802395209, | |
| "grad_norm": 0.711075484752655, | |
| "learning_rate": 5.8380498885569246e-06, | |
| "loss": 0.8923, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 6.790419161676647, | |
| "grad_norm": 0.8444594740867615, | |
| "learning_rate": 5.554211572597477e-06, | |
| "loss": 0.8989, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 6.880239520958084, | |
| "grad_norm": 0.9748265743255615, | |
| "learning_rate": 5.274769500437882e-06, | |
| "loss": 0.8956, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 6.970059880239521, | |
| "grad_norm": 0.8513533473014832, | |
| "learning_rate": 5.000000000000003e-06, | |
| "loss": 0.8906, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 7.053892215568863, | |
| "grad_norm": 1.3123410940170288, | |
| "learning_rate": 4.730174778705909e-06, | |
| "loss": 0.8249, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 7.1437125748503, | |
| "grad_norm": 1.0771753787994385, | |
| "learning_rate": 4.4655606547984165e-06, | |
| "loss": 0.8777, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 7.233532934131737, | |
| "grad_norm": 0.8778141736984253, | |
| "learning_rate": 4.206419293496333e-06, | |
| "loss": 0.8832, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 7.323353293413174, | |
| "grad_norm": 1.0552406311035156, | |
| "learning_rate": 3.953006948245247e-06, | |
| "loss": 0.872, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 7.413173652694611, | |
| "grad_norm": 0.6986867785453796, | |
| "learning_rate": 3.705574207319844e-06, | |
| "loss": 0.8807, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 7.502994011976048, | |
| "grad_norm": 0.7952435612678528, | |
| "learning_rate": 3.4643657460282078e-06, | |
| "loss": 0.8793, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 7.592814371257485, | |
| "grad_norm": 0.9549069404602051, | |
| "learning_rate": 3.2296200847632376e-06, | |
| "loss": 0.8751, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 7.682634730538922, | |
| "grad_norm": 0.6770684719085693, | |
| "learning_rate": 3.001569353140347e-06, | |
| "loss": 0.8772, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 7.772455089820359, | |
| "grad_norm": 0.8119450211524963, | |
| "learning_rate": 2.780439060454756e-06, | |
| "loss": 0.8721, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 7.862275449101796, | |
| "grad_norm": 0.7255963087081909, | |
| "learning_rate": 2.566447872685298e-06, | |
| "loss": 0.8734, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 7.952095808383233, | |
| "grad_norm": 0.6024225950241089, | |
| "learning_rate": 2.359807396265307e-06, | |
| "loss": 0.8779, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 8.035928143712574, | |
| "grad_norm": 0.6504734754562378, | |
| "learning_rate": 2.160721968834344e-06, | |
| "loss": 0.8174, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 8.125748502994012, | |
| "grad_norm": 0.8142725825309753, | |
| "learning_rate": 1.969388457177743e-06, | |
| "loss": 0.863, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 8.215568862275449, | |
| "grad_norm": 0.8136929869651794, | |
| "learning_rate": 1.7859960625537476e-06, | |
| "loss": 0.8677, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 8.305389221556887, | |
| "grad_norm": 0.6553688049316406, | |
| "learning_rate": 1.6107261336007284e-06, | |
| "loss": 0.853, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 8.395209580838323, | |
| "grad_norm": 0.8645008206367493, | |
| "learning_rate": 1.443751987009533e-06, | |
| "loss": 0.8569, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 8.48502994011976, | |
| "grad_norm": 0.7802151441574097, | |
| "learning_rate": 1.2852387361382768e-06, | |
| "loss": 0.8608, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 8.574850299401197, | |
| "grad_norm": 0.8659719824790955, | |
| "learning_rate": 1.1353431277390125e-06, | |
| "loss": 0.8723, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 8.664670658682635, | |
| "grad_norm": 1.073027491569519, | |
| "learning_rate": 9.942133869578164e-07, | |
| "loss": 0.8538, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 8.754491017964071, | |
| "grad_norm": 0.8132328391075134, | |
| "learning_rate": 8.619890707614687e-07, | |
| "loss": 0.8552, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 8.844311377245509, | |
| "grad_norm": 0.959922730922699, | |
| "learning_rate": 7.388009299357412e-07, | |
| "loss": 0.8658, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 8.934131736526947, | |
| "grad_norm": 0.8655376434326172, | |
| "learning_rate": 6.247707797917257e-07, | |
| "loss": 0.8522, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 9.017964071856287, | |
| "grad_norm": 0.8419204950332642, | |
| "learning_rate": 5.200113797080464e-07, | |
| "loss": 0.7919, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 9.107784431137725, | |
| "grad_norm": 0.9591242074966431, | |
| "learning_rate": 4.2462632162809103e-07, | |
| "loss": 0.857, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 9.197604790419161, | |
| "grad_norm": 0.4990997016429901, | |
| "learning_rate": 3.387099276225214e-07, | |
| "loss": 0.8503, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 9.2874251497006, | |
| "grad_norm": 0.9034556150436401, | |
| "learning_rate": 2.623471566183322e-07, | |
| "loss": 0.8465, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 9.377245508982035, | |
| "grad_norm": 0.8334706425666809, | |
| "learning_rate": 1.9561352038673264e-07, | |
| "loss": 0.8576, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 9.467065868263473, | |
| "grad_norm": 0.6150113344192505, | |
| "learning_rate": 1.3857500887288544e-07, | |
| "loss": 0.8544, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 9.55688622754491, | |
| "grad_norm": 0.7956252098083496, | |
| "learning_rate": 9.12880249413628e-08, | |
| "loss": 0.8488, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 9.646706586826348, | |
| "grad_norm": 0.755624532699585, | |
| "learning_rate": 5.379932860185122e-08, | |
| "loss": 0.8512, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 9.736526946107784, | |
| "grad_norm": 0.6398297548294067, | |
| "learning_rate": 2.6145990770238827e-08, | |
| "loss": 0.8559, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 9.826347305389222, | |
| "grad_norm": 0.9212434887886047, | |
| "learning_rate": 8.355356610822984e-09, | |
| "loss": 0.8499, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 9.91616766467066, | |
| "grad_norm": 0.868816614151001, | |
| "learning_rate": 4.4501849589040357e-10, | |
| "loss": 0.8618, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 9.91616766467066, | |
| "step": 1110, | |
| "total_flos": 3.0002693815874355e+17, | |
| "train_loss": 1.1883513411960087, | |
| "train_runtime": 36141.346, | |
| "train_samples_per_second": 0.553, | |
| "train_steps_per_second": 0.031 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 1110, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 10, | |
| "save_steps": 10000, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 3.0002693815874355e+17, | |
| "train_batch_size": 3, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |