| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.1381406271584473, | |
| "eval_steps": 500, | |
| "global_step": 1000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0001381406271584473, | |
| "grad_norm": 0.13134899735450745, | |
| "learning_rate": 0.0, | |
| "loss": 0.2546, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.0002762812543168946, | |
| "grad_norm": 0.1231406033039093, | |
| "learning_rate": 8.695768706980801e-06, | |
| "loss": 0.2051, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.0004144218814753419, | |
| "grad_norm": 0.08561521023511887, | |
| "learning_rate": 1.3782467315509068e-05, | |
| "loss": 0.1455, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.0005525625086337892, | |
| "grad_norm": 0.1449182778596878, | |
| "learning_rate": 1.7391537413961602e-05, | |
| "loss": 0.3489, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.0006907031357922365, | |
| "grad_norm": 0.07311359792947769, | |
| "learning_rate": 2.019094966738107e-05, | |
| "loss": 0.1115, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.0008288437629506838, | |
| "grad_norm": 0.09480362385511398, | |
| "learning_rate": 2.247823602248987e-05, | |
| "loss": 0.1602, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.0009669843901091311, | |
| "grad_norm": 0.08785154670476913, | |
| "learning_rate": 2.441210908061704e-05, | |
| "loss": 0.1448, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.0011051250172675784, | |
| "grad_norm": 0.10113295167684555, | |
| "learning_rate": 2.60873061209424e-05, | |
| "loss": 0.1726, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.0012432656444260257, | |
| "grad_norm": 0.10489944368600845, | |
| "learning_rate": 2.7564934631018137e-05, | |
| "loss": 0.1343, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.001381406271584473, | |
| "grad_norm": 0.07018060237169266, | |
| "learning_rate": 2.8886718374361883e-05, | |
| "loss": 0.0986, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.0015195468987429202, | |
| "grad_norm": 0.07492844760417938, | |
| "learning_rate": 3.008241721328615e-05, | |
| "loss": 0.1243, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.0016576875259013675, | |
| "grad_norm": 0.02947733923792839, | |
| "learning_rate": 3.117400472947067e-05, | |
| "loss": 0.0427, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.0017958281530598148, | |
| "grad_norm": 0.07098696380853653, | |
| "learning_rate": 3.217816790308016e-05, | |
| "loss": 0.074, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.0019339687802182623, | |
| "grad_norm": 0.06729952991008759, | |
| "learning_rate": 3.3107877787597844e-05, | |
| "loss": 0.0756, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.0020721094073767096, | |
| "grad_norm": 0.047634709626436234, | |
| "learning_rate": 3.397341698289014e-05, | |
| "loss": 0.0335, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.002210250034535157, | |
| "grad_norm": 0.01928865537047386, | |
| "learning_rate": 3.4783074827923204e-05, | |
| "loss": 0.0253, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.002348390661693604, | |
| "grad_norm": 0.10047394782304764, | |
| "learning_rate": 3.5543631465891534e-05, | |
| "loss": 0.118, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.0024865312888520514, | |
| "grad_norm": 0.05856269970536232, | |
| "learning_rate": 3.626070333799893e-05, | |
| "loss": 0.0372, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.0026246719160104987, | |
| "grad_norm": 0.011952841654419899, | |
| "learning_rate": 3.69389951409255e-05, | |
| "loss": 0.0123, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.002762812543168946, | |
| "grad_norm": 0.0623110830783844, | |
| "learning_rate": 3.7582487081342674e-05, | |
| "loss": 0.0615, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.002900953170327393, | |
| "grad_norm": 0.020977556705474854, | |
| "learning_rate": 3.819457639612611e-05, | |
| "loss": 0.0215, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.0030390937974858405, | |
| "grad_norm": 0.02789384499192238, | |
| "learning_rate": 3.8778185920266955e-05, | |
| "loss": 0.0263, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.0031772344246442877, | |
| "grad_norm": 0.04567106068134308, | |
| "learning_rate": 3.9335848501569437e-05, | |
| "loss": 0.0298, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.003315375051802735, | |
| "grad_norm": 0.09989040344953537, | |
| "learning_rate": 3.9869773436451475e-05, | |
| "loss": 0.0894, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.0034535156789611823, | |
| "grad_norm": 0.07369614392518997, | |
| "learning_rate": 4.038189933476214e-05, | |
| "loss": 0.0671, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.0035916563061196296, | |
| "grad_norm": 0.1597849428653717, | |
| "learning_rate": 4.087393661006097e-05, | |
| "loss": 0.0934, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.0037297969332780773, | |
| "grad_norm": 0.06050405278801918, | |
| "learning_rate": 4.13474019465272e-05, | |
| "loss": 0.0418, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.0038679375604365246, | |
| "grad_norm": 0.12362563610076904, | |
| "learning_rate": 4.180364649457864e-05, | |
| "loss": 0.1035, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.004006078187594971, | |
| "grad_norm": 0.0370742604136467, | |
| "learning_rate": 4.22438791165378e-05, | |
| "loss": 0.0298, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.004144218814753419, | |
| "grad_norm": 0.13234247267246246, | |
| "learning_rate": 4.2669185689870944e-05, | |
| "loss": 0.1149, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.004282359441911866, | |
| "grad_norm": 0.11454839259386063, | |
| "learning_rate": 4.308054524410194e-05, | |
| "loss": 0.0825, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.004420500069070314, | |
| "grad_norm": 0.055333029478788376, | |
| "learning_rate": 4.347884353490401e-05, | |
| "loss": 0.0309, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.0045586406962287605, | |
| "grad_norm": 0.010015130043029785, | |
| "learning_rate": 4.386488452879522e-05, | |
| "loss": 0.0112, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.004696781323387208, | |
| "grad_norm": 0.02108495496213436, | |
| "learning_rate": 4.423940017287234e-05, | |
| "loss": 0.0144, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.004834921950545655, | |
| "grad_norm": 0.045554034411907196, | |
| "learning_rate": 4.460305874799812e-05, | |
| "loss": 0.0325, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.004973062577704103, | |
| "grad_norm": 0.05177578702569008, | |
| "learning_rate": 4.495647204497974e-05, | |
| "loss": 0.0333, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.0051112032048625505, | |
| "grad_norm": 0.08447037637233734, | |
| "learning_rate": 4.5300201557312036e-05, | |
| "loss": 0.0489, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.005249343832020997, | |
| "grad_norm": 0.03207926079630852, | |
| "learning_rate": 4.56347638479063e-05, | |
| "loss": 0.0217, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.005387484459179445, | |
| "grad_norm": 0.04534975811839104, | |
| "learning_rate": 4.596063521858923e-05, | |
| "loss": 0.0332, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.005525625086337892, | |
| "grad_norm": 0.04216322675347328, | |
| "learning_rate": 4.627825578832348e-05, | |
| "loss": 0.0363, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.0056637657134963396, | |
| "grad_norm": 0.04155502840876579, | |
| "learning_rate": 4.65880330677802e-05, | |
| "loss": 0.0238, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.005801906340654786, | |
| "grad_norm": 0.15111327171325684, | |
| "learning_rate": 4.6890345103106905e-05, | |
| "loss": 0.1467, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.005940046967813234, | |
| "grad_norm": 0.057506952434778214, | |
| "learning_rate": 4.7185543249731356e-05, | |
| "loss": 0.0366, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.006078187594971681, | |
| "grad_norm": 0.15154612064361572, | |
| "learning_rate": 4.747395462724775e-05, | |
| "loss": 0.2312, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.006216328222130129, | |
| "grad_norm": 0.1836182028055191, | |
| "learning_rate": 4.775588429839921e-05, | |
| "loss": 0.0679, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.0063544688492885755, | |
| "grad_norm": 0.04606263339519501, | |
| "learning_rate": 4.803161720855024e-05, | |
| "loss": 0.0216, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.006492609476447023, | |
| "grad_norm": 0.07287360727787018, | |
| "learning_rate": 4.830141991656282e-05, | |
| "loss": 0.0355, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.00663075010360547, | |
| "grad_norm": 0.10376694798469543, | |
| "learning_rate": 4.856554214343227e-05, | |
| "loss": 0.0384, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.006768890730763918, | |
| "grad_norm": 0.0253737885504961, | |
| "learning_rate": 4.882421816123408e-05, | |
| "loss": 0.0145, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.006907031357922365, | |
| "grad_norm": 0.045460280030965805, | |
| "learning_rate": 4.9077668041742954e-05, | |
| "loss": 0.0272, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.007045171985080812, | |
| "grad_norm": 0.00954690482467413, | |
| "learning_rate": 4.93260987814006e-05, | |
| "loss": 0.0086, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.007183312612239259, | |
| "grad_norm": 0.12355463951826096, | |
| "learning_rate": 4.956970531704177e-05, | |
| "loss": 0.0398, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.007321453239397707, | |
| "grad_norm": 0.028379764407873154, | |
| "learning_rate": 4.9808671444865916e-05, | |
| "loss": 0.0166, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.0074595938665561546, | |
| "grad_norm": 0.029451319947838783, | |
| "learning_rate": 5.0043170653507996e-05, | |
| "loss": 0.0285, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.007597734493714601, | |
| "grad_norm": 0.1718253642320633, | |
| "learning_rate": 5.027336688066723e-05, | |
| "loss": 0.1021, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.007735875120873049, | |
| "grad_norm": 0.02496183291077614, | |
| "learning_rate": 5.049941520155945e-05, | |
| "loss": 0.0167, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.007874015748031496, | |
| "grad_norm": 0.14813463389873505, | |
| "learning_rate": 5.072146245643457e-05, | |
| "loss": 0.0571, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.008012156375189943, | |
| "grad_norm": 0.004978721961379051, | |
| "learning_rate": 5.09396478235186e-05, | |
| "loss": 0.0063, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.008150297002348391, | |
| "grad_norm": 0.027830282226204872, | |
| "learning_rate": 5.115410334297882e-05, | |
| "loss": 0.0209, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.008288437629506838, | |
| "grad_norm": 0.02240496128797531, | |
| "learning_rate": 5.136495439685174e-05, | |
| "loss": 0.0166, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.008426578256665285, | |
| "grad_norm": 0.004362566862255335, | |
| "learning_rate": 5.157232014930198e-05, | |
| "loss": 0.005, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.008564718883823732, | |
| "grad_norm": 0.31787729263305664, | |
| "learning_rate": 5.1776313951082735e-05, | |
| "loss": 0.1389, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.00870285951098218, | |
| "grad_norm": 0.009066242724657059, | |
| "learning_rate": 5.1977043711635176e-05, | |
| "loss": 0.0081, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.008841000138140627, | |
| "grad_norm": 0.006892406847327948, | |
| "learning_rate": 5.21746122418848e-05, | |
| "loss": 0.007, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.008979140765299074, | |
| "grad_norm": 0.06632635742425919, | |
| "learning_rate": 5.236911757046123e-05, | |
| "loss": 0.0397, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.009117281392457521, | |
| "grad_norm": 0.03089607134461403, | |
| "learning_rate": 5.2560653235776016e-05, | |
| "loss": 0.0194, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.00925542201961597, | |
| "grad_norm": 0.010297919623553753, | |
| "learning_rate": 5.27493085561372e-05, | |
| "loss": 0.0094, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.009393562646774416, | |
| "grad_norm": 0.00596292968839407, | |
| "learning_rate": 5.293516887985315e-05, | |
| "loss": 0.0056, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.009531703273932863, | |
| "grad_norm": 0.03583955764770508, | |
| "learning_rate": 5.31183158170785e-05, | |
| "loss": 0.034, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.00966984390109131, | |
| "grad_norm": 0.022406477481126785, | |
| "learning_rate": 5.3298827454978915e-05, | |
| "loss": 0.0177, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.009807984528249759, | |
| "grad_norm": 0.02521800622344017, | |
| "learning_rate": 5.347677855763414e-05, | |
| "loss": 0.0137, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.009946125155408206, | |
| "grad_norm": 0.023860285058617592, | |
| "learning_rate": 5.3652240751960537e-05, | |
| "loss": 0.0223, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.010084265782566652, | |
| "grad_norm": 0.044385574758052826, | |
| "learning_rate": 5.3825282700810095e-05, | |
| "loss": 0.023, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.010222406409725101, | |
| "grad_norm": 0.011455412954092026, | |
| "learning_rate": 5.399597026429285e-05, | |
| "loss": 0.0115, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.010360547036883548, | |
| "grad_norm": 0.04733498394489288, | |
| "learning_rate": 5.416436665027121e-05, | |
| "loss": 0.0301, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.010498687664041995, | |
| "grad_norm": 0.00907475408166647, | |
| "learning_rate": 5.43305325548871e-05, | |
| "loss": 0.0068, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.010636828291200441, | |
| "grad_norm": 0.14968900382518768, | |
| "learning_rate": 5.449452629390319e-05, | |
| "loss": 0.0619, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.01077496891835889, | |
| "grad_norm": 0.15661145746707916, | |
| "learning_rate": 5.4656403925570034e-05, | |
| "loss": 0.0722, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.010913109545517337, | |
| "grad_norm": 0.004802990239113569, | |
| "learning_rate": 5.481621936566648e-05, | |
| "loss": 0.0054, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.011051250172675784, | |
| "grad_norm": 0.0075927311554551125, | |
| "learning_rate": 5.4974024495304276e-05, | |
| "loss": 0.0071, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.01118939079983423, | |
| "grad_norm": 0.11258488148450851, | |
| "learning_rate": 5.5129869262036274e-05, | |
| "loss": 0.0985, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.011327531426992679, | |
| "grad_norm": 0.052467264235019684, | |
| "learning_rate": 5.5283801774761e-05, | |
| "loss": 0.0298, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.011465672054151126, | |
| "grad_norm": 0.05677814409136772, | |
| "learning_rate": 5.543586839287528e-05, | |
| "loss": 0.0272, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.011603812681309573, | |
| "grad_norm": 0.11758122593164444, | |
| "learning_rate": 5.55861138100877e-05, | |
| "loss": 0.0555, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.01174195330846802, | |
| "grad_norm": 0.021579308435320854, | |
| "learning_rate": 5.573458113327261e-05, | |
| "loss": 0.0141, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.011880093935626468, | |
| "grad_norm": 0.018838461488485336, | |
| "learning_rate": 5.5881311956712153e-05, | |
| "loss": 0.016, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.012018234562784915, | |
| "grad_norm": 0.07577263563871384, | |
| "learning_rate": 5.602634643204686e-05, | |
| "loss": 0.0568, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.012156375189943362, | |
| "grad_norm": 0.16609540581703186, | |
| "learning_rate": 5.6169723334228564e-05, | |
| "loss": 0.2139, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.01229451581710181, | |
| "grad_norm": 0.008787001483142376, | |
| "learning_rate": 5.631148012374702e-05, | |
| "loss": 0.0088, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.012432656444260257, | |
| "grad_norm": 0.005372604820877314, | |
| "learning_rate": 5.645165300538001e-05, | |
| "loss": 0.0064, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.012570797071418704, | |
| "grad_norm": 0.020076818764209747, | |
| "learning_rate": 5.65902769836972e-05, | |
| "loss": 0.0146, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.012708937698577151, | |
| "grad_norm": 0.02066107466816902, | |
| "learning_rate": 5.672738591553104e-05, | |
| "loss": 0.0245, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.0128470783257356, | |
| "grad_norm": 0.03774566575884819, | |
| "learning_rate": 5.6863012559611006e-05, | |
| "loss": 0.0254, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.012985218952894046, | |
| "grad_norm": 0.0144899757578969, | |
| "learning_rate": 5.699718862354363e-05, | |
| "loss": 0.0153, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.013123359580052493, | |
| "grad_norm": 0.06870149821043015, | |
| "learning_rate": 5.712994480830657e-05, | |
| "loss": 0.0499, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.01326150020721094, | |
| "grad_norm": 0.02667887881398201, | |
| "learning_rate": 5.726131085041308e-05, | |
| "loss": 0.0202, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.013399640834369389, | |
| "grad_norm": 0.012837301008403301, | |
| "learning_rate": 5.7391315561891545e-05, | |
| "loss": 0.0113, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.013537781461527836, | |
| "grad_norm": 0.052629996091127396, | |
| "learning_rate": 5.751998686821489e-05, | |
| "loss": 0.0505, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.013675922088686282, | |
| "grad_norm": 0.020552391186356544, | |
| "learning_rate": 5.764735184430429e-05, | |
| "loss": 0.0186, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.01381406271584473, | |
| "grad_norm": 0.019090712070465088, | |
| "learning_rate": 5.7773436748723766e-05, | |
| "loss": 0.0126, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.013952203343003178, | |
| "grad_norm": 0.05564524605870247, | |
| "learning_rate": 5.7898267056173305e-05, | |
| "loss": 0.0311, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.014090343970161625, | |
| "grad_norm": 0.003988645505160093, | |
| "learning_rate": 5.80218674883814e-05, | |
| "loss": 0.0043, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.014228484597320071, | |
| "grad_norm": 0.014693093486130238, | |
| "learning_rate": 5.814426204349046e-05, | |
| "loss": 0.0135, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.014366625224478518, | |
| "grad_norm": 0.010103048756718636, | |
| "learning_rate": 5.826547402402257e-05, | |
| "loss": 0.0088, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.014504765851636967, | |
| "grad_norm": 0.06549032032489777, | |
| "learning_rate": 5.838552606350718e-05, | |
| "loss": 0.0285, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.014642906478795414, | |
| "grad_norm": 0.014358027838170528, | |
| "learning_rate": 5.850444015184672e-05, | |
| "loss": 0.0131, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.01478104710595386, | |
| "grad_norm": 0.009209529496729374, | |
| "learning_rate": 5.862223765949125e-05, | |
| "loss": 0.0084, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.014919187733112309, | |
| "grad_norm": 0.004671560134738684, | |
| "learning_rate": 5.873893936048881e-05, | |
| "loss": 0.0058, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.015057328360270756, | |
| "grad_norm": 0.018950628116726875, | |
| "learning_rate": 5.885456545447318e-05, | |
| "loss": 0.0122, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.015195468987429203, | |
| "grad_norm": 0.009532412514090538, | |
| "learning_rate": 5.8969135587648026e-05, | |
| "loss": 0.0081, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.01533360961458765, | |
| "grad_norm": 0.11297975480556488, | |
| "learning_rate": 5.90826688728211e-05, | |
| "loss": 0.0985, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.015471750241746098, | |
| "grad_norm": 0.021297922357916832, | |
| "learning_rate": 5.9195183908540244e-05, | |
| "loss": 0.0172, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.015609890868904545, | |
| "grad_norm": 0.011318448930978775, | |
| "learning_rate": 5.930669879737879e-05, | |
| "loss": 0.0106, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.015748031496062992, | |
| "grad_norm": 0.01391973253339529, | |
| "learning_rate": 5.9417231163415365e-05, | |
| "loss": 0.0135, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.01588617212322144, | |
| "grad_norm": 0.004168018698692322, | |
| "learning_rate": 5.9526798168950515e-05, | |
| "loss": 0.0048, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.016024312750379886, | |
| "grad_norm": 0.08173353970050812, | |
| "learning_rate": 5.96354165304994e-05, | |
| "loss": 0.0481, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.016162453377538334, | |
| "grad_norm": 0.03505641967058182, | |
| "learning_rate": 5.9743102534098305e-05, | |
| "loss": 0.0212, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.016300594004696783, | |
| "grad_norm": 0.020770832896232605, | |
| "learning_rate": 5.9849872049959624e-05, | |
| "loss": 0.0154, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.016438734631855228, | |
| "grad_norm": 0.013789334334433079, | |
| "learning_rate": 5.9955740546508573e-05, | |
| "loss": 0.0138, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.016576875259013676, | |
| "grad_norm": 0.004777516704052687, | |
| "learning_rate": 6.006072310383254e-05, | |
| "loss": 0.0051, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.01671501588617212, | |
| "grad_norm": 0.09303977340459824, | |
| "learning_rate": 6.01648344265723e-05, | |
| "loss": 0.0718, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.01685315651333057, | |
| "grad_norm": 0.006040315609425306, | |
| "learning_rate": 6.026808885628279e-05, | |
| "loss": 0.006, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.01699129714048902, | |
| "grad_norm": 0.0029070712625980377, | |
| "learning_rate": 6.0370500383289253e-05, | |
| "loss": 0.0041, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.017129437767647464, | |
| "grad_norm": 0.018031740561127663, | |
| "learning_rate": 6.0472082658063546e-05, | |
| "loss": 0.0118, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.017267578394805912, | |
| "grad_norm": 0.0054465411230921745, | |
| "learning_rate": 6.057284900214323e-05, | |
| "loss": 0.006, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.01740571902196436, | |
| "grad_norm": 0.024758221581578255, | |
| "learning_rate": 6.0672812418615974e-05, | |
| "loss": 0.0181, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.017543859649122806, | |
| "grad_norm": 0.060118090361356735, | |
| "learning_rate": 6.077198560218932e-05, | |
| "loss": 0.0314, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.017682000276281255, | |
| "grad_norm": 0.0032233481761068106, | |
| "learning_rate": 6.087038094886561e-05, | |
| "loss": 0.0044, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.017820140903439703, | |
| "grad_norm": 0.023020045831799507, | |
| "learning_rate": 6.096801056524042e-05, | |
| "loss": 0.0205, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.01795828153059815, | |
| "grad_norm": 0.012697757221758366, | |
| "learning_rate": 6.106488627744203e-05, | |
| "loss": 0.0126, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.018096422157756597, | |
| "grad_norm": 0.023354003205895424, | |
| "learning_rate": 6.116101963972834e-05, | |
| "loss": 0.012, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.018234562784915042, | |
| "grad_norm": 0.020162884145975113, | |
| "learning_rate": 6.125642194275681e-05, | |
| "loss": 0.0168, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.01837270341207349, | |
| "grad_norm": 0.02640388533473015, | |
| "learning_rate": 6.135110422154255e-05, | |
| "loss": 0.0148, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.01851084403923194, | |
| "grad_norm": 0.004422938451170921, | |
| "learning_rate": 6.1445077263118e-05, | |
| "loss": 0.0053, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.018648984666390384, | |
| "grad_norm": 0.0029348707757890224, | |
| "learning_rate": 6.153835161390827e-05, | |
| "loss": 0.0041, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.018787125293548833, | |
| "grad_norm": 0.008122744038701057, | |
| "learning_rate": 6.163093758683395e-05, | |
| "loss": 0.0072, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.01892526592070728, | |
| "grad_norm": 0.012685508467257023, | |
| "learning_rate": 6.172284526815391e-05, | |
| "loss": 0.0108, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.019063406547865726, | |
| "grad_norm": 0.07143422216176987, | |
| "learning_rate": 6.181408452405932e-05, | |
| "loss": 0.0354, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.019201547175024175, | |
| "grad_norm": 0.013023455627262592, | |
| "learning_rate": 6.190466500702942e-05, | |
| "loss": 0.0121, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.01933968780218262, | |
| "grad_norm": 0.11455456912517548, | |
| "learning_rate": 6.199459616195971e-05, | |
| "loss": 0.0916, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.01947782842934107, | |
| "grad_norm": 0.004115304443985224, | |
| "learning_rate": 6.208388723207189e-05, | |
| "loss": 0.005, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.019615969056499517, | |
| "grad_norm": 0.006390172056853771, | |
| "learning_rate": 6.217254726461495e-05, | |
| "loss": 0.0056, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.019754109683657962, | |
| "grad_norm": 0.0032778135500848293, | |
| "learning_rate": 6.22605851163663e-05, | |
| "loss": 0.0043, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.01989225031081641, | |
| "grad_norm": 0.005337041802704334, | |
| "learning_rate": 6.234800945894134e-05, | |
| "loss": 0.0054, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.02003039093797486, | |
| "grad_norm": 0.004124751314520836, | |
| "learning_rate": 6.243482878391886e-05, | |
| "loss": 0.0051, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.020168531565133305, | |
| "grad_norm": 0.005773474462330341, | |
| "learning_rate": 6.25210514077909e-05, | |
| "loss": 0.006, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.020306672192291753, | |
| "grad_norm": 0.003976911772042513, | |
| "learning_rate": 6.260668547674315e-05, | |
| "loss": 0.005, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.020444812819450202, | |
| "grad_norm": 0.015533742494881153, | |
| "learning_rate": 6.269173897127364e-05, | |
| "loss": 0.0115, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.020582953446608647, | |
| "grad_norm": 0.007217871956527233, | |
| "learning_rate": 6.277621971065575e-05, | |
| "loss": 0.007, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.020721094073767096, | |
| "grad_norm": 0.02741631306707859, | |
| "learning_rate": 6.286013535725201e-05, | |
| "loss": 0.0262, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.02085923470092554, | |
| "grad_norm": 0.009277957491576672, | |
| "learning_rate": 6.294349342068456e-05, | |
| "loss": 0.0074, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.02099737532808399, | |
| "grad_norm": 0.03387785702943802, | |
| "learning_rate": 6.30263012618679e-05, | |
| "loss": 0.02, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.021135515955242438, | |
| "grad_norm": 0.02018630877137184, | |
| "learning_rate": 6.310856609690967e-05, | |
| "loss": 0.0135, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.021273656582400883, | |
| "grad_norm": 0.0030231664422899485, | |
| "learning_rate": 6.3190295000884e-05, | |
| "loss": 0.0039, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.02141179720955933, | |
| "grad_norm": 0.038589298725128174, | |
| "learning_rate": 6.327149491148301e-05, | |
| "loss": 0.0358, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.02154993783671778, | |
| "grad_norm": 0.11658922582864761, | |
| "learning_rate": 6.335217263255083e-05, | |
| "loss": 0.0583, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.021688078463876225, | |
| "grad_norm": 0.015740083530545235, | |
| "learning_rate": 6.343233483750467e-05, | |
| "loss": 0.0134, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.021826219091034674, | |
| "grad_norm": 0.04937628656625748, | |
| "learning_rate": 6.351198807264728e-05, | |
| "loss": 0.0245, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.021964359718193122, | |
| "grad_norm": 0.03167060390114784, | |
| "learning_rate": 6.359113876037498e-05, | |
| "loss": 0.0142, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.022102500345351567, | |
| "grad_norm": 0.0016946723917499185, | |
| "learning_rate": 6.366979320228508e-05, | |
| "loss": 0.0025, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.022240640972510016, | |
| "grad_norm": 0.017945002764463425, | |
| "learning_rate": 6.374795758218648e-05, | |
| "loss": 0.0114, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.02237878159966846, | |
| "grad_norm": 0.015089811757206917, | |
| "learning_rate": 6.382563796901706e-05, | |
| "loss": 0.0104, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.02251692222682691, | |
| "grad_norm": 0.037023257464170456, | |
| "learning_rate": 6.390284031967139e-05, | |
| "loss": 0.0212, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.022655062853985358, | |
| "grad_norm": 0.00312859402038157, | |
| "learning_rate": 6.397957048174181e-05, | |
| "loss": 0.0041, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.022793203481143803, | |
| "grad_norm": 0.016800010576844215, | |
| "learning_rate": 6.405583419617628e-05, | |
| "loss": 0.0117, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.022931344108302252, | |
| "grad_norm": 0.007050058338791132, | |
| "learning_rate": 6.413163709985608e-05, | |
| "loss": 0.0077, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.0230694847354607, | |
| "grad_norm": 0.010559717193245888, | |
| "learning_rate": 6.420698472809568e-05, | |
| "loss": 0.0101, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.023207625362619146, | |
| "grad_norm": 0.007447375915944576, | |
| "learning_rate": 6.428188251706851e-05, | |
| "loss": 0.0063, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.023345765989777594, | |
| "grad_norm": 0.016936203464865685, | |
| "learning_rate": 6.435633580616032e-05, | |
| "loss": 0.0123, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.02348390661693604, | |
| "grad_norm": 0.012545132078230381, | |
| "learning_rate": 6.443034984025341e-05, | |
| "loss": 0.0106, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.023622047244094488, | |
| "grad_norm": 0.004759025294333696, | |
| "learning_rate": 6.450392977194364e-05, | |
| "loss": 0.0055, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.023760187871252936, | |
| "grad_norm": 0.004177019465714693, | |
| "learning_rate": 6.457708066369296e-05, | |
| "loss": 0.005, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.02389832849841138, | |
| "grad_norm": 0.005387986544519663, | |
| "learning_rate": 6.464980748991957e-05, | |
| "loss": 0.0052, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.02403646912556983, | |
| "grad_norm": 0.014265534467995167, | |
| "learning_rate": 6.472211513902766e-05, | |
| "loss": 0.0105, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.02417460975272828, | |
| "grad_norm": 0.009414114989340305, | |
| "learning_rate": 6.47940084153792e-05, | |
| "loss": 0.0082, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.024312750379886724, | |
| "grad_norm": 0.006094431504607201, | |
| "learning_rate": 6.486549204120935e-05, | |
| "loss": 0.007, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.024450891007045172, | |
| "grad_norm": 0.01183071918785572, | |
| "learning_rate": 6.493657065848789e-05, | |
| "loss": 0.0051, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.02458903163420362, | |
| "grad_norm": 0.01014156173914671, | |
| "learning_rate": 6.500724883072781e-05, | |
| "loss": 0.008, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.024727172261362066, | |
| "grad_norm": 0.011028112843632698, | |
| "learning_rate": 6.507753104474373e-05, | |
| "loss": 0.0117, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.024865312888520515, | |
| "grad_norm": 0.09993032366037369, | |
| "learning_rate": 6.514742171236081e-05, | |
| "loss": 0.0368, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.02500345351567896, | |
| "grad_norm": 0.00730351684615016, | |
| "learning_rate": 6.521692517207681e-05, | |
| "loss": 0.0055, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.02514159414283741, | |
| "grad_norm": 0.012081836350262165, | |
| "learning_rate": 6.5286045690678e-05, | |
| "loss": 0.0096, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.025279734769995857, | |
| "grad_norm": 0.03789973631501198, | |
| "learning_rate": 6.535478746481105e-05, | |
| "loss": 0.0214, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.025417875397154302, | |
| "grad_norm": 0.15203222632408142, | |
| "learning_rate": 6.542315462251184e-05, | |
| "loss": 0.173, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.02555601602431275, | |
| "grad_norm": 0.01032046414911747, | |
| "learning_rate": 6.54911512246931e-05, | |
| "loss": 0.0096, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.0256941566514712, | |
| "grad_norm": 0.028638096526265144, | |
| "learning_rate": 6.555878126659181e-05, | |
| "loss": 0.0145, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.025832297278629644, | |
| "grad_norm": 0.003815449308604002, | |
| "learning_rate": 6.562604867917768e-05, | |
| "loss": 0.005, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.025970437905788093, | |
| "grad_norm": 0.02196453884243965, | |
| "learning_rate": 6.569295733052443e-05, | |
| "loss": 0.0163, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.026108578532946538, | |
| "grad_norm": 0.027504578232765198, | |
| "learning_rate": 6.575951102714424e-05, | |
| "loss": 0.0236, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.026246719160104987, | |
| "grad_norm": 0.01594018191099167, | |
| "learning_rate": 6.582571351528738e-05, | |
| "loss": 0.0151, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.026384859787263435, | |
| "grad_norm": 0.014933554455637932, | |
| "learning_rate": 6.589156848220747e-05, | |
| "loss": 0.0081, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.02652300041442188, | |
| "grad_norm": 0.0055963508784770966, | |
| "learning_rate": 6.595707955739388e-05, | |
| "loss": 0.0062, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.02666114104158033, | |
| "grad_norm": 0.050148047506809235, | |
| "learning_rate": 6.602225031377195e-05, | |
| "loss": 0.0194, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.026799281668738777, | |
| "grad_norm": 0.029830805957317352, | |
| "learning_rate": 6.608708426887235e-05, | |
| "loss": 0.0205, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.026937422295897222, | |
| "grad_norm": 0.011922827921807766, | |
| "learning_rate": 6.61515848859703e-05, | |
| "loss": 0.0096, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.02707556292305567, | |
| "grad_norm": 0.009799070656299591, | |
| "learning_rate": 6.621575557519569e-05, | |
| "loss": 0.01, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.02721370355021412, | |
| "grad_norm": 0.01243631262332201, | |
| "learning_rate": 6.627959969461484e-05, | |
| "loss": 0.0099, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.027351844177372565, | |
| "grad_norm": 0.07121865451335907, | |
| "learning_rate": 6.634312055128508e-05, | |
| "loss": 0.0304, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.027489984804531013, | |
| "grad_norm": 0.017733663320541382, | |
| "learning_rate": 6.64063214022826e-05, | |
| "loss": 0.0137, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.02762812543168946, | |
| "grad_norm": 0.012368605472147465, | |
| "learning_rate": 6.646920545570455e-05, | |
| "loss": 0.0084, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.027766266058847907, | |
| "grad_norm": 0.02112652361392975, | |
| "learning_rate": 6.653177587164626e-05, | |
| "loss": 0.0179, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.027904406686006356, | |
| "grad_norm": 0.02623148076236248, | |
| "learning_rate": 6.659403576315411e-05, | |
| "loss": 0.0197, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.0280425473131648, | |
| "grad_norm": 0.03918753191828728, | |
| "learning_rate": 6.665598819715483e-05, | |
| "loss": 0.0258, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.02818068794032325, | |
| "grad_norm": 0.013951225206255913, | |
| "learning_rate": 6.67176361953622e-05, | |
| "loss": 0.0135, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.028318828567481698, | |
| "grad_norm": 0.007952759973704815, | |
| "learning_rate": 6.677898273516128e-05, | |
| "loss": 0.0089, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.028456969194640143, | |
| "grad_norm": 0.018430588766932487, | |
| "learning_rate": 6.684003075047128e-05, | |
| "loss": 0.0149, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.02859510982179859, | |
| "grad_norm": 0.023949814960360527, | |
| "learning_rate": 6.690078313258756e-05, | |
| "loss": 0.0132, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.028733250448957037, | |
| "grad_norm": 0.024928990751504898, | |
| "learning_rate": 6.696124273100337e-05, | |
| "loss": 0.0167, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.028871391076115485, | |
| "grad_norm": 0.030003704130649567, | |
| "learning_rate": 6.702141235421166e-05, | |
| "loss": 0.0172, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.029009531703273934, | |
| "grad_norm": 0.0721101239323616, | |
| "learning_rate": 6.708129477048798e-05, | |
| "loss": 0.0368, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.02914767233043238, | |
| "grad_norm": 0.08490179479122162, | |
| "learning_rate": 6.71408927086548e-05, | |
| "loss": 0.0536, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.029285812957590827, | |
| "grad_norm": 0.03323817253112793, | |
| "learning_rate": 6.720020885882752e-05, | |
| "loss": 0.0178, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.029423953584749276, | |
| "grad_norm": 0.1837650090456009, | |
| "learning_rate": 6.72592458731432e-05, | |
| "loss": 0.1725, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.02956209421190772, | |
| "grad_norm": 0.019839327782392502, | |
| "learning_rate": 6.731800636647207e-05, | |
| "loss": 0.0181, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.02970023483906617, | |
| "grad_norm": 0.1045767292380333, | |
| "learning_rate": 6.737649291711243e-05, | |
| "loss": 0.0833, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.029838375466224618, | |
| "grad_norm": 0.006606792565435171, | |
| "learning_rate": 6.743470806746961e-05, | |
| "loss": 0.0075, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.029976516093383063, | |
| "grad_norm": 0.03129244223237038, | |
| "learning_rate": 6.749265432471897e-05, | |
| "loss": 0.0193, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.030114656720541512, | |
| "grad_norm": 0.045653145760297775, | |
| "learning_rate": 6.755033416145399e-05, | |
| "loss": 0.0199, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.030252797347699957, | |
| "grad_norm": 0.010944471694529057, | |
| "learning_rate": 6.760775001631917e-05, | |
| "loss": 0.0095, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.030390937974858406, | |
| "grad_norm": 0.01569080539047718, | |
| "learning_rate": 6.766490429462884e-05, | |
| "loss": 0.0104, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.030529078602016854, | |
| "grad_norm": 0.021444043144583702, | |
| "learning_rate": 6.772179936897169e-05, | |
| "loss": 0.0142, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 0.0306672192291753, | |
| "grad_norm": 0.28277868032455444, | |
| "learning_rate": 6.77784375798019e-05, | |
| "loss": 0.2374, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.030805359856333748, | |
| "grad_norm": 0.45913952589035034, | |
| "learning_rate": 6.783482123601665e-05, | |
| "loss": 0.4551, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 0.030943500483492196, | |
| "grad_norm": 0.020372901111841202, | |
| "learning_rate": 6.789095261552104e-05, | |
| "loss": 0.0168, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.03108164111065064, | |
| "grad_norm": 0.004885848145931959, | |
| "learning_rate": 6.794683396578028e-05, | |
| "loss": 0.005, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.03121978173780909, | |
| "grad_norm": 0.01693105325102806, | |
| "learning_rate": 6.80024675043596e-05, | |
| "loss": 0.0163, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 0.03135792236496754, | |
| "grad_norm": 0.11799801141023636, | |
| "learning_rate": 6.805785541945228e-05, | |
| "loss": 0.0717, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 0.031496062992125984, | |
| "grad_norm": 0.036976467818021774, | |
| "learning_rate": 6.811299987039618e-05, | |
| "loss": 0.0245, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.03163420361928443, | |
| "grad_norm": 0.02155802771449089, | |
| "learning_rate": 6.816790298817876e-05, | |
| "loss": 0.0177, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 0.03177234424644288, | |
| "grad_norm": 0.07842877507209778, | |
| "learning_rate": 6.822256687593131e-05, | |
| "loss": 0.0581, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.031910484873601326, | |
| "grad_norm": 0.03338398411870003, | |
| "learning_rate": 6.827699360941226e-05, | |
| "loss": 0.0325, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.03204862550075977, | |
| "grad_norm": 0.032553404569625854, | |
| "learning_rate": 6.83311852374802e-05, | |
| "loss": 0.0322, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 0.03218676612791822, | |
| "grad_norm": 0.04135012626647949, | |
| "learning_rate": 6.838514378255668e-05, | |
| "loss": 0.0309, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 0.03232490675507667, | |
| "grad_norm": 0.015881696715950966, | |
| "learning_rate": 6.84388712410791e-05, | |
| "loss": 0.0122, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.03246304738223511, | |
| "grad_norm": 0.007995386607944965, | |
| "learning_rate": 6.849236958394389e-05, | |
| "loss": 0.0064, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.032601188009393565, | |
| "grad_norm": 0.02736758068203926, | |
| "learning_rate": 6.854564075694041e-05, | |
| "loss": 0.0235, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 0.03273932863655201, | |
| "grad_norm": 0.002618682337924838, | |
| "learning_rate": 6.859868668117555e-05, | |
| "loss": 0.0042, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.032877469263710456, | |
| "grad_norm": 0.005231305491179228, | |
| "learning_rate": 6.865150925348938e-05, | |
| "loss": 0.0068, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 0.03301560989086891, | |
| "grad_norm": 0.09725295752286911, | |
| "learning_rate": 6.870411034686229e-05, | |
| "loss": 0.0792, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 0.03315375051802735, | |
| "grad_norm": 0.050293877720832825, | |
| "learning_rate": 6.875649181081335e-05, | |
| "loss": 0.0283, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.0332918911451858, | |
| "grad_norm": 0.028621939942240715, | |
| "learning_rate": 6.880865547179058e-05, | |
| "loss": 0.0281, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 0.03343003177234424, | |
| "grad_norm": 0.008942426182329655, | |
| "learning_rate": 6.886060313355311e-05, | |
| "loss": 0.0102, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 0.033568172399502695, | |
| "grad_norm": 0.017691925168037415, | |
| "learning_rate": 6.891233657754533e-05, | |
| "loss": 0.0144, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.03370631302666114, | |
| "grad_norm": 0.06508173048496246, | |
| "learning_rate": 6.896385756326359e-05, | |
| "loss": 0.0417, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 0.033844453653819585, | |
| "grad_norm": 0.04150944948196411, | |
| "learning_rate": 6.901516782861516e-05, | |
| "loss": 0.0291, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.03398259428097804, | |
| "grad_norm": 0.03793828561902046, | |
| "learning_rate": 6.906626909027006e-05, | |
| "loss": 0.0233, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.03412073490813648, | |
| "grad_norm": 0.010130356065928936, | |
| "learning_rate": 6.911716304400567e-05, | |
| "loss": 0.0099, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 0.03425887553529493, | |
| "grad_norm": 0.13118235766887665, | |
| "learning_rate": 6.916785136504435e-05, | |
| "loss": 0.0867, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 0.03439701616245338, | |
| "grad_norm": 0.01260526105761528, | |
| "learning_rate": 6.921833570838434e-05, | |
| "loss": 0.0125, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.034535156789611825, | |
| "grad_norm": 0.008999558165669441, | |
| "learning_rate": 6.926861770912402e-05, | |
| "loss": 0.0073, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.03467329741677027, | |
| "grad_norm": 0.00830650795251131, | |
| "learning_rate": 6.931869898277965e-05, | |
| "loss": 0.0091, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 0.03481143804392872, | |
| "grad_norm": 0.008377288468182087, | |
| "learning_rate": 6.936858112559677e-05, | |
| "loss": 0.0075, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.03494957867108717, | |
| "grad_norm": 0.053149525076150894, | |
| "learning_rate": 6.941826571485559e-05, | |
| "loss": 0.0127, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 0.03508771929824561, | |
| "grad_norm": 0.00812042597681284, | |
| "learning_rate": 6.946775430917013e-05, | |
| "loss": 0.0089, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 0.035225859925404064, | |
| "grad_norm": 0.11070810258388519, | |
| "learning_rate": 6.951704844878168e-05, | |
| "loss": 0.0831, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.03536400055256251, | |
| "grad_norm": 0.012675175443291664, | |
| "learning_rate": 6.956614965584641e-05, | |
| "loss": 0.0114, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.035502141179720954, | |
| "grad_norm": 0.04479183256626129, | |
| "learning_rate": 6.96150594347174e-05, | |
| "loss": 0.0212, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 0.035640281806879406, | |
| "grad_norm": 0.0070395260117948055, | |
| "learning_rate": 6.966377927222123e-05, | |
| "loss": 0.0085, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.03577842243403785, | |
| "grad_norm": 0.0065799071453511715, | |
| "learning_rate": 6.971231063792908e-05, | |
| "loss": 0.0077, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 0.0359165630611963, | |
| "grad_norm": 0.011478329077363014, | |
| "learning_rate": 6.976065498442284e-05, | |
| "loss": 0.0096, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.03605470368835474, | |
| "grad_norm": 0.005571560934185982, | |
| "learning_rate": 6.980881374755593e-05, | |
| "loss": 0.0058, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 0.036192844315513194, | |
| "grad_norm": 0.15769460797309875, | |
| "learning_rate": 6.985678834670915e-05, | |
| "loss": 0.057, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 0.03633098494267164, | |
| "grad_norm": 0.010631869547069073, | |
| "learning_rate": 6.990458018504173e-05, | |
| "loss": 0.0106, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 0.036469125569830084, | |
| "grad_norm": 0.020581720396876335, | |
| "learning_rate": 6.995219064973763e-05, | |
| "loss": 0.0151, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.036607266196988536, | |
| "grad_norm": 0.15721233189105988, | |
| "learning_rate": 6.9999621112247e-05, | |
| "loss": 0.1997, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.03674540682414698, | |
| "grad_norm": 0.08665527403354645, | |
| "learning_rate": 7.004687292852334e-05, | |
| "loss": 0.0567, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 0.036883547451305426, | |
| "grad_norm": 0.01629851944744587, | |
| "learning_rate": 7.009394743925609e-05, | |
| "loss": 0.0129, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 0.03702168807846388, | |
| "grad_norm": 0.017079714685678482, | |
| "learning_rate": 7.01408459700988e-05, | |
| "loss": 0.0122, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 0.03715982870562232, | |
| "grad_norm": 0.011573737487196922, | |
| "learning_rate": 7.018756983189322e-05, | |
| "loss": 0.0111, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 0.03729796933278077, | |
| "grad_norm": 0.057232815772295, | |
| "learning_rate": 7.023412032088907e-05, | |
| "loss": 0.0334, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.03743610995993922, | |
| "grad_norm": 0.01176871731877327, | |
| "learning_rate": 7.028049871895989e-05, | |
| "loss": 0.0109, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 0.037574250587097666, | |
| "grad_norm": 0.005038474686443806, | |
| "learning_rate": 7.032670629381475e-05, | |
| "loss": 0.0057, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.03771239121425611, | |
| "grad_norm": 0.02611648477613926, | |
| "learning_rate": 7.037274429920628e-05, | |
| "loss": 0.0212, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 0.03785053184141456, | |
| "grad_norm": 0.04503704607486725, | |
| "learning_rate": 7.04186139751347e-05, | |
| "loss": 0.0316, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 0.03798867246857301, | |
| "grad_norm": 0.02312229387462139, | |
| "learning_rate": 7.046431654804831e-05, | |
| "loss": 0.0195, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.03812681309573145, | |
| "grad_norm": 0.07315024733543396, | |
| "learning_rate": 7.05098532310401e-05, | |
| "loss": 0.0422, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.038264953722889905, | |
| "grad_norm": 0.01006357092410326, | |
| "learning_rate": 7.055522522404115e-05, | |
| "loss": 0.0083, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 0.03840309435004835, | |
| "grad_norm": 0.02410217933356762, | |
| "learning_rate": 7.060043371401022e-05, | |
| "loss": 0.0164, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 0.038541234977206795, | |
| "grad_norm": 0.029703989624977112, | |
| "learning_rate": 7.064547987512007e-05, | |
| "loss": 0.0232, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 0.03867937560436524, | |
| "grad_norm": 0.03558236360549927, | |
| "learning_rate": 7.069036486894051e-05, | |
| "loss": 0.0194, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.03881751623152369, | |
| "grad_norm": 0.004356220830231905, | |
| "learning_rate": 7.073508984461812e-05, | |
| "loss": 0.006, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 0.03895565685868214, | |
| "grad_norm": 0.06183575093746185, | |
| "learning_rate": 7.077965593905269e-05, | |
| "loss": 0.0374, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.03909379748584058, | |
| "grad_norm": 0.03367290645837784, | |
| "learning_rate": 7.082406427707072e-05, | |
| "loss": 0.0147, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 0.039231938112999035, | |
| "grad_norm": 0.05731538310647011, | |
| "learning_rate": 7.086831597159574e-05, | |
| "loss": 0.0267, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 0.03937007874015748, | |
| "grad_norm": 0.013960711658000946, | |
| "learning_rate": 7.091241212381564e-05, | |
| "loss": 0.0115, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.039508219367315925, | |
| "grad_norm": 0.04287987947463989, | |
| "learning_rate": 7.095635382334712e-05, | |
| "loss": 0.0186, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.03964635999447438, | |
| "grad_norm": 0.02721596322953701, | |
| "learning_rate": 7.100014214839724e-05, | |
| "loss": 0.0153, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 0.03978450062163282, | |
| "grad_norm": 0.09616994112730026, | |
| "learning_rate": 7.104377816592214e-05, | |
| "loss": 0.0479, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.03992264124879127, | |
| "grad_norm": 0.01681303046643734, | |
| "learning_rate": 7.108726293178307e-05, | |
| "loss": 0.0099, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 0.04006078187594972, | |
| "grad_norm": 0.007514542900025845, | |
| "learning_rate": 7.113059749089967e-05, | |
| "loss": 0.007, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.040198922503108164, | |
| "grad_norm": 0.00759028410539031, | |
| "learning_rate": 7.117378287740062e-05, | |
| "loss": 0.0093, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 0.04033706313026661, | |
| "grad_norm": 0.04445767030119896, | |
| "learning_rate": 7.12168201147717e-05, | |
| "loss": 0.0192, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 0.04047520375742506, | |
| "grad_norm": 0.004815852735191584, | |
| "learning_rate": 7.125971021600142e-05, | |
| "loss": 0.0058, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 0.04061334438458351, | |
| "grad_norm": 0.01803855411708355, | |
| "learning_rate": 7.130245418372394e-05, | |
| "loss": 0.0189, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 0.04075148501174195, | |
| "grad_norm": 0.022579418495297432, | |
| "learning_rate": 7.13450530103599e-05, | |
| "loss": 0.0124, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.040889625638900404, | |
| "grad_norm": 0.006007243413478136, | |
| "learning_rate": 7.138750767825444e-05, | |
| "loss": 0.0067, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 0.04102776626605885, | |
| "grad_norm": 0.01632005348801613, | |
| "learning_rate": 7.142981915981336e-05, | |
| "loss": 0.0115, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 0.041165906893217294, | |
| "grad_norm": 0.009706880897283554, | |
| "learning_rate": 7.147198841763656e-05, | |
| "loss": 0.0096, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 0.041304047520375746, | |
| "grad_norm": 0.004515258129686117, | |
| "learning_rate": 7.15140164046496e-05, | |
| "loss": 0.0058, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 0.04144218814753419, | |
| "grad_norm": 0.05710865557193756, | |
| "learning_rate": 7.155590406423282e-05, | |
| "loss": 0.0388, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.041580328774692636, | |
| "grad_norm": 0.012932428158819675, | |
| "learning_rate": 7.15976523303484e-05, | |
| "loss": 0.0075, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 0.04171846940185108, | |
| "grad_norm": 0.08358635753393173, | |
| "learning_rate": 7.163926212766535e-05, | |
| "loss": 0.0512, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 0.04185661002900953, | |
| "grad_norm": 0.00845788512378931, | |
| "learning_rate": 7.168073437168235e-05, | |
| "loss": 0.0085, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 0.04199475065616798, | |
| "grad_norm": 0.008571326732635498, | |
| "learning_rate": 7.172206996884871e-05, | |
| "loss": 0.0074, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 0.042132891283326424, | |
| "grad_norm": 0.1380782574415207, | |
| "learning_rate": 7.176326981668306e-05, | |
| "loss": 0.0893, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.042271031910484876, | |
| "grad_norm": 0.10377588123083115, | |
| "learning_rate": 7.180433480389047e-05, | |
| "loss": 0.1076, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 0.04240917253764332, | |
| "grad_norm": 0.005449495278298855, | |
| "learning_rate": 7.184526581047739e-05, | |
| "loss": 0.0056, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 0.042547313164801766, | |
| "grad_norm": 0.016402093693614006, | |
| "learning_rate": 7.18860637078648e-05, | |
| "loss": 0.0144, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 0.04268545379196022, | |
| "grad_norm": 0.057461198419332504, | |
| "learning_rate": 7.192672935899954e-05, | |
| "loss": 0.0335, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 0.04282359441911866, | |
| "grad_norm": 0.018814753741025925, | |
| "learning_rate": 7.196726361846382e-05, | |
| "loss": 0.019, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.04296173504627711, | |
| "grad_norm": 0.021789349615573883, | |
| "learning_rate": 7.200766733258301e-05, | |
| "loss": 0.0174, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 0.04309987567343556, | |
| "grad_norm": 0.05062391981482506, | |
| "learning_rate": 7.204794133953164e-05, | |
| "loss": 0.0256, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.043238016300594005, | |
| "grad_norm": 0.05671432614326477, | |
| "learning_rate": 7.208808646943773e-05, | |
| "loss": 0.0402, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 0.04337615692775245, | |
| "grad_norm": 0.02019777148962021, | |
| "learning_rate": 7.212810354448547e-05, | |
| "loss": 0.0152, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 0.0435142975549109, | |
| "grad_norm": 0.020292505621910095, | |
| "learning_rate": 7.216799337901625e-05, | |
| "loss": 0.0189, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.04365243818206935, | |
| "grad_norm": 0.03928740695118904, | |
| "learning_rate": 7.220775677962808e-05, | |
| "loss": 0.0383, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 0.04379057880922779, | |
| "grad_norm": 0.005962614435702562, | |
| "learning_rate": 7.224739454527347e-05, | |
| "loss": 0.0069, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 0.043928719436386245, | |
| "grad_norm": 0.011562081053853035, | |
| "learning_rate": 7.228690746735578e-05, | |
| "loss": 0.0113, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 0.04406686006354469, | |
| "grad_norm": 0.009331891313195229, | |
| "learning_rate": 7.232629632982394e-05, | |
| "loss": 0.0118, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 0.044205000690703135, | |
| "grad_norm": 0.07156947255134583, | |
| "learning_rate": 7.236556190926588e-05, | |
| "loss": 0.0474, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.04434314131786158, | |
| "grad_norm": 0.0070722345262765884, | |
| "learning_rate": 7.240470497500033e-05, | |
| "loss": 0.0071, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 0.04448128194502003, | |
| "grad_norm": 0.24819540977478027, | |
| "learning_rate": 7.244372628916727e-05, | |
| "loss": 0.0668, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 0.04461942257217848, | |
| "grad_norm": 0.012170674279332161, | |
| "learning_rate": 7.248262660681704e-05, | |
| "loss": 0.0139, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 0.04475756319933692, | |
| "grad_norm": 0.01726563833653927, | |
| "learning_rate": 7.252140667599786e-05, | |
| "loss": 0.0148, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 0.044895703826495374, | |
| "grad_norm": 0.006273672450333834, | |
| "learning_rate": 7.256006723784231e-05, | |
| "loss": 0.0073, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.04503384445365382, | |
| "grad_norm": 0.02530139684677124, | |
| "learning_rate": 7.25986090266522e-05, | |
| "loss": 0.0145, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 0.045171985080812264, | |
| "grad_norm": 0.043121714144945145, | |
| "learning_rate": 7.263703276998225e-05, | |
| "loss": 0.0374, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 0.045310125707970716, | |
| "grad_norm": 0.011094048619270325, | |
| "learning_rate": 7.26753391887226e-05, | |
| "loss": 0.0098, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 0.04544826633512916, | |
| "grad_norm": 0.05844907462596893, | |
| "learning_rate": 7.271352899717988e-05, | |
| "loss": 0.0314, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 0.04558640696228761, | |
| "grad_norm": 0.008745652623474598, | |
| "learning_rate": 7.27516029031571e-05, | |
| "loss": 0.0107, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.04572454758944606, | |
| "grad_norm": 0.12875714898109436, | |
| "learning_rate": 7.278956160803247e-05, | |
| "loss": 0.0946, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 0.045862688216604504, | |
| "grad_norm": 0.09134670346975327, | |
| "learning_rate": 7.282740580683686e-05, | |
| "loss": 0.0767, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 0.04600082884376295, | |
| "grad_norm": 0.01726178452372551, | |
| "learning_rate": 7.286513618833017e-05, | |
| "loss": 0.0131, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 0.0461389694709214, | |
| "grad_norm": 0.006260738708078861, | |
| "learning_rate": 7.290275343507648e-05, | |
| "loss": 0.0069, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 0.046277110098079846, | |
| "grad_norm": 0.02264183759689331, | |
| "learning_rate": 7.294025822351828e-05, | |
| "loss": 0.0157, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.04641525072523829, | |
| "grad_norm": 0.033730894327163696, | |
| "learning_rate": 7.297765122404931e-05, | |
| "loss": 0.0338, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 0.04655339135239674, | |
| "grad_norm": 0.010735332034528255, | |
| "learning_rate": 7.301493310108658e-05, | |
| "loss": 0.0117, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 0.04669153197955519, | |
| "grad_norm": 0.13428980112075806, | |
| "learning_rate": 7.305210451314113e-05, | |
| "loss": 0.0662, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.04682967260671363, | |
| "grad_norm": 0.007416439242660999, | |
| "learning_rate": 7.308916611288784e-05, | |
| "loss": 0.0093, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 0.04696781323387208, | |
| "grad_norm": 0.029140817001461983, | |
| "learning_rate": 7.312611854723422e-05, | |
| "loss": 0.0268, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.04710595386103053, | |
| "grad_norm": 0.002688678679987788, | |
| "learning_rate": 7.316296245738808e-05, | |
| "loss": 0.0041, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 0.047244094488188976, | |
| "grad_norm": 0.09536877274513245, | |
| "learning_rate": 7.319969847892443e-05, | |
| "loss": 0.0513, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 0.04738223511534742, | |
| "grad_norm": 0.0194696132093668, | |
| "learning_rate": 7.323632724185112e-05, | |
| "loss": 0.0156, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 0.04752037574250587, | |
| "grad_norm": 0.012019694782793522, | |
| "learning_rate": 7.327284937067376e-05, | |
| "loss": 0.0128, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 0.04765851636966432, | |
| "grad_norm": 0.07920549809932709, | |
| "learning_rate": 7.330926548445958e-05, | |
| "loss": 0.0496, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.04779665699682276, | |
| "grad_norm": 0.012154892086982727, | |
| "learning_rate": 7.334557619690038e-05, | |
| "loss": 0.0163, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 0.047934797623981215, | |
| "grad_norm": 0.02246297337114811, | |
| "learning_rate": 7.338178211637459e-05, | |
| "loss": 0.016, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 0.04807293825113966, | |
| "grad_norm": 0.1408204585313797, | |
| "learning_rate": 7.341788384600846e-05, | |
| "loss": 0.1492, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 0.048211078878298105, | |
| "grad_norm": 0.03998423367738724, | |
| "learning_rate": 7.345388198373633e-05, | |
| "loss": 0.0386, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 0.04834921950545656, | |
| "grad_norm": 0.0037037734873592854, | |
| "learning_rate": 7.348977712236e-05, | |
| "loss": 0.0052, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.048487360132615, | |
| "grad_norm": 0.01490323431789875, | |
| "learning_rate": 7.352556984960736e-05, | |
| "loss": 0.0154, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 0.04862550075977345, | |
| "grad_norm": 0.022191043943166733, | |
| "learning_rate": 7.356126074819015e-05, | |
| "loss": 0.0173, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 0.0487636413869319, | |
| "grad_norm": 0.016748666763305664, | |
| "learning_rate": 7.359685039586083e-05, | |
| "loss": 0.0147, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 0.048901782014090345, | |
| "grad_norm": 0.018724625930190086, | |
| "learning_rate": 7.363233936546869e-05, | |
| "loss": 0.0196, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 0.04903992264124879, | |
| "grad_norm": 0.04513927921652794, | |
| "learning_rate": 7.366772822501523e-05, | |
| "loss": 0.024, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.04917806326840724, | |
| "grad_norm": 0.017063314095139503, | |
| "learning_rate": 7.370301753770863e-05, | |
| "loss": 0.0164, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 0.04931620389556569, | |
| "grad_norm": 0.01662248931825161, | |
| "learning_rate": 7.373820786201764e-05, | |
| "loss": 0.0148, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 0.04945434452272413, | |
| "grad_norm": 0.023914910852909088, | |
| "learning_rate": 7.377329975172453e-05, | |
| "loss": 0.0205, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 0.04959248514988258, | |
| "grad_norm": 0.01144400890916586, | |
| "learning_rate": 7.380829375597736e-05, | |
| "loss": 0.0135, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 0.04973062577704103, | |
| "grad_norm": 0.04367682710289955, | |
| "learning_rate": 7.384319041934161e-05, | |
| "loss": 0.0376, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.049868766404199474, | |
| "grad_norm": 0.05894111096858978, | |
| "learning_rate": 7.3877990281851e-05, | |
| "loss": 0.0239, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 0.05000690703135792, | |
| "grad_norm": 0.008600580506026745, | |
| "learning_rate": 7.391269387905761e-05, | |
| "loss": 0.01, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 0.05014504765851637, | |
| "grad_norm": 0.012402276508510113, | |
| "learning_rate": 7.394730174208137e-05, | |
| "loss": 0.0165, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 0.05028318828567482, | |
| "grad_norm": 0.017745958641171455, | |
| "learning_rate": 7.398181439765882e-05, | |
| "loss": 0.0218, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.05042132891283326, | |
| "grad_norm": 0.05549019202589989, | |
| "learning_rate": 7.401623236819118e-05, | |
| "loss": 0.0429, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.050559469539991714, | |
| "grad_norm": 0.0038389600813388824, | |
| "learning_rate": 7.405055617179185e-05, | |
| "loss": 0.0056, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 0.05069761016715016, | |
| "grad_norm": 0.14552558958530426, | |
| "learning_rate": 7.408478632233319e-05, | |
| "loss": 0.0786, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 0.050835750794308604, | |
| "grad_norm": 0.011212456971406937, | |
| "learning_rate": 7.411892332949265e-05, | |
| "loss": 0.008, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 0.050973891421467056, | |
| "grad_norm": 0.09112522006034851, | |
| "learning_rate": 7.415296769879832e-05, | |
| "loss": 0.0501, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 0.0511120320486255, | |
| "grad_norm": 0.0417109876871109, | |
| "learning_rate": 7.418691993167391e-05, | |
| "loss": 0.0289, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.051250172675783946, | |
| "grad_norm": 0.03118273988366127, | |
| "learning_rate": 7.422078052548295e-05, | |
| "loss": 0.023, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 0.0513883133029424, | |
| "grad_norm": 0.014401586726307869, | |
| "learning_rate": 7.42545499735726e-05, | |
| "loss": 0.0159, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 0.05152645393010084, | |
| "grad_norm": 0.0387452095746994, | |
| "learning_rate": 7.428822876531674e-05, | |
| "loss": 0.0359, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 0.05166459455725929, | |
| "grad_norm": 0.0016639974201098084, | |
| "learning_rate": 7.43218173861585e-05, | |
| "loss": 0.0028, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 0.05180273518441774, | |
| "grad_norm": 0.007370566017925739, | |
| "learning_rate": 7.435531631765229e-05, | |
| "loss": 0.0091, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.051940875811576186, | |
| "grad_norm": 0.008531290106475353, | |
| "learning_rate": 7.438872603750523e-05, | |
| "loss": 0.0091, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 0.05207901643873463, | |
| "grad_norm": 0.02294515073299408, | |
| "learning_rate": 7.442204701961796e-05, | |
| "loss": 0.0254, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 0.052217157065893076, | |
| "grad_norm": 0.02039843425154686, | |
| "learning_rate": 7.445527973412506e-05, | |
| "loss": 0.0208, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 0.05235529769305153, | |
| "grad_norm": 0.050735872238874435, | |
| "learning_rate": 7.448842464743471e-05, | |
| "loss": 0.0319, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 0.05249343832020997, | |
| "grad_norm": 0.016932478174567223, | |
| "learning_rate": 7.452148222226818e-05, | |
| "loss": 0.0109, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.05263157894736842, | |
| "grad_norm": 0.01010705903172493, | |
| "learning_rate": 7.455445291769839e-05, | |
| "loss": 0.0106, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 0.05276971957452687, | |
| "grad_norm": 0.2523377239704132, | |
| "learning_rate": 7.458733718918828e-05, | |
| "loss": 0.1156, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 0.052907860201685315, | |
| "grad_norm": 0.006449633743613958, | |
| "learning_rate": 7.462013548862848e-05, | |
| "loss": 0.0086, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 0.05304600082884376, | |
| "grad_norm": 0.0557723194360733, | |
| "learning_rate": 7.465284826437468e-05, | |
| "loss": 0.0106, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 0.05318414145600221, | |
| "grad_norm": 0.0064123859629035, | |
| "learning_rate": 7.468547596128427e-05, | |
| "loss": 0.0085, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.05332228208316066, | |
| "grad_norm": 0.009457213804125786, | |
| "learning_rate": 7.471801902075274e-05, | |
| "loss": 0.0117, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 0.0534604227103191, | |
| "grad_norm": 0.0031539765186607838, | |
| "learning_rate": 7.475047788074949e-05, | |
| "loss": 0.0047, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 0.053598563337477555, | |
| "grad_norm": 0.023614773526787758, | |
| "learning_rate": 7.478285297585316e-05, | |
| "loss": 0.0219, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 0.053736703964636, | |
| "grad_norm": 0.10714581608772278, | |
| "learning_rate": 7.481514473728654e-05, | |
| "loss": 0.0712, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 0.053874844591794445, | |
| "grad_norm": 0.014973816461861134, | |
| "learning_rate": 7.484735359295111e-05, | |
| "loss": 0.0115, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.0540129852189529, | |
| "grad_norm": 0.035133246332407, | |
| "learning_rate": 7.487947996746097e-05, | |
| "loss": 0.0189, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 0.05415112584611134, | |
| "grad_norm": 0.01953275129199028, | |
| "learning_rate": 7.491152428217649e-05, | |
| "loss": 0.0255, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 0.05428926647326979, | |
| "grad_norm": 0.012475092895328999, | |
| "learning_rate": 7.494348695523741e-05, | |
| "loss": 0.0117, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 0.05442740710042824, | |
| "grad_norm": 0.018982090055942535, | |
| "learning_rate": 7.497536840159565e-05, | |
| "loss": 0.0186, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 0.054565547727586684, | |
| "grad_norm": 0.0650404840707779, | |
| "learning_rate": 7.500716903304755e-05, | |
| "loss": 0.0505, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.05470368835474513, | |
| "grad_norm": 0.02308955229818821, | |
| "learning_rate": 7.503888925826588e-05, | |
| "loss": 0.0208, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 0.054841828981903575, | |
| "grad_norm": 0.006316799204796553, | |
| "learning_rate": 7.50705294828313e-05, | |
| "loss": 0.009, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 0.05497996960906203, | |
| "grad_norm": 0.03284195438027382, | |
| "learning_rate": 7.510209010926341e-05, | |
| "loss": 0.0286, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 0.05511811023622047, | |
| "grad_norm": 0.03874586522579193, | |
| "learning_rate": 7.51335715370516e-05, | |
| "loss": 0.0282, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 0.05525625086337892, | |
| "grad_norm": 0.017570259049534798, | |
| "learning_rate": 7.516497416268535e-05, | |
| "loss": 0.0159, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.05539439149053737, | |
| "grad_norm": 0.1028120145201683, | |
| "learning_rate": 7.519629837968414e-05, | |
| "loss": 0.0638, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 0.055532532117695814, | |
| "grad_norm": 0.04638221859931946, | |
| "learning_rate": 7.522754457862707e-05, | |
| "loss": 0.0261, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 0.05567067274485426, | |
| "grad_norm": 0.006671587936580181, | |
| "learning_rate": 7.52587131471821e-05, | |
| "loss": 0.0098, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 0.05580881337201271, | |
| "grad_norm": 0.0016172940377146006, | |
| "learning_rate": 7.528980447013491e-05, | |
| "loss": 0.0032, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 0.055946953999171156, | |
| "grad_norm": 0.009872229769825935, | |
| "learning_rate": 7.532081892941734e-05, | |
| "loss": 0.0102, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.0560850946263296, | |
| "grad_norm": 0.014002018608152866, | |
| "learning_rate": 7.535175690413565e-05, | |
| "loss": 0.0126, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 0.05622323525348805, | |
| "grad_norm": 0.01186671108007431, | |
| "learning_rate": 7.538261877059817e-05, | |
| "loss": 0.0151, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 0.0563613758806465, | |
| "grad_norm": 0.010572116822004318, | |
| "learning_rate": 7.541340490234301e-05, | |
| "loss": 0.0128, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 0.056499516507804944, | |
| "grad_norm": 0.006026785355061293, | |
| "learning_rate": 7.544411567016487e-05, | |
| "loss": 0.009, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 0.056637657134963396, | |
| "grad_norm": 0.0200533214956522, | |
| "learning_rate": 7.547475144214207e-05, | |
| "loss": 0.0171, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.05677579776212184, | |
| "grad_norm": 0.016931114718317986, | |
| "learning_rate": 7.550531258366296e-05, | |
| "loss": 0.015, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 0.056913938389280286, | |
| "grad_norm": 0.011858226731419563, | |
| "learning_rate": 7.553579945745208e-05, | |
| "loss": 0.0113, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 0.05705207901643874, | |
| "grad_norm": 0.02081177569925785, | |
| "learning_rate": 7.556621242359586e-05, | |
| "loss": 0.0161, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 0.05719021964359718, | |
| "grad_norm": 0.008650501258671284, | |
| "learning_rate": 7.559655183956836e-05, | |
| "loss": 0.0096, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 0.05732836027075563, | |
| "grad_norm": 0.008785598911345005, | |
| "learning_rate": 7.562681806025635e-05, | |
| "loss": 0.0096, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.05746650089791407, | |
| "grad_norm": 0.004337130580097437, | |
| "learning_rate": 7.565701143798417e-05, | |
| "loss": 0.006, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.057604641525072525, | |
| "grad_norm": 0.01971902698278427, | |
| "learning_rate": 7.568713232253847e-05, | |
| "loss": 0.0144, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 0.05774278215223097, | |
| "grad_norm": 0.003371615894138813, | |
| "learning_rate": 7.571718106119245e-05, | |
| "loss": 0.004, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 0.057880922779389415, | |
| "grad_norm": 0.08240024745464325, | |
| "learning_rate": 7.574715799872985e-05, | |
| "loss": 0.0443, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 0.05801906340654787, | |
| "grad_norm": 0.013157228007912636, | |
| "learning_rate": 7.577706347746878e-05, | |
| "loss": 0.01, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.05815720403370631, | |
| "grad_norm": 0.08836307376623154, | |
| "learning_rate": 7.58068978372851e-05, | |
| "loss": 0.053, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 0.05829534466086476, | |
| "grad_norm": 0.008075368590652943, | |
| "learning_rate": 7.58366614156356e-05, | |
| "loss": 0.0103, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 0.05843348528802321, | |
| "grad_norm": 0.013115230947732925, | |
| "learning_rate": 7.586635454758096e-05, | |
| "loss": 0.0145, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 0.058571625915181655, | |
| "grad_norm": 0.041858140379190445, | |
| "learning_rate": 7.589597756580832e-05, | |
| "loss": 0.0266, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 0.0587097665423401, | |
| "grad_norm": 0.006484493613243103, | |
| "learning_rate": 7.592553080065369e-05, | |
| "loss": 0.007, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.05884790716949855, | |
| "grad_norm": 0.08854290097951889, | |
| "learning_rate": 7.5955014580124e-05, | |
| "loss": 0.0769, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 0.058986047796657, | |
| "grad_norm": 0.03287418931722641, | |
| "learning_rate": 7.598442922991903e-05, | |
| "loss": 0.0359, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 0.05912418842381544, | |
| "grad_norm": 0.0041122897528111935, | |
| "learning_rate": 7.601377507345287e-05, | |
| "loss": 0.0056, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 0.059262329050973894, | |
| "grad_norm": 0.01932491548359394, | |
| "learning_rate": 7.604305243187539e-05, | |
| "loss": 0.0185, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 0.05940046967813234, | |
| "grad_norm": 0.034386295825242996, | |
| "learning_rate": 7.607226162409324e-05, | |
| "loss": 0.0259, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.059538610305290784, | |
| "grad_norm": 0.014645845629274845, | |
| "learning_rate": 7.610140296679071e-05, | |
| "loss": 0.0159, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 0.059676750932449237, | |
| "grad_norm": 0.025981949642300606, | |
| "learning_rate": 7.613047677445041e-05, | |
| "loss": 0.0199, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 0.05981489155960768, | |
| "grad_norm": 0.00898673851042986, | |
| "learning_rate": 7.615948335937346e-05, | |
| "loss": 0.0087, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 0.05995303218676613, | |
| "grad_norm": 0.005743580870330334, | |
| "learning_rate": 7.618842303169978e-05, | |
| "loss": 0.0069, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 0.06009117281392457, | |
| "grad_norm": 0.00816387590020895, | |
| "learning_rate": 7.621729609942794e-05, | |
| "loss": 0.0101, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.060229313441083024, | |
| "grad_norm": 0.021145803853869438, | |
| "learning_rate": 7.624610286843479e-05, | |
| "loss": 0.0145, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 0.06036745406824147, | |
| "grad_norm": 0.01337206270545721, | |
| "learning_rate": 7.627484364249493e-05, | |
| "loss": 0.0123, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 0.060505594695399914, | |
| "grad_norm": 0.03297748416662216, | |
| "learning_rate": 7.630351872329997e-05, | |
| "loss": 0.0201, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 0.060643735322558366, | |
| "grad_norm": 0.05304405465722084, | |
| "learning_rate": 7.633212841047743e-05, | |
| "loss": 0.026, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 0.06078187594971681, | |
| "grad_norm": 0.009639502502977848, | |
| "learning_rate": 7.636067300160962e-05, | |
| "loss": 0.0098, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.060920016576875256, | |
| "grad_norm": 0.08749550580978394, | |
| "learning_rate": 7.638915279225222e-05, | |
| "loss": 0.0489, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 0.06105815720403371, | |
| "grad_norm": 0.0455004945397377, | |
| "learning_rate": 7.64175680759525e-05, | |
| "loss": 0.0276, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.061196297831192153, | |
| "grad_norm": 0.07244568318128586, | |
| "learning_rate": 7.644591914426769e-05, | |
| "loss": 0.0554, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 0.0613344384583506, | |
| "grad_norm": 0.007952029816806316, | |
| "learning_rate": 7.64742062867827e-05, | |
| "loss": 0.0101, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 0.06147257908550905, | |
| "grad_norm": 0.045144565403461456, | |
| "learning_rate": 7.65024297911281e-05, | |
| "loss": 0.0249, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.061610719712667496, | |
| "grad_norm": 0.00880199670791626, | |
| "learning_rate": 7.653058994299745e-05, | |
| "loss": 0.0077, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 0.06174886033982594, | |
| "grad_norm": 0.015727581456303596, | |
| "learning_rate": 7.655868702616483e-05, | |
| "loss": 0.0163, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 0.06188700096698439, | |
| "grad_norm": 0.002647354966029525, | |
| "learning_rate": 7.658672132250185e-05, | |
| "loss": 0.0046, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 0.06202514159414284, | |
| "grad_norm": 0.0048969099298119545, | |
| "learning_rate": 7.661469311199475e-05, | |
| "loss": 0.0063, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 0.06216328222130128, | |
| "grad_norm": 0.010342615656554699, | |
| "learning_rate": 7.664260267276109e-05, | |
| "loss": 0.011, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.062301422848459735, | |
| "grad_norm": 0.004208502359688282, | |
| "learning_rate": 7.667045028106635e-05, | |
| "loss": 0.0059, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 0.06243956347561818, | |
| "grad_norm": 0.012316681444644928, | |
| "learning_rate": 7.66982362113404e-05, | |
| "loss": 0.0106, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 0.06257770410277663, | |
| "grad_norm": 0.006314276251941919, | |
| "learning_rate": 7.672596073619361e-05, | |
| "loss": 0.0086, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 0.06271584472993508, | |
| "grad_norm": 0.004086961969733238, | |
| "learning_rate": 7.675362412643307e-05, | |
| "loss": 0.0052, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 0.06285398535709352, | |
| "grad_norm": 0.048082876950502396, | |
| "learning_rate": 7.678122665107829e-05, | |
| "loss": 0.0333, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.06299212598425197, | |
| "grad_norm": 0.01329890824854374, | |
| "learning_rate": 7.680876857737698e-05, | |
| "loss": 0.0115, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 0.06313026661141041, | |
| "grad_norm": 0.02435714565217495, | |
| "learning_rate": 7.683625017082056e-05, | |
| "loss": 0.0225, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 0.06326840723856886, | |
| "grad_norm": 0.09511169046163559, | |
| "learning_rate": 7.686367169515956e-05, | |
| "loss": 0.084, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 0.06340654786572732, | |
| "grad_norm": 0.01172893587499857, | |
| "learning_rate": 7.689103341241874e-05, | |
| "loss": 0.0115, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 0.06354468849288576, | |
| "grad_norm": 0.022892113775014877, | |
| "learning_rate": 7.691833558291211e-05, | |
| "loss": 0.0145, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.06368282912004421, | |
| "grad_norm": 0.005384576041251421, | |
| "learning_rate": 7.694557846525789e-05, | |
| "loss": 0.0055, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 0.06382096974720265, | |
| "grad_norm": 0.007982897572219372, | |
| "learning_rate": 7.697276231639306e-05, | |
| "loss": 0.009, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 0.0639591103743611, | |
| "grad_norm": 0.014571700245141983, | |
| "learning_rate": 7.699988739158804e-05, | |
| "loss": 0.0134, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 0.06409725100151954, | |
| "grad_norm": 0.048437412828207016, | |
| "learning_rate": 7.7026953944461e-05, | |
| "loss": 0.0285, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 0.06423539162867799, | |
| "grad_norm": 0.1020222008228302, | |
| "learning_rate": 7.705396222699208e-05, | |
| "loss": 0.0604, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.06437353225583645, | |
| "grad_norm": 0.002950117690488696, | |
| "learning_rate": 7.708091248953748e-05, | |
| "loss": 0.0039, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 0.06451167288299489, | |
| "grad_norm": 0.011465366929769516, | |
| "learning_rate": 7.710780498084345e-05, | |
| "loss": 0.0105, | |
| "step": 467 | |
| }, | |
| { | |
| "epoch": 0.06464981351015334, | |
| "grad_norm": 0.009519292041659355, | |
| "learning_rate": 7.71346399480599e-05, | |
| "loss": 0.0091, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 0.06478795413731178, | |
| "grad_norm": 0.009986692108213902, | |
| "learning_rate": 7.716141763675424e-05, | |
| "loss": 0.0111, | |
| "step": 469 | |
| }, | |
| { | |
| "epoch": 0.06492609476447023, | |
| "grad_norm": 0.01423732005059719, | |
| "learning_rate": 7.718813829092471e-05, | |
| "loss": 0.0126, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.06506423539162867, | |
| "grad_norm": 0.008685347624123096, | |
| "learning_rate": 7.721480215301373e-05, | |
| "loss": 0.0114, | |
| "step": 471 | |
| }, | |
| { | |
| "epoch": 0.06520237601878713, | |
| "grad_norm": 0.003593148896470666, | |
| "learning_rate": 7.724140946392123e-05, | |
| "loss": 0.0048, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 0.06534051664594558, | |
| "grad_norm": 0.019208243116736412, | |
| "learning_rate": 7.726796046301751e-05, | |
| "loss": 0.0131, | |
| "step": 473 | |
| }, | |
| { | |
| "epoch": 0.06547865727310402, | |
| "grad_norm": 0.0028865146450698376, | |
| "learning_rate": 7.729445538815635e-05, | |
| "loss": 0.004, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 0.06561679790026247, | |
| "grad_norm": 0.006914031691849232, | |
| "learning_rate": 7.732089447568765e-05, | |
| "loss": 0.0073, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.06575493852742091, | |
| "grad_norm": 0.005969736259430647, | |
| "learning_rate": 7.734727796047018e-05, | |
| "loss": 0.0061, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 0.06589307915457936, | |
| "grad_norm": 0.009544942528009415, | |
| "learning_rate": 7.737360607588405e-05, | |
| "loss": 0.0109, | |
| "step": 477 | |
| }, | |
| { | |
| "epoch": 0.06603121978173782, | |
| "grad_norm": 0.017079075798392296, | |
| "learning_rate": 7.73998790538431e-05, | |
| "loss": 0.0119, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 0.06616936040889626, | |
| "grad_norm": 0.006762394681572914, | |
| "learning_rate": 7.742609712480713e-05, | |
| "loss": 0.006, | |
| "step": 479 | |
| }, | |
| { | |
| "epoch": 0.0663075010360547, | |
| "grad_norm": 0.006828032899647951, | |
| "learning_rate": 7.745226051779415e-05, | |
| "loss": 0.0069, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.06644564166321315, | |
| "grad_norm": 0.008298908360302448, | |
| "learning_rate": 7.74783694603922e-05, | |
| "loss": 0.0083, | |
| "step": 481 | |
| }, | |
| { | |
| "epoch": 0.0665837822903716, | |
| "grad_norm": 0.15002965927124023, | |
| "learning_rate": 7.750442417877138e-05, | |
| "loss": 0.0388, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 0.06672192291753004, | |
| "grad_norm": 0.00954232458025217, | |
| "learning_rate": 7.753042489769554e-05, | |
| "loss": 0.0088, | |
| "step": 483 | |
| }, | |
| { | |
| "epoch": 0.06686006354468849, | |
| "grad_norm": 0.10171414166688919, | |
| "learning_rate": 7.755637184053391e-05, | |
| "loss": 0.058, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 0.06699820417184695, | |
| "grad_norm": 0.0051827989518642426, | |
| "learning_rate": 7.758226522927262e-05, | |
| "loss": 0.0043, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.06713634479900539, | |
| "grad_norm": 0.00380541174672544, | |
| "learning_rate": 7.760810528452614e-05, | |
| "loss": 0.0052, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 0.06727448542616384, | |
| "grad_norm": 0.0701836422085762, | |
| "learning_rate": 7.763389222554847e-05, | |
| "loss": 0.0342, | |
| "step": 487 | |
| }, | |
| { | |
| "epoch": 0.06741262605332228, | |
| "grad_norm": 0.0025863810442388058, | |
| "learning_rate": 7.765962627024439e-05, | |
| "loss": 0.0041, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 0.06755076668048073, | |
| "grad_norm": 0.004156921990215778, | |
| "learning_rate": 7.768530763518046e-05, | |
| "loss": 0.0057, | |
| "step": 489 | |
| }, | |
| { | |
| "epoch": 0.06768890730763917, | |
| "grad_norm": 0.020250679925084114, | |
| "learning_rate": 7.771093653559595e-05, | |
| "loss": 0.0098, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.06782704793479763, | |
| "grad_norm": 0.008420642465353012, | |
| "learning_rate": 7.773651318541372e-05, | |
| "loss": 0.0073, | |
| "step": 491 | |
| }, | |
| { | |
| "epoch": 0.06796518856195607, | |
| "grad_norm": 0.021448403596878052, | |
| "learning_rate": 7.776203779725086e-05, | |
| "loss": 0.0201, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 0.06810332918911452, | |
| "grad_norm": 0.00401209807023406, | |
| "learning_rate": 7.778751058242933e-05, | |
| "loss": 0.0044, | |
| "step": 493 | |
| }, | |
| { | |
| "epoch": 0.06824146981627296, | |
| "grad_norm": 0.004071755334734917, | |
| "learning_rate": 7.781293175098647e-05, | |
| "loss": 0.0048, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 0.06837961044343141, | |
| "grad_norm": 0.07812987267971039, | |
| "learning_rate": 7.783830151168537e-05, | |
| "loss": 0.0596, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.06851775107058986, | |
| "grad_norm": 0.006245663855224848, | |
| "learning_rate": 7.786362007202515e-05, | |
| "loss": 0.0084, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 0.06865589169774831, | |
| "grad_norm": 0.029952548444271088, | |
| "learning_rate": 7.788888763825119e-05, | |
| "loss": 0.0229, | |
| "step": 497 | |
| }, | |
| { | |
| "epoch": 0.06879403232490676, | |
| "grad_norm": 0.00528394291177392, | |
| "learning_rate": 7.791410441536515e-05, | |
| "loss": 0.0072, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 0.0689321729520652, | |
| "grad_norm": 0.004051877185702324, | |
| "learning_rate": 7.793927060713498e-05, | |
| "loss": 0.0049, | |
| "step": 499 | |
| }, | |
| { | |
| "epoch": 0.06907031357922365, | |
| "grad_norm": 0.036640387028455734, | |
| "learning_rate": 7.796438641610483e-05, | |
| "loss": 0.0186, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.0692084542063821, | |
| "grad_norm": 0.004171683453023434, | |
| "learning_rate": 7.798945204360475e-05, | |
| "loss": 0.0067, | |
| "step": 501 | |
| }, | |
| { | |
| "epoch": 0.06934659483354054, | |
| "grad_norm": 0.012148253619670868, | |
| "learning_rate": 7.801446768976046e-05, | |
| "loss": 0.0156, | |
| "step": 502 | |
| }, | |
| { | |
| "epoch": 0.06948473546069898, | |
| "grad_norm": 0.0024077165871858597, | |
| "learning_rate": 7.803943355350285e-05, | |
| "loss": 0.0042, | |
| "step": 503 | |
| }, | |
| { | |
| "epoch": 0.06962287608785744, | |
| "grad_norm": 0.018709257245063782, | |
| "learning_rate": 7.806434983257758e-05, | |
| "loss": 0.0147, | |
| "step": 504 | |
| }, | |
| { | |
| "epoch": 0.06976101671501589, | |
| "grad_norm": 0.004529232159256935, | |
| "learning_rate": 7.808921672355438e-05, | |
| "loss": 0.0051, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 0.06989915734217433, | |
| "grad_norm": 0.08686839044094086, | |
| "learning_rate": 7.811403442183638e-05, | |
| "loss": 0.0485, | |
| "step": 506 | |
| }, | |
| { | |
| "epoch": 0.07003729796933278, | |
| "grad_norm": 0.009920633397996426, | |
| "learning_rate": 7.81388031216694e-05, | |
| "loss": 0.0093, | |
| "step": 507 | |
| }, | |
| { | |
| "epoch": 0.07017543859649122, | |
| "grad_norm": 0.11127305775880814, | |
| "learning_rate": 7.816352301615093e-05, | |
| "loss": 0.0544, | |
| "step": 508 | |
| }, | |
| { | |
| "epoch": 0.07031357922364967, | |
| "grad_norm": 0.010723556391894817, | |
| "learning_rate": 7.818819429723929e-05, | |
| "loss": 0.0121, | |
| "step": 509 | |
| }, | |
| { | |
| "epoch": 0.07045171985080813, | |
| "grad_norm": 0.009187200106680393, | |
| "learning_rate": 7.821281715576248e-05, | |
| "loss": 0.0091, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.07058986047796657, | |
| "grad_norm": 0.012122333981096745, | |
| "learning_rate": 7.823739178142714e-05, | |
| "loss": 0.0117, | |
| "step": 511 | |
| }, | |
| { | |
| "epoch": 0.07072800110512502, | |
| "grad_norm": 0.008281617425382137, | |
| "learning_rate": 7.826191836282722e-05, | |
| "loss": 0.0073, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 0.07086614173228346, | |
| "grad_norm": 0.018169116228818893, | |
| "learning_rate": 7.82863970874527e-05, | |
| "loss": 0.0199, | |
| "step": 513 | |
| }, | |
| { | |
| "epoch": 0.07100428235944191, | |
| "grad_norm": 0.02657570131123066, | |
| "learning_rate": 7.831082814169822e-05, | |
| "loss": 0.0206, | |
| "step": 514 | |
| }, | |
| { | |
| "epoch": 0.07114242298660035, | |
| "grad_norm": 0.006763003766536713, | |
| "learning_rate": 7.833521171087153e-05, | |
| "loss": 0.0081, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 0.07128056361375881, | |
| "grad_norm": 0.14291508495807648, | |
| "learning_rate": 7.835954797920203e-05, | |
| "loss": 0.1127, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 0.07141870424091726, | |
| "grad_norm": 0.015182922594249249, | |
| "learning_rate": 7.838383712984899e-05, | |
| "loss": 0.0183, | |
| "step": 517 | |
| }, | |
| { | |
| "epoch": 0.0715568448680757, | |
| "grad_norm": 0.006437589880079031, | |
| "learning_rate": 7.840807934490987e-05, | |
| "loss": 0.007, | |
| "step": 518 | |
| }, | |
| { | |
| "epoch": 0.07169498549523415, | |
| "grad_norm": 0.006644203793257475, | |
| "learning_rate": 7.843227480542863e-05, | |
| "loss": 0.007, | |
| "step": 519 | |
| }, | |
| { | |
| "epoch": 0.0718331261223926, | |
| "grad_norm": 0.0024770349264144897, | |
| "learning_rate": 7.845642369140365e-05, | |
| "loss": 0.0037, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.07197126674955104, | |
| "grad_norm": 0.007362083997577429, | |
| "learning_rate": 7.848052618179587e-05, | |
| "loss": 0.0083, | |
| "step": 521 | |
| }, | |
| { | |
| "epoch": 0.07210940737670948, | |
| "grad_norm": 0.019205942749977112, | |
| "learning_rate": 7.850458245453673e-05, | |
| "loss": 0.0222, | |
| "step": 522 | |
| }, | |
| { | |
| "epoch": 0.07224754800386794, | |
| "grad_norm": 0.009427196346223354, | |
| "learning_rate": 7.852859268653608e-05, | |
| "loss": 0.0113, | |
| "step": 523 | |
| }, | |
| { | |
| "epoch": 0.07238568863102639, | |
| "grad_norm": 0.006181191653013229, | |
| "learning_rate": 7.855255705368995e-05, | |
| "loss": 0.006, | |
| "step": 524 | |
| }, | |
| { | |
| "epoch": 0.07252382925818483, | |
| "grad_norm": 0.0038815487641841173, | |
| "learning_rate": 7.857647573088825e-05, | |
| "loss": 0.006, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.07266196988534328, | |
| "grad_norm": 0.002488507889211178, | |
| "learning_rate": 7.860034889202254e-05, | |
| "loss": 0.0038, | |
| "step": 526 | |
| }, | |
| { | |
| "epoch": 0.07280011051250172, | |
| "grad_norm": 0.007900571450591087, | |
| "learning_rate": 7.862417670999348e-05, | |
| "loss": 0.0086, | |
| "step": 527 | |
| }, | |
| { | |
| "epoch": 0.07293825113966017, | |
| "grad_norm": 0.025396516546607018, | |
| "learning_rate": 7.864795935671842e-05, | |
| "loss": 0.0152, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 0.07307639176681863, | |
| "grad_norm": 0.03293461725115776, | |
| "learning_rate": 7.867169700313887e-05, | |
| "loss": 0.0226, | |
| "step": 529 | |
| }, | |
| { | |
| "epoch": 0.07321453239397707, | |
| "grad_norm": 0.017437491565942764, | |
| "learning_rate": 7.869538981922779e-05, | |
| "loss": 0.0159, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.07335267302113552, | |
| "grad_norm": 0.005919649265706539, | |
| "learning_rate": 7.871903797399694e-05, | |
| "loss": 0.0073, | |
| "step": 531 | |
| }, | |
| { | |
| "epoch": 0.07349081364829396, | |
| "grad_norm": 0.009071402251720428, | |
| "learning_rate": 7.874264163550414e-05, | |
| "loss": 0.0095, | |
| "step": 532 | |
| }, | |
| { | |
| "epoch": 0.07362895427545241, | |
| "grad_norm": 0.005668788682669401, | |
| "learning_rate": 7.876620097086035e-05, | |
| "loss": 0.0073, | |
| "step": 533 | |
| }, | |
| { | |
| "epoch": 0.07376709490261085, | |
| "grad_norm": 0.02198665961623192, | |
| "learning_rate": 7.87897161462369e-05, | |
| "loss": 0.0133, | |
| "step": 534 | |
| }, | |
| { | |
| "epoch": 0.07390523552976931, | |
| "grad_norm": 0.00984006654471159, | |
| "learning_rate": 7.881318732687234e-05, | |
| "loss": 0.008, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 0.07404337615692776, | |
| "grad_norm": 0.006679135840386152, | |
| "learning_rate": 7.88366146770796e-05, | |
| "loss": 0.0078, | |
| "step": 536 | |
| }, | |
| { | |
| "epoch": 0.0741815167840862, | |
| "grad_norm": 0.016889598220586777, | |
| "learning_rate": 7.88599983602528e-05, | |
| "loss": 0.0151, | |
| "step": 537 | |
| }, | |
| { | |
| "epoch": 0.07431965741124465, | |
| "grad_norm": 0.09604030102491379, | |
| "learning_rate": 7.888333853887401e-05, | |
| "loss": 0.0759, | |
| "step": 538 | |
| }, | |
| { | |
| "epoch": 0.07445779803840309, | |
| "grad_norm": 0.01921008713543415, | |
| "learning_rate": 7.890663537452024e-05, | |
| "loss": 0.0166, | |
| "step": 539 | |
| }, | |
| { | |
| "epoch": 0.07459593866556154, | |
| "grad_norm": 0.008759930729866028, | |
| "learning_rate": 7.892988902786988e-05, | |
| "loss": 0.01, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.07473407929271998, | |
| "grad_norm": 0.0019715612288564444, | |
| "learning_rate": 7.895309965870956e-05, | |
| "loss": 0.0036, | |
| "step": 541 | |
| }, | |
| { | |
| "epoch": 0.07487221991987844, | |
| "grad_norm": 0.012198393233120441, | |
| "learning_rate": 7.897626742594068e-05, | |
| "loss": 0.0103, | |
| "step": 542 | |
| }, | |
| { | |
| "epoch": 0.07501036054703689, | |
| "grad_norm": 0.009106654673814774, | |
| "learning_rate": 7.899939248758587e-05, | |
| "loss": 0.0087, | |
| "step": 543 | |
| }, | |
| { | |
| "epoch": 0.07514850117419533, | |
| "grad_norm": 0.010135483928024769, | |
| "learning_rate": 7.902247500079554e-05, | |
| "loss": 0.0102, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 0.07528664180135378, | |
| "grad_norm": 0.03742365166544914, | |
| "learning_rate": 7.904551512185426e-05, | |
| "loss": 0.016, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 0.07542478242851222, | |
| "grad_norm": 0.004954047035425901, | |
| "learning_rate": 7.906851300618707e-05, | |
| "loss": 0.0065, | |
| "step": 546 | |
| }, | |
| { | |
| "epoch": 0.07556292305567067, | |
| "grad_norm": 0.02959989383816719, | |
| "learning_rate": 7.909146880836586e-05, | |
| "loss": 0.0156, | |
| "step": 547 | |
| }, | |
| { | |
| "epoch": 0.07570106368282913, | |
| "grad_norm": 0.016155797988176346, | |
| "learning_rate": 7.911438268211552e-05, | |
| "loss": 0.0175, | |
| "step": 548 | |
| }, | |
| { | |
| "epoch": 0.07583920430998757, | |
| "grad_norm": 0.007207171525806189, | |
| "learning_rate": 7.913725478032012e-05, | |
| "loss": 0.0074, | |
| "step": 549 | |
| }, | |
| { | |
| "epoch": 0.07597734493714602, | |
| "grad_norm": 0.004485739395022392, | |
| "learning_rate": 7.916008525502909e-05, | |
| "loss": 0.0054, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.07611548556430446, | |
| "grad_norm": 0.02434372715651989, | |
| "learning_rate": 7.91828742574633e-05, | |
| "loss": 0.0146, | |
| "step": 551 | |
| }, | |
| { | |
| "epoch": 0.0762536261914629, | |
| "grad_norm": 0.01013586763292551, | |
| "learning_rate": 7.92056219380209e-05, | |
| "loss": 0.0116, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 0.07639176681862135, | |
| "grad_norm": 0.0093398317694664, | |
| "learning_rate": 7.922832844628351e-05, | |
| "loss": 0.0085, | |
| "step": 553 | |
| }, | |
| { | |
| "epoch": 0.07652990744577981, | |
| "grad_norm": 0.018571248278021812, | |
| "learning_rate": 7.925099393102196e-05, | |
| "loss": 0.0153, | |
| "step": 554 | |
| }, | |
| { | |
| "epoch": 0.07666804807293826, | |
| "grad_norm": 0.006554738152772188, | |
| "learning_rate": 7.927361854020218e-05, | |
| "loss": 0.0079, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 0.0768061887000967, | |
| "grad_norm": 0.004420983139425516, | |
| "learning_rate": 7.929620242099101e-05, | |
| "loss": 0.0057, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 0.07694432932725515, | |
| "grad_norm": 0.0557420551776886, | |
| "learning_rate": 7.931874571976197e-05, | |
| "loss": 0.0338, | |
| "step": 557 | |
| }, | |
| { | |
| "epoch": 0.07708246995441359, | |
| "grad_norm": 0.031253259629011154, | |
| "learning_rate": 7.934124858210087e-05, | |
| "loss": 0.028, | |
| "step": 558 | |
| }, | |
| { | |
| "epoch": 0.07722061058157204, | |
| "grad_norm": 0.005481123924255371, | |
| "learning_rate": 7.936371115281153e-05, | |
| "loss": 0.0075, | |
| "step": 559 | |
| }, | |
| { | |
| "epoch": 0.07735875120873048, | |
| "grad_norm": 0.013984930701553822, | |
| "learning_rate": 7.938613357592132e-05, | |
| "loss": 0.0168, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.07749689183588894, | |
| "grad_norm": 0.052795786410570145, | |
| "learning_rate": 7.940851599468675e-05, | |
| "loss": 0.0454, | |
| "step": 561 | |
| }, | |
| { | |
| "epoch": 0.07763503246304738, | |
| "grad_norm": 0.02815697155892849, | |
| "learning_rate": 7.943085855159892e-05, | |
| "loss": 0.0204, | |
| "step": 562 | |
| }, | |
| { | |
| "epoch": 0.07777317309020583, | |
| "grad_norm": 0.0057924659922719, | |
| "learning_rate": 7.945316138838898e-05, | |
| "loss": 0.0071, | |
| "step": 563 | |
| }, | |
| { | |
| "epoch": 0.07791131371736428, | |
| "grad_norm": 0.015068231150507927, | |
| "learning_rate": 7.94754246460335e-05, | |
| "loss": 0.0121, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 0.07804945434452272, | |
| "grad_norm": 0.006565915886312723, | |
| "learning_rate": 7.949764846475986e-05, | |
| "loss": 0.006, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 0.07818759497168117, | |
| "grad_norm": 0.026122109964489937, | |
| "learning_rate": 7.951983298405152e-05, | |
| "loss": 0.0229, | |
| "step": 566 | |
| }, | |
| { | |
| "epoch": 0.07832573559883962, | |
| "grad_norm": 0.009172670543193817, | |
| "learning_rate": 7.95419783426533e-05, | |
| "loss": 0.0109, | |
| "step": 567 | |
| }, | |
| { | |
| "epoch": 0.07846387622599807, | |
| "grad_norm": 0.012615257874131203, | |
| "learning_rate": 7.956408467857654e-05, | |
| "loss": 0.0137, | |
| "step": 568 | |
| }, | |
| { | |
| "epoch": 0.07860201685315651, | |
| "grad_norm": 0.015784192830324173, | |
| "learning_rate": 7.95861521291043e-05, | |
| "loss": 0.0144, | |
| "step": 569 | |
| }, | |
| { | |
| "epoch": 0.07874015748031496, | |
| "grad_norm": 0.009694431908428669, | |
| "learning_rate": 7.960818083079644e-05, | |
| "loss": 0.0091, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.0788782981074734, | |
| "grad_norm": 0.010170293040573597, | |
| "learning_rate": 7.963017091949477e-05, | |
| "loss": 0.0113, | |
| "step": 571 | |
| }, | |
| { | |
| "epoch": 0.07901643873463185, | |
| "grad_norm": 0.016057651489973068, | |
| "learning_rate": 7.965212253032793e-05, | |
| "loss": 0.0123, | |
| "step": 572 | |
| }, | |
| { | |
| "epoch": 0.07915457936179031, | |
| "grad_norm": 0.002096309559419751, | |
| "learning_rate": 7.967403579771654e-05, | |
| "loss": 0.0037, | |
| "step": 573 | |
| }, | |
| { | |
| "epoch": 0.07929271998894875, | |
| "grad_norm": 0.002688502660021186, | |
| "learning_rate": 7.969591085537804e-05, | |
| "loss": 0.0038, | |
| "step": 574 | |
| }, | |
| { | |
| "epoch": 0.0794308606161072, | |
| "grad_norm": 0.019958576187491417, | |
| "learning_rate": 7.97177478363316e-05, | |
| "loss": 0.0188, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 0.07956900124326564, | |
| "grad_norm": 0.03227323293685913, | |
| "learning_rate": 7.973954687290295e-05, | |
| "loss": 0.0295, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 0.07970714187042409, | |
| "grad_norm": 0.0040475050918757915, | |
| "learning_rate": 7.976130809672927e-05, | |
| "loss": 0.0051, | |
| "step": 577 | |
| }, | |
| { | |
| "epoch": 0.07984528249758253, | |
| "grad_norm": 0.029720835387706757, | |
| "learning_rate": 7.978303163876388e-05, | |
| "loss": 0.012, | |
| "step": 578 | |
| }, | |
| { | |
| "epoch": 0.079983423124741, | |
| "grad_norm": 0.02928655594587326, | |
| "learning_rate": 7.980471762928102e-05, | |
| "loss": 0.019, | |
| "step": 579 | |
| }, | |
| { | |
| "epoch": 0.08012156375189944, | |
| "grad_norm": 0.0031133065931499004, | |
| "learning_rate": 7.982636619788048e-05, | |
| "loss": 0.0047, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.08025970437905788, | |
| "grad_norm": 0.009925676509737968, | |
| "learning_rate": 7.984797747349231e-05, | |
| "loss": 0.0104, | |
| "step": 581 | |
| }, | |
| { | |
| "epoch": 0.08039784500621633, | |
| "grad_norm": 0.013947058469057083, | |
| "learning_rate": 7.986955158438142e-05, | |
| "loss": 0.0148, | |
| "step": 582 | |
| }, | |
| { | |
| "epoch": 0.08053598563337477, | |
| "grad_norm": 0.07294765114784241, | |
| "learning_rate": 7.989108865815207e-05, | |
| "loss": 0.0709, | |
| "step": 583 | |
| }, | |
| { | |
| "epoch": 0.08067412626053322, | |
| "grad_norm": 0.021420951932668686, | |
| "learning_rate": 7.99125888217525e-05, | |
| "loss": 0.0134, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 0.08081226688769166, | |
| "grad_norm": 0.002957011340186, | |
| "learning_rate": 7.993405220147938e-05, | |
| "loss": 0.0042, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 0.08095040751485012, | |
| "grad_norm": 0.0029216560069471598, | |
| "learning_rate": 7.995547892298221e-05, | |
| "loss": 0.0036, | |
| "step": 586 | |
| }, | |
| { | |
| "epoch": 0.08108854814200857, | |
| "grad_norm": 0.010200831107795238, | |
| "learning_rate": 7.997686911126785e-05, | |
| "loss": 0.011, | |
| "step": 587 | |
| }, | |
| { | |
| "epoch": 0.08122668876916701, | |
| "grad_norm": 0.009586147964000702, | |
| "learning_rate": 7.999822289070476e-05, | |
| "loss": 0.0102, | |
| "step": 588 | |
| }, | |
| { | |
| "epoch": 0.08136482939632546, | |
| "grad_norm": 0.024438267573714256, | |
| "learning_rate": 8.001954038502744e-05, | |
| "loss": 0.0247, | |
| "step": 589 | |
| }, | |
| { | |
| "epoch": 0.0815029700234839, | |
| "grad_norm": 0.01673005148768425, | |
| "learning_rate": 8.004082171734068e-05, | |
| "loss": 0.0135, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.08164111065064235, | |
| "grad_norm": 0.03330973908305168, | |
| "learning_rate": 8.006206701012391e-05, | |
| "loss": 0.0231, | |
| "step": 591 | |
| }, | |
| { | |
| "epoch": 0.08177925127780081, | |
| "grad_norm": 0.015957612544298172, | |
| "learning_rate": 8.008327638523524e-05, | |
| "loss": 0.0117, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 0.08191739190495925, | |
| "grad_norm": 0.0030719093047082424, | |
| "learning_rate": 8.010444996391589e-05, | |
| "loss": 0.005, | |
| "step": 593 | |
| }, | |
| { | |
| "epoch": 0.0820555325321177, | |
| "grad_norm": 0.029894646257162094, | |
| "learning_rate": 8.012558786679415e-05, | |
| "loss": 0.0171, | |
| "step": 594 | |
| }, | |
| { | |
| "epoch": 0.08219367315927614, | |
| "grad_norm": 0.0035989086609333754, | |
| "learning_rate": 8.014669021388965e-05, | |
| "loss": 0.0043, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 0.08233181378643459, | |
| "grad_norm": 0.0025325831957161427, | |
| "learning_rate": 8.016775712461736e-05, | |
| "loss": 0.0041, | |
| "step": 596 | |
| }, | |
| { | |
| "epoch": 0.08246995441359303, | |
| "grad_norm": 0.023065388202667236, | |
| "learning_rate": 8.018878871779166e-05, | |
| "loss": 0.0185, | |
| "step": 597 | |
| }, | |
| { | |
| "epoch": 0.08260809504075149, | |
| "grad_norm": 0.003533895593136549, | |
| "learning_rate": 8.02097851116304e-05, | |
| "loss": 0.0057, | |
| "step": 598 | |
| }, | |
| { | |
| "epoch": 0.08274623566790994, | |
| "grad_norm": 0.0017970139160752296, | |
| "learning_rate": 8.023074642375884e-05, | |
| "loss": 0.0034, | |
| "step": 599 | |
| }, | |
| { | |
| "epoch": 0.08288437629506838, | |
| "grad_norm": 0.009932472370564938, | |
| "learning_rate": 8.025167277121362e-05, | |
| "loss": 0.0103, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.08302251692222683, | |
| "grad_norm": 0.010555654764175415, | |
| "learning_rate": 8.02725642704467e-05, | |
| "loss": 0.009, | |
| "step": 601 | |
| }, | |
| { | |
| "epoch": 0.08316065754938527, | |
| "grad_norm": 0.008634793572127819, | |
| "learning_rate": 8.02934210373292e-05, | |
| "loss": 0.0087, | |
| "step": 602 | |
| }, | |
| { | |
| "epoch": 0.08329879817654372, | |
| "grad_norm": 0.010200063697993755, | |
| "learning_rate": 8.031424318715534e-05, | |
| "loss": 0.0088, | |
| "step": 603 | |
| }, | |
| { | |
| "epoch": 0.08343693880370216, | |
| "grad_norm": 0.012205686420202255, | |
| "learning_rate": 8.033503083464615e-05, | |
| "loss": 0.0106, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 0.08357507943086062, | |
| "grad_norm": 0.010989578440785408, | |
| "learning_rate": 8.035578409395338e-05, | |
| "loss": 0.0085, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 0.08371322005801907, | |
| "grad_norm": 0.028562815859913826, | |
| "learning_rate": 8.037650307866316e-05, | |
| "loss": 0.0207, | |
| "step": 606 | |
| }, | |
| { | |
| "epoch": 0.08385136068517751, | |
| "grad_norm": 0.011450926773250103, | |
| "learning_rate": 8.039718790179982e-05, | |
| "loss": 0.0141, | |
| "step": 607 | |
| }, | |
| { | |
| "epoch": 0.08398950131233596, | |
| "grad_norm": 0.0031172465533018112, | |
| "learning_rate": 8.041783867582952e-05, | |
| "loss": 0.0047, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 0.0841276419394944, | |
| "grad_norm": 0.033424630761146545, | |
| "learning_rate": 8.04384555126639e-05, | |
| "loss": 0.0168, | |
| "step": 609 | |
| }, | |
| { | |
| "epoch": 0.08426578256665285, | |
| "grad_norm": 0.033151671290397644, | |
| "learning_rate": 8.045903852366387e-05, | |
| "loss": 0.0197, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.0844039231938113, | |
| "grad_norm": 0.00440793065354228, | |
| "learning_rate": 8.0479587819643e-05, | |
| "loss": 0.0046, | |
| "step": 611 | |
| }, | |
| { | |
| "epoch": 0.08454206382096975, | |
| "grad_norm": 0.019544310867786407, | |
| "learning_rate": 8.050010351087127e-05, | |
| "loss": 0.0152, | |
| "step": 612 | |
| }, | |
| { | |
| "epoch": 0.0846802044481282, | |
| "grad_norm": 0.012192309834063053, | |
| "learning_rate": 8.052058570707858e-05, | |
| "loss": 0.0107, | |
| "step": 613 | |
| }, | |
| { | |
| "epoch": 0.08481834507528664, | |
| "grad_norm": 0.00563014717772603, | |
| "learning_rate": 8.054103451745819e-05, | |
| "loss": 0.0059, | |
| "step": 614 | |
| }, | |
| { | |
| "epoch": 0.08495648570244509, | |
| "grad_norm": 0.006608523894101381, | |
| "learning_rate": 8.056145005067033e-05, | |
| "loss": 0.0072, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 0.08509462632960353, | |
| "grad_norm": 0.006565593648701906, | |
| "learning_rate": 8.05818324148456e-05, | |
| "loss": 0.007, | |
| "step": 616 | |
| }, | |
| { | |
| "epoch": 0.08523276695676199, | |
| "grad_norm": 0.0052340151742100716, | |
| "learning_rate": 8.060218171758839e-05, | |
| "loss": 0.0051, | |
| "step": 617 | |
| }, | |
| { | |
| "epoch": 0.08537090758392044, | |
| "grad_norm": 0.0058824713341891766, | |
| "learning_rate": 8.062249806598032e-05, | |
| "loss": 0.0068, | |
| "step": 618 | |
| }, | |
| { | |
| "epoch": 0.08550904821107888, | |
| "grad_norm": 0.008156881667673588, | |
| "learning_rate": 8.064278156658369e-05, | |
| "loss": 0.0073, | |
| "step": 619 | |
| }, | |
| { | |
| "epoch": 0.08564718883823733, | |
| "grad_norm": 0.003906435798853636, | |
| "learning_rate": 8.066303232544462e-05, | |
| "loss": 0.0044, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.08578532946539577, | |
| "grad_norm": 0.004190659616142511, | |
| "learning_rate": 8.068325044809664e-05, | |
| "loss": 0.0045, | |
| "step": 621 | |
| }, | |
| { | |
| "epoch": 0.08592347009255422, | |
| "grad_norm": 0.004072085488587618, | |
| "learning_rate": 8.070343603956381e-05, | |
| "loss": 0.0048, | |
| "step": 622 | |
| }, | |
| { | |
| "epoch": 0.08606161071971266, | |
| "grad_norm": 0.08516935259103775, | |
| "learning_rate": 8.072358920436406e-05, | |
| "loss": 0.065, | |
| "step": 623 | |
| }, | |
| { | |
| "epoch": 0.08619975134687112, | |
| "grad_norm": 0.12318930774927139, | |
| "learning_rate": 8.074371004651244e-05, | |
| "loss": 0.0462, | |
| "step": 624 | |
| }, | |
| { | |
| "epoch": 0.08633789197402957, | |
| "grad_norm": 0.026846202090382576, | |
| "learning_rate": 8.076379866952429e-05, | |
| "loss": 0.0188, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 0.08647603260118801, | |
| "grad_norm": 0.036953698843717575, | |
| "learning_rate": 8.078385517641853e-05, | |
| "loss": 0.0196, | |
| "step": 626 | |
| }, | |
| { | |
| "epoch": 0.08661417322834646, | |
| "grad_norm": 0.07914821058511734, | |
| "learning_rate": 8.080387966972071e-05, | |
| "loss": 0.0722, | |
| "step": 627 | |
| }, | |
| { | |
| "epoch": 0.0867523138555049, | |
| "grad_norm": 0.05335409194231033, | |
| "learning_rate": 8.082387225146627e-05, | |
| "loss": 0.043, | |
| "step": 628 | |
| }, | |
| { | |
| "epoch": 0.08689045448266335, | |
| "grad_norm": 0.008553317748010159, | |
| "learning_rate": 8.084383302320358e-05, | |
| "loss": 0.0077, | |
| "step": 629 | |
| }, | |
| { | |
| "epoch": 0.0870285951098218, | |
| "grad_norm": 0.005956161767244339, | |
| "learning_rate": 8.086376208599705e-05, | |
| "loss": 0.0067, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.08716673573698025, | |
| "grad_norm": 0.03032430075109005, | |
| "learning_rate": 8.088365954043025e-05, | |
| "loss": 0.021, | |
| "step": 631 | |
| }, | |
| { | |
| "epoch": 0.0873048763641387, | |
| "grad_norm": 0.0037184508983045816, | |
| "learning_rate": 8.090352548660889e-05, | |
| "loss": 0.0053, | |
| "step": 632 | |
| }, | |
| { | |
| "epoch": 0.08744301699129714, | |
| "grad_norm": 0.014337614178657532, | |
| "learning_rate": 8.092336002416385e-05, | |
| "loss": 0.0122, | |
| "step": 633 | |
| }, | |
| { | |
| "epoch": 0.08758115761845559, | |
| "grad_norm": 0.007453493308275938, | |
| "learning_rate": 8.094316325225429e-05, | |
| "loss": 0.0097, | |
| "step": 634 | |
| }, | |
| { | |
| "epoch": 0.08771929824561403, | |
| "grad_norm": 0.019370267167687416, | |
| "learning_rate": 8.096293526957041e-05, | |
| "loss": 0.0165, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 0.08785743887277249, | |
| "grad_norm": 0.01445276290178299, | |
| "learning_rate": 8.098267617433659e-05, | |
| "loss": 0.016, | |
| "step": 636 | |
| }, | |
| { | |
| "epoch": 0.08799557949993093, | |
| "grad_norm": 0.017797252163290977, | |
| "learning_rate": 8.100238606431425e-05, | |
| "loss": 0.0122, | |
| "step": 637 | |
| }, | |
| { | |
| "epoch": 0.08813372012708938, | |
| "grad_norm": 0.014362436719238758, | |
| "learning_rate": 8.102206503680476e-05, | |
| "loss": 0.0126, | |
| "step": 638 | |
| }, | |
| { | |
| "epoch": 0.08827186075424782, | |
| "grad_norm": 0.10174919664859772, | |
| "learning_rate": 8.104171318865227e-05, | |
| "loss": 0.0851, | |
| "step": 639 | |
| }, | |
| { | |
| "epoch": 0.08841000138140627, | |
| "grad_norm": 0.015217412263154984, | |
| "learning_rate": 8.106133061624669e-05, | |
| "loss": 0.0163, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.08854814200856471, | |
| "grad_norm": 0.002396307187154889, | |
| "learning_rate": 8.108091741552639e-05, | |
| "loss": 0.0039, | |
| "step": 641 | |
| }, | |
| { | |
| "epoch": 0.08868628263572316, | |
| "grad_norm": 0.002735432470217347, | |
| "learning_rate": 8.110047368198112e-05, | |
| "loss": 0.0045, | |
| "step": 642 | |
| }, | |
| { | |
| "epoch": 0.08882442326288162, | |
| "grad_norm": 0.007928053848445415, | |
| "learning_rate": 8.111999951065476e-05, | |
| "loss": 0.0095, | |
| "step": 643 | |
| }, | |
| { | |
| "epoch": 0.08896256389004006, | |
| "grad_norm": 0.004873833619058132, | |
| "learning_rate": 8.113949499614808e-05, | |
| "loss": 0.0054, | |
| "step": 644 | |
| }, | |
| { | |
| "epoch": 0.08910070451719851, | |
| "grad_norm": 0.005578899290412664, | |
| "learning_rate": 8.11589602326215e-05, | |
| "loss": 0.0066, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 0.08923884514435695, | |
| "grad_norm": 0.004303886089473963, | |
| "learning_rate": 8.117839531379784e-05, | |
| "loss": 0.0062, | |
| "step": 646 | |
| }, | |
| { | |
| "epoch": 0.0893769857715154, | |
| "grad_norm": 0.038032252341508865, | |
| "learning_rate": 8.1197800332965e-05, | |
| "loss": 0.0369, | |
| "step": 647 | |
| }, | |
| { | |
| "epoch": 0.08951512639867384, | |
| "grad_norm": 0.06186262145638466, | |
| "learning_rate": 8.121717538297867e-05, | |
| "loss": 0.031, | |
| "step": 648 | |
| }, | |
| { | |
| "epoch": 0.0896532670258323, | |
| "grad_norm": 0.029956836253404617, | |
| "learning_rate": 8.123652055626497e-05, | |
| "loss": 0.0184, | |
| "step": 649 | |
| }, | |
| { | |
| "epoch": 0.08979140765299075, | |
| "grad_norm": 0.005330318119376898, | |
| "learning_rate": 8.125583594482312e-05, | |
| "loss": 0.0073, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.0899295482801492, | |
| "grad_norm": 0.006597398314625025, | |
| "learning_rate": 8.127512164022804e-05, | |
| "loss": 0.0071, | |
| "step": 651 | |
| }, | |
| { | |
| "epoch": 0.09006768890730764, | |
| "grad_norm": 0.012712563388049603, | |
| "learning_rate": 8.1294377733633e-05, | |
| "loss": 0.0152, | |
| "step": 652 | |
| }, | |
| { | |
| "epoch": 0.09020582953446608, | |
| "grad_norm": 0.0039291055873036385, | |
| "learning_rate": 8.131360431577212e-05, | |
| "loss": 0.0049, | |
| "step": 653 | |
| }, | |
| { | |
| "epoch": 0.09034397016162453, | |
| "grad_norm": 0.009396882727742195, | |
| "learning_rate": 8.133280147696304e-05, | |
| "loss": 0.0122, | |
| "step": 654 | |
| }, | |
| { | |
| "epoch": 0.09048211078878299, | |
| "grad_norm": 0.003607046790421009, | |
| "learning_rate": 8.13519693071094e-05, | |
| "loss": 0.0062, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 0.09062025141594143, | |
| "grad_norm": 0.060046661645174026, | |
| "learning_rate": 8.137110789570339e-05, | |
| "loss": 0.0501, | |
| "step": 656 | |
| }, | |
| { | |
| "epoch": 0.09075839204309988, | |
| "grad_norm": 0.020894410088658333, | |
| "learning_rate": 8.139021733182823e-05, | |
| "loss": 0.0164, | |
| "step": 657 | |
| }, | |
| { | |
| "epoch": 0.09089653267025832, | |
| "grad_norm": 0.01064255740493536, | |
| "learning_rate": 8.140929770416067e-05, | |
| "loss": 0.0113, | |
| "step": 658 | |
| }, | |
| { | |
| "epoch": 0.09103467329741677, | |
| "grad_norm": 0.01144749391824007, | |
| "learning_rate": 8.142834910097348e-05, | |
| "loss": 0.0116, | |
| "step": 659 | |
| }, | |
| { | |
| "epoch": 0.09117281392457521, | |
| "grad_norm": 0.0037209605798125267, | |
| "learning_rate": 8.144737161013789e-05, | |
| "loss": 0.0042, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.09131095455173366, | |
| "grad_norm": 0.014418653212487698, | |
| "learning_rate": 8.146636531912602e-05, | |
| "loss": 0.019, | |
| "step": 661 | |
| }, | |
| { | |
| "epoch": 0.09144909517889212, | |
| "grad_norm": 0.004248477052897215, | |
| "learning_rate": 8.148533031501328e-05, | |
| "loss": 0.004, | |
| "step": 662 | |
| }, | |
| { | |
| "epoch": 0.09158723580605056, | |
| "grad_norm": 0.023498015478253365, | |
| "learning_rate": 8.150426668448077e-05, | |
| "loss": 0.0156, | |
| "step": 663 | |
| }, | |
| { | |
| "epoch": 0.09172537643320901, | |
| "grad_norm": 0.02856908179819584, | |
| "learning_rate": 8.152317451381767e-05, | |
| "loss": 0.021, | |
| "step": 664 | |
| }, | |
| { | |
| "epoch": 0.09186351706036745, | |
| "grad_norm": 0.03920527175068855, | |
| "learning_rate": 8.154205388892362e-05, | |
| "loss": 0.041, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 0.0920016576875259, | |
| "grad_norm": 0.006108124274760485, | |
| "learning_rate": 8.156090489531097e-05, | |
| "loss": 0.0053, | |
| "step": 666 | |
| }, | |
| { | |
| "epoch": 0.09213979831468434, | |
| "grad_norm": 0.040040936321020126, | |
| "learning_rate": 8.157972761810723e-05, | |
| "loss": 0.0089, | |
| "step": 667 | |
| }, | |
| { | |
| "epoch": 0.0922779389418428, | |
| "grad_norm": 0.008494898676872253, | |
| "learning_rate": 8.159852214205728e-05, | |
| "loss": 0.0117, | |
| "step": 668 | |
| }, | |
| { | |
| "epoch": 0.09241607956900125, | |
| "grad_norm": 0.07821296900510788, | |
| "learning_rate": 8.161728855152573e-05, | |
| "loss": 0.0578, | |
| "step": 669 | |
| }, | |
| { | |
| "epoch": 0.09255422019615969, | |
| "grad_norm": 0.027581755071878433, | |
| "learning_rate": 8.163602693049908e-05, | |
| "loss": 0.0198, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.09269236082331814, | |
| "grad_norm": 0.007122738752514124, | |
| "learning_rate": 8.165473736258814e-05, | |
| "loss": 0.009, | |
| "step": 671 | |
| }, | |
| { | |
| "epoch": 0.09283050145047658, | |
| "grad_norm": 0.028153732419013977, | |
| "learning_rate": 8.167341993103011e-05, | |
| "loss": 0.0241, | |
| "step": 672 | |
| }, | |
| { | |
| "epoch": 0.09296864207763503, | |
| "grad_norm": 0.05543315038084984, | |
| "learning_rate": 8.169207471869094e-05, | |
| "loss": 0.0428, | |
| "step": 673 | |
| }, | |
| { | |
| "epoch": 0.09310678270479349, | |
| "grad_norm": 0.013588045723736286, | |
| "learning_rate": 8.171070180806739e-05, | |
| "loss": 0.0097, | |
| "step": 674 | |
| }, | |
| { | |
| "epoch": 0.09324492333195193, | |
| "grad_norm": 0.009861117228865623, | |
| "learning_rate": 8.172930128128935e-05, | |
| "loss": 0.009, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 0.09338306395911038, | |
| "grad_norm": 0.004336031153798103, | |
| "learning_rate": 8.174787322012195e-05, | |
| "loss": 0.0061, | |
| "step": 676 | |
| }, | |
| { | |
| "epoch": 0.09352120458626882, | |
| "grad_norm": 0.008924839086830616, | |
| "learning_rate": 8.176641770596768e-05, | |
| "loss": 0.0101, | |
| "step": 677 | |
| }, | |
| { | |
| "epoch": 0.09365934521342727, | |
| "grad_norm": 0.005986085161566734, | |
| "learning_rate": 8.178493481986864e-05, | |
| "loss": 0.0083, | |
| "step": 678 | |
| }, | |
| { | |
| "epoch": 0.09379748584058571, | |
| "grad_norm": 0.006522186566144228, | |
| "learning_rate": 8.180342464250859e-05, | |
| "loss": 0.0075, | |
| "step": 679 | |
| }, | |
| { | |
| "epoch": 0.09393562646774416, | |
| "grad_norm": 0.007981918752193451, | |
| "learning_rate": 8.182188725421502e-05, | |
| "loss": 0.0058, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.09407376709490262, | |
| "grad_norm": 0.018126454204320908, | |
| "learning_rate": 8.184032273496133e-05, | |
| "loss": 0.013, | |
| "step": 681 | |
| }, | |
| { | |
| "epoch": 0.09421190772206106, | |
| "grad_norm": 0.0011805971153080463, | |
| "learning_rate": 8.185873116436889e-05, | |
| "loss": 0.0023, | |
| "step": 682 | |
| }, | |
| { | |
| "epoch": 0.0943500483492195, | |
| "grad_norm": 0.016573583707213402, | |
| "learning_rate": 8.187711262170905e-05, | |
| "loss": 0.0152, | |
| "step": 683 | |
| }, | |
| { | |
| "epoch": 0.09448818897637795, | |
| "grad_norm": 0.0179180596023798, | |
| "learning_rate": 8.189546718590522e-05, | |
| "loss": 0.0141, | |
| "step": 684 | |
| }, | |
| { | |
| "epoch": 0.0946263296035364, | |
| "grad_norm": 0.012282337993383408, | |
| "learning_rate": 8.191379493553499e-05, | |
| "loss": 0.0115, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 0.09476447023069484, | |
| "grad_norm": 0.01766025647521019, | |
| "learning_rate": 8.193209594883192e-05, | |
| "loss": 0.0159, | |
| "step": 686 | |
| }, | |
| { | |
| "epoch": 0.0949026108578533, | |
| "grad_norm": 0.002552238991484046, | |
| "learning_rate": 8.195037030368783e-05, | |
| "loss": 0.0041, | |
| "step": 687 | |
| }, | |
| { | |
| "epoch": 0.09504075148501175, | |
| "grad_norm": 0.004465501289814711, | |
| "learning_rate": 8.196861807765457e-05, | |
| "loss": 0.0051, | |
| "step": 688 | |
| }, | |
| { | |
| "epoch": 0.09517889211217019, | |
| "grad_norm": 0.012139561586081982, | |
| "learning_rate": 8.198683934794608e-05, | |
| "loss": 0.0082, | |
| "step": 689 | |
| }, | |
| { | |
| "epoch": 0.09531703273932864, | |
| "grad_norm": 0.010561689734458923, | |
| "learning_rate": 8.200503419144038e-05, | |
| "loss": 0.0114, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.09545517336648708, | |
| "grad_norm": 0.025389693677425385, | |
| "learning_rate": 8.202320268468144e-05, | |
| "loss": 0.0187, | |
| "step": 691 | |
| }, | |
| { | |
| "epoch": 0.09559331399364553, | |
| "grad_norm": 0.0038718711584806442, | |
| "learning_rate": 8.204134490388117e-05, | |
| "loss": 0.0048, | |
| "step": 692 | |
| }, | |
| { | |
| "epoch": 0.09573145462080399, | |
| "grad_norm": 0.007652644068002701, | |
| "learning_rate": 8.205946092492133e-05, | |
| "loss": 0.0082, | |
| "step": 693 | |
| }, | |
| { | |
| "epoch": 0.09586959524796243, | |
| "grad_norm": 0.006432169582694769, | |
| "learning_rate": 8.207755082335539e-05, | |
| "loss": 0.007, | |
| "step": 694 | |
| }, | |
| { | |
| "epoch": 0.09600773587512088, | |
| "grad_norm": 0.0037059388123452663, | |
| "learning_rate": 8.20956146744105e-05, | |
| "loss": 0.0063, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 0.09614587650227932, | |
| "grad_norm": 0.0046118441969156265, | |
| "learning_rate": 8.211365255298927e-05, | |
| "loss": 0.0059, | |
| "step": 696 | |
| }, | |
| { | |
| "epoch": 0.09628401712943777, | |
| "grad_norm": 0.0039950888603925705, | |
| "learning_rate": 8.213166453367173e-05, | |
| "loss": 0.0057, | |
| "step": 697 | |
| }, | |
| { | |
| "epoch": 0.09642215775659621, | |
| "grad_norm": 0.010843550786376, | |
| "learning_rate": 8.214965069071713e-05, | |
| "loss": 0.0119, | |
| "step": 698 | |
| }, | |
| { | |
| "epoch": 0.09656029838375466, | |
| "grad_norm": 0.0019335473189130425, | |
| "learning_rate": 8.216761109806576e-05, | |
| "loss": 0.0034, | |
| "step": 699 | |
| }, | |
| { | |
| "epoch": 0.09669843901091311, | |
| "grad_norm": 0.031584322452545166, | |
| "learning_rate": 8.218554582934079e-05, | |
| "loss": 0.0217, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.09683657963807156, | |
| "grad_norm": 0.001754386001266539, | |
| "learning_rate": 8.220345495785013e-05, | |
| "loss": 0.0031, | |
| "step": 701 | |
| }, | |
| { | |
| "epoch": 0.09697472026523, | |
| "grad_norm": 0.02052963897585869, | |
| "learning_rate": 8.222133855658817e-05, | |
| "loss": 0.0121, | |
| "step": 702 | |
| }, | |
| { | |
| "epoch": 0.09711286089238845, | |
| "grad_norm": 0.02404898777604103, | |
| "learning_rate": 8.223919669823754e-05, | |
| "loss": 0.0218, | |
| "step": 703 | |
| }, | |
| { | |
| "epoch": 0.0972510015195469, | |
| "grad_norm": 0.004977188538759947, | |
| "learning_rate": 8.225702945517095e-05, | |
| "loss": 0.0062, | |
| "step": 704 | |
| }, | |
| { | |
| "epoch": 0.09738914214670534, | |
| "grad_norm": 0.007673116400837898, | |
| "learning_rate": 8.227483689945297e-05, | |
| "loss": 0.0083, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 0.0975272827738638, | |
| "grad_norm": 0.016122104600071907, | |
| "learning_rate": 8.229261910284163e-05, | |
| "loss": 0.0114, | |
| "step": 706 | |
| }, | |
| { | |
| "epoch": 0.09766542340102224, | |
| "grad_norm": 0.0029692344833165407, | |
| "learning_rate": 8.231037613679035e-05, | |
| "loss": 0.0032, | |
| "step": 707 | |
| }, | |
| { | |
| "epoch": 0.09780356402818069, | |
| "grad_norm": 0.016377883031964302, | |
| "learning_rate": 8.232810807244948e-05, | |
| "loss": 0.0107, | |
| "step": 708 | |
| }, | |
| { | |
| "epoch": 0.09794170465533913, | |
| "grad_norm": 0.0028426761273294687, | |
| "learning_rate": 8.23458149806682e-05, | |
| "loss": 0.005, | |
| "step": 709 | |
| }, | |
| { | |
| "epoch": 0.09807984528249758, | |
| "grad_norm": 0.004359726328402758, | |
| "learning_rate": 8.236349693199602e-05, | |
| "loss": 0.0051, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.09821798590965602, | |
| "grad_norm": 0.014035112224519253, | |
| "learning_rate": 8.23811539966846e-05, | |
| "loss": 0.013, | |
| "step": 711 | |
| }, | |
| { | |
| "epoch": 0.09835612653681448, | |
| "grad_norm": 0.0015871514333412051, | |
| "learning_rate": 8.239878624468942e-05, | |
| "loss": 0.0029, | |
| "step": 712 | |
| }, | |
| { | |
| "epoch": 0.09849426716397293, | |
| "grad_norm": 0.0035029149148613214, | |
| "learning_rate": 8.241639374567137e-05, | |
| "loss": 0.0044, | |
| "step": 713 | |
| }, | |
| { | |
| "epoch": 0.09863240779113137, | |
| "grad_norm": 0.008812729269266129, | |
| "learning_rate": 8.243397656899844e-05, | |
| "loss": 0.0065, | |
| "step": 714 | |
| }, | |
| { | |
| "epoch": 0.09877054841828982, | |
| "grad_norm": 0.03456171602010727, | |
| "learning_rate": 8.245153478374738e-05, | |
| "loss": 0.025, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 0.09890868904544826, | |
| "grad_norm": 0.0055639478377997875, | |
| "learning_rate": 8.246906845870534e-05, | |
| "loss": 0.0058, | |
| "step": 716 | |
| }, | |
| { | |
| "epoch": 0.09904682967260671, | |
| "grad_norm": 0.013393756933510303, | |
| "learning_rate": 8.248657766237134e-05, | |
| "loss": 0.011, | |
| "step": 717 | |
| }, | |
| { | |
| "epoch": 0.09918497029976515, | |
| "grad_norm": 0.006130352150648832, | |
| "learning_rate": 8.250406246295815e-05, | |
| "loss": 0.0049, | |
| "step": 718 | |
| }, | |
| { | |
| "epoch": 0.09932311092692361, | |
| "grad_norm": 0.008264213800430298, | |
| "learning_rate": 8.252152292839364e-05, | |
| "loss": 0.0096, | |
| "step": 719 | |
| }, | |
| { | |
| "epoch": 0.09946125155408206, | |
| "grad_norm": 0.05160053074359894, | |
| "learning_rate": 8.253895912632242e-05, | |
| "loss": 0.0319, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.0995993921812405, | |
| "grad_norm": 0.026709305122494698, | |
| "learning_rate": 8.25563711241075e-05, | |
| "loss": 0.018, | |
| "step": 721 | |
| }, | |
| { | |
| "epoch": 0.09973753280839895, | |
| "grad_norm": 0.04814797267317772, | |
| "learning_rate": 8.25737589888318e-05, | |
| "loss": 0.0403, | |
| "step": 722 | |
| }, | |
| { | |
| "epoch": 0.0998756734355574, | |
| "grad_norm": 0.009802248328924179, | |
| "learning_rate": 8.259112278729965e-05, | |
| "loss": 0.0069, | |
| "step": 723 | |
| }, | |
| { | |
| "epoch": 0.10001381406271584, | |
| "grad_norm": 0.004918436054140329, | |
| "learning_rate": 8.260846258603841e-05, | |
| "loss": 0.0061, | |
| "step": 724 | |
| }, | |
| { | |
| "epoch": 0.1001519546898743, | |
| "grad_norm": 0.00288878520950675, | |
| "learning_rate": 8.262577845129994e-05, | |
| "loss": 0.0035, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 0.10029009531703274, | |
| "grad_norm": 0.0023648052010685205, | |
| "learning_rate": 8.264307044906217e-05, | |
| "loss": 0.0035, | |
| "step": 726 | |
| }, | |
| { | |
| "epoch": 0.10042823594419119, | |
| "grad_norm": 0.001214659190736711, | |
| "learning_rate": 8.266033864503055e-05, | |
| "loss": 0.0023, | |
| "step": 727 | |
| }, | |
| { | |
| "epoch": 0.10056637657134963, | |
| "grad_norm": 0.00870110746473074, | |
| "learning_rate": 8.267758310463961e-05, | |
| "loss": 0.0082, | |
| "step": 728 | |
| }, | |
| { | |
| "epoch": 0.10070451719850808, | |
| "grad_norm": 0.001736936392262578, | |
| "learning_rate": 8.26948038930544e-05, | |
| "loss": 0.0032, | |
| "step": 729 | |
| }, | |
| { | |
| "epoch": 0.10084265782566652, | |
| "grad_norm": 0.00857970304787159, | |
| "learning_rate": 8.271200107517198e-05, | |
| "loss": 0.0084, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.10098079845282498, | |
| "grad_norm": 0.0043541984632611275, | |
| "learning_rate": 8.27291747156229e-05, | |
| "loss": 0.0059, | |
| "step": 731 | |
| }, | |
| { | |
| "epoch": 0.10111893907998343, | |
| "grad_norm": 0.017549846321344376, | |
| "learning_rate": 8.274632487877266e-05, | |
| "loss": 0.0156, | |
| "step": 732 | |
| }, | |
| { | |
| "epoch": 0.10125707970714187, | |
| "grad_norm": 0.02662530541419983, | |
| "learning_rate": 8.276345162872311e-05, | |
| "loss": 0.0148, | |
| "step": 733 | |
| }, | |
| { | |
| "epoch": 0.10139522033430032, | |
| "grad_norm": 0.016454922035336494, | |
| "learning_rate": 8.278055502931398e-05, | |
| "loss": 0.0126, | |
| "step": 734 | |
| }, | |
| { | |
| "epoch": 0.10153336096145876, | |
| "grad_norm": 0.0410042479634285, | |
| "learning_rate": 8.279763514412423e-05, | |
| "loss": 0.0273, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 0.10167150158861721, | |
| "grad_norm": 0.03968954086303711, | |
| "learning_rate": 8.281469203647345e-05, | |
| "loss": 0.0259, | |
| "step": 736 | |
| }, | |
| { | |
| "epoch": 0.10180964221577565, | |
| "grad_norm": 0.014899387024343014, | |
| "learning_rate": 8.283172576942335e-05, | |
| "loss": 0.0072, | |
| "step": 737 | |
| }, | |
| { | |
| "epoch": 0.10194778284293411, | |
| "grad_norm": 0.012564278207719326, | |
| "learning_rate": 8.284873640577912e-05, | |
| "loss": 0.0108, | |
| "step": 738 | |
| }, | |
| { | |
| "epoch": 0.10208592347009256, | |
| "grad_norm": 0.0013456420274451375, | |
| "learning_rate": 8.286572400809081e-05, | |
| "loss": 0.0028, | |
| "step": 739 | |
| }, | |
| { | |
| "epoch": 0.102224064097251, | |
| "grad_norm": 0.014120531268417835, | |
| "learning_rate": 8.288268863865471e-05, | |
| "loss": 0.0127, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.10236220472440945, | |
| "grad_norm": 0.0025778827257454395, | |
| "learning_rate": 8.289963035951473e-05, | |
| "loss": 0.0037, | |
| "step": 741 | |
| }, | |
| { | |
| "epoch": 0.10250034535156789, | |
| "grad_norm": 0.06253144890069962, | |
| "learning_rate": 8.291654923246375e-05, | |
| "loss": 0.0239, | |
| "step": 742 | |
| }, | |
| { | |
| "epoch": 0.10263848597872634, | |
| "grad_norm": 0.0070442091673612595, | |
| "learning_rate": 8.293344531904503e-05, | |
| "loss": 0.0078, | |
| "step": 743 | |
| }, | |
| { | |
| "epoch": 0.1027766266058848, | |
| "grad_norm": 0.02459808997809887, | |
| "learning_rate": 8.29503186805534e-05, | |
| "loss": 0.0195, | |
| "step": 744 | |
| }, | |
| { | |
| "epoch": 0.10291476723304324, | |
| "grad_norm": 0.0076606678776443005, | |
| "learning_rate": 8.296716937803683e-05, | |
| "loss": 0.0096, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 0.10305290786020169, | |
| "grad_norm": 0.011984504759311676, | |
| "learning_rate": 8.298399747229752e-05, | |
| "loss": 0.0139, | |
| "step": 746 | |
| }, | |
| { | |
| "epoch": 0.10319104848736013, | |
| "grad_norm": 0.00821708794683218, | |
| "learning_rate": 8.30008030238934e-05, | |
| "loss": 0.0089, | |
| "step": 747 | |
| }, | |
| { | |
| "epoch": 0.10332918911451858, | |
| "grad_norm": 0.03523290902376175, | |
| "learning_rate": 8.30175860931393e-05, | |
| "loss": 0.0264, | |
| "step": 748 | |
| }, | |
| { | |
| "epoch": 0.10346732974167702, | |
| "grad_norm": 0.018691513687372208, | |
| "learning_rate": 8.303434674010831e-05, | |
| "loss": 0.0121, | |
| "step": 749 | |
| }, | |
| { | |
| "epoch": 0.10360547036883548, | |
| "grad_norm": 0.004369445610791445, | |
| "learning_rate": 8.305108502463309e-05, | |
| "loss": 0.0072, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.10374361099599393, | |
| "grad_norm": 0.034502919763326645, | |
| "learning_rate": 8.306780100630714e-05, | |
| "loss": 0.0242, | |
| "step": 751 | |
| }, | |
| { | |
| "epoch": 0.10388175162315237, | |
| "grad_norm": 0.0035339398309588432, | |
| "learning_rate": 8.308449474448603e-05, | |
| "loss": 0.0039, | |
| "step": 752 | |
| }, | |
| { | |
| "epoch": 0.10401989225031082, | |
| "grad_norm": 0.0015408458421006799, | |
| "learning_rate": 8.310116629828871e-05, | |
| "loss": 0.0027, | |
| "step": 753 | |
| }, | |
| { | |
| "epoch": 0.10415803287746926, | |
| "grad_norm": 0.09164151549339294, | |
| "learning_rate": 8.311781572659877e-05, | |
| "loss": 0.0502, | |
| "step": 754 | |
| }, | |
| { | |
| "epoch": 0.1042961735046277, | |
| "grad_norm": 0.009650464169681072, | |
| "learning_rate": 8.313444308806562e-05, | |
| "loss": 0.0087, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 0.10443431413178615, | |
| "grad_norm": 0.038072582334280014, | |
| "learning_rate": 8.315104844110585e-05, | |
| "loss": 0.0405, | |
| "step": 756 | |
| }, | |
| { | |
| "epoch": 0.10457245475894461, | |
| "grad_norm": 0.019803661853075027, | |
| "learning_rate": 8.316763184390432e-05, | |
| "loss": 0.0156, | |
| "step": 757 | |
| }, | |
| { | |
| "epoch": 0.10471059538610306, | |
| "grad_norm": 0.02868686057627201, | |
| "learning_rate": 8.318419335441551e-05, | |
| "loss": 0.0259, | |
| "step": 758 | |
| }, | |
| { | |
| "epoch": 0.1048487360132615, | |
| "grad_norm": 0.04778165742754936, | |
| "learning_rate": 8.320073303036467e-05, | |
| "loss": 0.0396, | |
| "step": 759 | |
| }, | |
| { | |
| "epoch": 0.10498687664041995, | |
| "grad_norm": 0.009861230850219727, | |
| "learning_rate": 8.321725092924898e-05, | |
| "loss": 0.0119, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.10512501726757839, | |
| "grad_norm": 0.014904690906405449, | |
| "learning_rate": 8.32337471083389e-05, | |
| "loss": 0.0121, | |
| "step": 761 | |
| }, | |
| { | |
| "epoch": 0.10526315789473684, | |
| "grad_norm": 0.022462379187345505, | |
| "learning_rate": 8.32502216246792e-05, | |
| "loss": 0.0252, | |
| "step": 762 | |
| }, | |
| { | |
| "epoch": 0.1054012985218953, | |
| "grad_norm": 0.03091195411980152, | |
| "learning_rate": 8.326667453509023e-05, | |
| "loss": 0.0193, | |
| "step": 763 | |
| }, | |
| { | |
| "epoch": 0.10553943914905374, | |
| "grad_norm": 0.00843064021319151, | |
| "learning_rate": 8.328310589616908e-05, | |
| "loss": 0.0073, | |
| "step": 764 | |
| }, | |
| { | |
| "epoch": 0.10567757977621219, | |
| "grad_norm": 0.00403135921806097, | |
| "learning_rate": 8.329951576429075e-05, | |
| "loss": 0.0061, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 0.10581572040337063, | |
| "grad_norm": 0.024939602240920067, | |
| "learning_rate": 8.331590419560928e-05, | |
| "loss": 0.0218, | |
| "step": 766 | |
| }, | |
| { | |
| "epoch": 0.10595386103052908, | |
| "grad_norm": 0.00784094538539648, | |
| "learning_rate": 8.333227124605897e-05, | |
| "loss": 0.0069, | |
| "step": 767 | |
| }, | |
| { | |
| "epoch": 0.10609200165768752, | |
| "grad_norm": 0.014025907963514328, | |
| "learning_rate": 8.334861697135548e-05, | |
| "loss": 0.0129, | |
| "step": 768 | |
| }, | |
| { | |
| "epoch": 0.10623014228484598, | |
| "grad_norm": 0.004523341543972492, | |
| "learning_rate": 8.336494142699693e-05, | |
| "loss": 0.0049, | |
| "step": 769 | |
| }, | |
| { | |
| "epoch": 0.10636828291200442, | |
| "grad_norm": 0.09783921390771866, | |
| "learning_rate": 8.338124466826507e-05, | |
| "loss": 0.0225, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.10650642353916287, | |
| "grad_norm": 0.0024887770414352417, | |
| "learning_rate": 8.339752675022647e-05, | |
| "loss": 0.0039, | |
| "step": 771 | |
| }, | |
| { | |
| "epoch": 0.10664456416632132, | |
| "grad_norm": 0.022068385034799576, | |
| "learning_rate": 8.341378772773356e-05, | |
| "loss": 0.0169, | |
| "step": 772 | |
| }, | |
| { | |
| "epoch": 0.10678270479347976, | |
| "grad_norm": 0.02197118103504181, | |
| "learning_rate": 8.343002765542567e-05, | |
| "loss": 0.0161, | |
| "step": 773 | |
| }, | |
| { | |
| "epoch": 0.1069208454206382, | |
| "grad_norm": 0.004046064335852861, | |
| "learning_rate": 8.34462465877303e-05, | |
| "loss": 0.0058, | |
| "step": 774 | |
| }, | |
| { | |
| "epoch": 0.10705898604779665, | |
| "grad_norm": 0.013067352585494518, | |
| "learning_rate": 8.346244457886409e-05, | |
| "loss": 0.0119, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 0.10719712667495511, | |
| "grad_norm": 0.017200065776705742, | |
| "learning_rate": 8.347862168283396e-05, | |
| "loss": 0.0146, | |
| "step": 776 | |
| }, | |
| { | |
| "epoch": 0.10733526730211355, | |
| "grad_norm": 0.007307767402380705, | |
| "learning_rate": 8.349477795343814e-05, | |
| "loss": 0.0089, | |
| "step": 777 | |
| }, | |
| { | |
| "epoch": 0.107473407929272, | |
| "grad_norm": 0.06693489849567413, | |
| "learning_rate": 8.351091344426733e-05, | |
| "loss": 0.0603, | |
| "step": 778 | |
| }, | |
| { | |
| "epoch": 0.10761154855643044, | |
| "grad_norm": 0.00934355054050684, | |
| "learning_rate": 8.352702820870569e-05, | |
| "loss": 0.0108, | |
| "step": 779 | |
| }, | |
| { | |
| "epoch": 0.10774968918358889, | |
| "grad_norm": 0.0020181015133857727, | |
| "learning_rate": 8.354312229993191e-05, | |
| "loss": 0.0036, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.10788782981074733, | |
| "grad_norm": 0.018208837136626244, | |
| "learning_rate": 8.35591957709203e-05, | |
| "loss": 0.0162, | |
| "step": 781 | |
| }, | |
| { | |
| "epoch": 0.1080259704379058, | |
| "grad_norm": 0.026561802253127098, | |
| "learning_rate": 8.357524867444177e-05, | |
| "loss": 0.0139, | |
| "step": 782 | |
| }, | |
| { | |
| "epoch": 0.10816411106506424, | |
| "grad_norm": 0.03126946836709976, | |
| "learning_rate": 8.359128106306499e-05, | |
| "loss": 0.0312, | |
| "step": 783 | |
| }, | |
| { | |
| "epoch": 0.10830225169222268, | |
| "grad_norm": 0.002232447499409318, | |
| "learning_rate": 8.360729298915728e-05, | |
| "loss": 0.0038, | |
| "step": 784 | |
| }, | |
| { | |
| "epoch": 0.10844039231938113, | |
| "grad_norm": 0.012769825756549835, | |
| "learning_rate": 8.362328450488575e-05, | |
| "loss": 0.0112, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 0.10857853294653957, | |
| "grad_norm": 0.005585674196481705, | |
| "learning_rate": 8.36392556622182e-05, | |
| "loss": 0.0063, | |
| "step": 786 | |
| }, | |
| { | |
| "epoch": 0.10871667357369802, | |
| "grad_norm": 0.0049601225182414055, | |
| "learning_rate": 8.365520651292431e-05, | |
| "loss": 0.0055, | |
| "step": 787 | |
| }, | |
| { | |
| "epoch": 0.10885481420085648, | |
| "grad_norm": 0.008773828856647015, | |
| "learning_rate": 8.367113710857645e-05, | |
| "loss": 0.0102, | |
| "step": 788 | |
| }, | |
| { | |
| "epoch": 0.10899295482801492, | |
| "grad_norm": 0.009176612831652164, | |
| "learning_rate": 8.36870475005508e-05, | |
| "loss": 0.0098, | |
| "step": 789 | |
| }, | |
| { | |
| "epoch": 0.10913109545517337, | |
| "grad_norm": 0.04428212344646454, | |
| "learning_rate": 8.370293774002835e-05, | |
| "loss": 0.0264, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.10926923608233181, | |
| "grad_norm": 0.0029104934073984623, | |
| "learning_rate": 8.371880787799582e-05, | |
| "loss": 0.0039, | |
| "step": 791 | |
| }, | |
| { | |
| "epoch": 0.10940737670949026, | |
| "grad_norm": 0.01047749724239111, | |
| "learning_rate": 8.37346579652467e-05, | |
| "loss": 0.0093, | |
| "step": 792 | |
| }, | |
| { | |
| "epoch": 0.1095455173366487, | |
| "grad_norm": 0.01725156232714653, | |
| "learning_rate": 8.375048805238215e-05, | |
| "loss": 0.0174, | |
| "step": 793 | |
| }, | |
| { | |
| "epoch": 0.10968365796380715, | |
| "grad_norm": 0.0077973040752112865, | |
| "learning_rate": 8.37662981898121e-05, | |
| "loss": 0.0073, | |
| "step": 794 | |
| }, | |
| { | |
| "epoch": 0.10982179859096561, | |
| "grad_norm": 0.0034406129270792007, | |
| "learning_rate": 8.378208842775606e-05, | |
| "loss": 0.0047, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 0.10995993921812405, | |
| "grad_norm": 0.004409421235322952, | |
| "learning_rate": 8.37978588162442e-05, | |
| "loss": 0.0052, | |
| "step": 796 | |
| }, | |
| { | |
| "epoch": 0.1100980798452825, | |
| "grad_norm": 0.020356660708785057, | |
| "learning_rate": 8.381360940511824e-05, | |
| "loss": 0.0189, | |
| "step": 797 | |
| }, | |
| { | |
| "epoch": 0.11023622047244094, | |
| "grad_norm": 0.009040805511176586, | |
| "learning_rate": 8.382934024403241e-05, | |
| "loss": 0.0075, | |
| "step": 798 | |
| }, | |
| { | |
| "epoch": 0.11037436109959939, | |
| "grad_norm": 0.004525200929492712, | |
| "learning_rate": 8.384505138245437e-05, | |
| "loss": 0.0055, | |
| "step": 799 | |
| }, | |
| { | |
| "epoch": 0.11051250172675783, | |
| "grad_norm": 0.0028125497046858072, | |
| "learning_rate": 8.386074286966616e-05, | |
| "loss": 0.0045, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.11065064235391629, | |
| "grad_norm": 0.007167664356529713, | |
| "learning_rate": 8.387641475476517e-05, | |
| "loss": 0.0089, | |
| "step": 801 | |
| }, | |
| { | |
| "epoch": 0.11078878298107474, | |
| "grad_norm": 0.02145918644964695, | |
| "learning_rate": 8.389206708666495e-05, | |
| "loss": 0.0109, | |
| "step": 802 | |
| }, | |
| { | |
| "epoch": 0.11092692360823318, | |
| "grad_norm": 0.07421177625656128, | |
| "learning_rate": 8.390769991409625e-05, | |
| "loss": 0.0339, | |
| "step": 803 | |
| }, | |
| { | |
| "epoch": 0.11106506423539163, | |
| "grad_norm": 0.016273144632577896, | |
| "learning_rate": 8.392331328560788e-05, | |
| "loss": 0.0138, | |
| "step": 804 | |
| }, | |
| { | |
| "epoch": 0.11120320486255007, | |
| "grad_norm": 0.006689921021461487, | |
| "learning_rate": 8.393890724956755e-05, | |
| "loss": 0.0051, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 0.11134134548970852, | |
| "grad_norm": 0.013465960510075092, | |
| "learning_rate": 8.39544818541629e-05, | |
| "loss": 0.0142, | |
| "step": 806 | |
| }, | |
| { | |
| "epoch": 0.11147948611686698, | |
| "grad_norm": 0.003067552577704191, | |
| "learning_rate": 8.39700371474023e-05, | |
| "loss": 0.0035, | |
| "step": 807 | |
| }, | |
| { | |
| "epoch": 0.11161762674402542, | |
| "grad_norm": 0.002693150658160448, | |
| "learning_rate": 8.398557317711569e-05, | |
| "loss": 0.0042, | |
| "step": 808 | |
| }, | |
| { | |
| "epoch": 0.11175576737118387, | |
| "grad_norm": 0.0046765501610934734, | |
| "learning_rate": 8.400108999095567e-05, | |
| "loss": 0.0049, | |
| "step": 809 | |
| }, | |
| { | |
| "epoch": 0.11189390799834231, | |
| "grad_norm": 0.004729992710053921, | |
| "learning_rate": 8.401658763639814e-05, | |
| "loss": 0.0059, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.11203204862550076, | |
| "grad_norm": 0.011979191564023495, | |
| "learning_rate": 8.403206616074328e-05, | |
| "loss": 0.015, | |
| "step": 811 | |
| }, | |
| { | |
| "epoch": 0.1121701892526592, | |
| "grad_norm": 0.002755200956016779, | |
| "learning_rate": 8.404752561111644e-05, | |
| "loss": 0.0038, | |
| "step": 812 | |
| }, | |
| { | |
| "epoch": 0.11230832987981765, | |
| "grad_norm": 0.010400881990790367, | |
| "learning_rate": 8.406296603446894e-05, | |
| "loss": 0.0117, | |
| "step": 813 | |
| }, | |
| { | |
| "epoch": 0.1124464705069761, | |
| "grad_norm": 0.006119609344750643, | |
| "learning_rate": 8.407838747757898e-05, | |
| "loss": 0.0071, | |
| "step": 814 | |
| }, | |
| { | |
| "epoch": 0.11258461113413455, | |
| "grad_norm": 0.005961157381534576, | |
| "learning_rate": 8.409378998705247e-05, | |
| "loss": 0.0072, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 0.112722751761293, | |
| "grad_norm": 0.0021235195454210043, | |
| "learning_rate": 8.41091736093238e-05, | |
| "loss": 0.0033, | |
| "step": 816 | |
| }, | |
| { | |
| "epoch": 0.11286089238845144, | |
| "grad_norm": 0.004867125768214464, | |
| "learning_rate": 8.412453839065685e-05, | |
| "loss": 0.0047, | |
| "step": 817 | |
| }, | |
| { | |
| "epoch": 0.11299903301560989, | |
| "grad_norm": 0.021859489381313324, | |
| "learning_rate": 8.413988437714566e-05, | |
| "loss": 0.0218, | |
| "step": 818 | |
| }, | |
| { | |
| "epoch": 0.11313717364276833, | |
| "grad_norm": 0.004767335951328278, | |
| "learning_rate": 8.415521161471534e-05, | |
| "loss": 0.005, | |
| "step": 819 | |
| }, | |
| { | |
| "epoch": 0.11327531426992679, | |
| "grad_norm": 0.0023131791967898607, | |
| "learning_rate": 8.417052014912287e-05, | |
| "loss": 0.003, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.11341345489708524, | |
| "grad_norm": 0.001652119099162519, | |
| "learning_rate": 8.418581002595795e-05, | |
| "loss": 0.003, | |
| "step": 821 | |
| }, | |
| { | |
| "epoch": 0.11355159552424368, | |
| "grad_norm": 0.06345450133085251, | |
| "learning_rate": 8.420108129064377e-05, | |
| "loss": 0.0511, | |
| "step": 822 | |
| }, | |
| { | |
| "epoch": 0.11368973615140213, | |
| "grad_norm": 0.001477993791922927, | |
| "learning_rate": 8.421633398843785e-05, | |
| "loss": 0.0027, | |
| "step": 823 | |
| }, | |
| { | |
| "epoch": 0.11382787677856057, | |
| "grad_norm": 0.010723377577960491, | |
| "learning_rate": 8.423156816443287e-05, | |
| "loss": 0.0119, | |
| "step": 824 | |
| }, | |
| { | |
| "epoch": 0.11396601740571902, | |
| "grad_norm": 0.005179021507501602, | |
| "learning_rate": 8.424678386355736e-05, | |
| "loss": 0.0044, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 0.11410415803287748, | |
| "grad_norm": 0.004340772982686758, | |
| "learning_rate": 8.426198113057666e-05, | |
| "loss": 0.0051, | |
| "step": 826 | |
| }, | |
| { | |
| "epoch": 0.11424229866003592, | |
| "grad_norm": 0.011915227398276329, | |
| "learning_rate": 8.427716001009356e-05, | |
| "loss": 0.0123, | |
| "step": 827 | |
| }, | |
| { | |
| "epoch": 0.11438043928719437, | |
| "grad_norm": 0.006282525602728128, | |
| "learning_rate": 8.429232054654917e-05, | |
| "loss": 0.0064, | |
| "step": 828 | |
| }, | |
| { | |
| "epoch": 0.11451857991435281, | |
| "grad_norm": 0.10944747924804688, | |
| "learning_rate": 8.43074627842237e-05, | |
| "loss": 0.0614, | |
| "step": 829 | |
| }, | |
| { | |
| "epoch": 0.11465672054151126, | |
| "grad_norm": 0.002959243021905422, | |
| "learning_rate": 8.432258676723714e-05, | |
| "loss": 0.0042, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.1147948611686697, | |
| "grad_norm": 0.010976544581353664, | |
| "learning_rate": 8.433769253955023e-05, | |
| "loss": 0.0088, | |
| "step": 831 | |
| }, | |
| { | |
| "epoch": 0.11493300179582815, | |
| "grad_norm": 0.015172791667282581, | |
| "learning_rate": 8.435278014496498e-05, | |
| "loss": 0.0099, | |
| "step": 832 | |
| }, | |
| { | |
| "epoch": 0.1150711424229866, | |
| "grad_norm": 0.0273448433727026, | |
| "learning_rate": 8.436784962712563e-05, | |
| "loss": 0.0216, | |
| "step": 833 | |
| }, | |
| { | |
| "epoch": 0.11520928305014505, | |
| "grad_norm": 0.05054100230336189, | |
| "learning_rate": 8.438290102951929e-05, | |
| "loss": 0.0301, | |
| "step": 834 | |
| }, | |
| { | |
| "epoch": 0.1153474236773035, | |
| "grad_norm": 0.006080084480345249, | |
| "learning_rate": 8.439793439547675e-05, | |
| "loss": 0.0086, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 0.11548556430446194, | |
| "grad_norm": 0.008105085231363773, | |
| "learning_rate": 8.441294976817325e-05, | |
| "loss": 0.0088, | |
| "step": 836 | |
| }, | |
| { | |
| "epoch": 0.11562370493162039, | |
| "grad_norm": 0.030704544857144356, | |
| "learning_rate": 8.442794719062914e-05, | |
| "loss": 0.029, | |
| "step": 837 | |
| }, | |
| { | |
| "epoch": 0.11576184555877883, | |
| "grad_norm": 0.09055589139461517, | |
| "learning_rate": 8.444292670571065e-05, | |
| "loss": 0.0448, | |
| "step": 838 | |
| }, | |
| { | |
| "epoch": 0.11589998618593729, | |
| "grad_norm": 0.03190995752811432, | |
| "learning_rate": 8.445788835613072e-05, | |
| "loss": 0.0238, | |
| "step": 839 | |
| }, | |
| { | |
| "epoch": 0.11603812681309573, | |
| "grad_norm": 0.0042504677549004555, | |
| "learning_rate": 8.447283218444959e-05, | |
| "loss": 0.0047, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.11617626744025418, | |
| "grad_norm": 0.006505970843136311, | |
| "learning_rate": 8.44877582330756e-05, | |
| "loss": 0.0067, | |
| "step": 841 | |
| }, | |
| { | |
| "epoch": 0.11631440806741263, | |
| "grad_norm": 0.0037945907097309828, | |
| "learning_rate": 8.450266654426589e-05, | |
| "loss": 0.0043, | |
| "step": 842 | |
| }, | |
| { | |
| "epoch": 0.11645254869457107, | |
| "grad_norm": 0.14548099040985107, | |
| "learning_rate": 8.451755716012719e-05, | |
| "loss": 0.2458, | |
| "step": 843 | |
| }, | |
| { | |
| "epoch": 0.11659068932172952, | |
| "grad_norm": 0.01771988905966282, | |
| "learning_rate": 8.453243012261641e-05, | |
| "loss": 0.0138, | |
| "step": 844 | |
| }, | |
| { | |
| "epoch": 0.11672882994888797, | |
| "grad_norm": 0.01083680521696806, | |
| "learning_rate": 8.454728547354141e-05, | |
| "loss": 0.0101, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 0.11686697057604642, | |
| "grad_norm": 0.006191121879965067, | |
| "learning_rate": 8.456212325456176e-05, | |
| "loss": 0.007, | |
| "step": 846 | |
| }, | |
| { | |
| "epoch": 0.11700511120320486, | |
| "grad_norm": 0.035239238291978836, | |
| "learning_rate": 8.457694350718935e-05, | |
| "loss": 0.0178, | |
| "step": 847 | |
| }, | |
| { | |
| "epoch": 0.11714325183036331, | |
| "grad_norm": 0.020794779062271118, | |
| "learning_rate": 8.459174627278912e-05, | |
| "loss": 0.0143, | |
| "step": 848 | |
| }, | |
| { | |
| "epoch": 0.11728139245752175, | |
| "grad_norm": 0.0027690723072737455, | |
| "learning_rate": 8.460653159257979e-05, | |
| "loss": 0.0043, | |
| "step": 849 | |
| }, | |
| { | |
| "epoch": 0.1174195330846802, | |
| "grad_norm": 0.011040976271033287, | |
| "learning_rate": 8.46212995076345e-05, | |
| "loss": 0.0097, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.11755767371183865, | |
| "grad_norm": 0.05898001417517662, | |
| "learning_rate": 8.463605005888147e-05, | |
| "loss": 0.0432, | |
| "step": 851 | |
| }, | |
| { | |
| "epoch": 0.1176958143389971, | |
| "grad_norm": 0.03841552138328552, | |
| "learning_rate": 8.465078328710481e-05, | |
| "loss": 0.0265, | |
| "step": 852 | |
| }, | |
| { | |
| "epoch": 0.11783395496615555, | |
| "grad_norm": 0.09727146476507187, | |
| "learning_rate": 8.466549923294504e-05, | |
| "loss": 0.062, | |
| "step": 853 | |
| }, | |
| { | |
| "epoch": 0.117972095593314, | |
| "grad_norm": 0.006218411959707737, | |
| "learning_rate": 8.468019793689983e-05, | |
| "loss": 0.0076, | |
| "step": 854 | |
| }, | |
| { | |
| "epoch": 0.11811023622047244, | |
| "grad_norm": 0.008461462333798409, | |
| "learning_rate": 8.469487943932471e-05, | |
| "loss": 0.0089, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 0.11824837684763088, | |
| "grad_norm": 0.011293224059045315, | |
| "learning_rate": 8.470954378043366e-05, | |
| "loss": 0.0106, | |
| "step": 856 | |
| }, | |
| { | |
| "epoch": 0.11838651747478933, | |
| "grad_norm": 0.001236687763594091, | |
| "learning_rate": 8.472419100029984e-05, | |
| "loss": 0.0026, | |
| "step": 857 | |
| }, | |
| { | |
| "epoch": 0.11852465810194779, | |
| "grad_norm": 0.017730748280882835, | |
| "learning_rate": 8.473882113885618e-05, | |
| "loss": 0.0171, | |
| "step": 858 | |
| }, | |
| { | |
| "epoch": 0.11866279872910623, | |
| "grad_norm": 0.017721619457006454, | |
| "learning_rate": 8.475343423589608e-05, | |
| "loss": 0.0151, | |
| "step": 859 | |
| }, | |
| { | |
| "epoch": 0.11880093935626468, | |
| "grad_norm": 0.005204800516366959, | |
| "learning_rate": 8.476803033107404e-05, | |
| "loss": 0.0048, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.11893907998342312, | |
| "grad_norm": 0.013895371928811073, | |
| "learning_rate": 8.47826094639063e-05, | |
| "loss": 0.0091, | |
| "step": 861 | |
| }, | |
| { | |
| "epoch": 0.11907722061058157, | |
| "grad_norm": 0.046134259551763535, | |
| "learning_rate": 8.479717167377152e-05, | |
| "loss": 0.0408, | |
| "step": 862 | |
| }, | |
| { | |
| "epoch": 0.11921536123774001, | |
| "grad_norm": 0.00713316909968853, | |
| "learning_rate": 8.481171699991137e-05, | |
| "loss": 0.0094, | |
| "step": 863 | |
| }, | |
| { | |
| "epoch": 0.11935350186489847, | |
| "grad_norm": 0.08703207969665527, | |
| "learning_rate": 8.48262454814312e-05, | |
| "loss": 0.0581, | |
| "step": 864 | |
| }, | |
| { | |
| "epoch": 0.11949164249205692, | |
| "grad_norm": 0.0035085680428892374, | |
| "learning_rate": 8.484075715730064e-05, | |
| "loss": 0.0053, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 0.11962978311921536, | |
| "grad_norm": 0.01007845625281334, | |
| "learning_rate": 8.485525206635425e-05, | |
| "loss": 0.0089, | |
| "step": 866 | |
| }, | |
| { | |
| "epoch": 0.11976792374637381, | |
| "grad_norm": 0.02679300867021084, | |
| "learning_rate": 8.486973024729214e-05, | |
| "loss": 0.0161, | |
| "step": 867 | |
| }, | |
| { | |
| "epoch": 0.11990606437353225, | |
| "grad_norm": 0.006040680222213268, | |
| "learning_rate": 8.488419173868058e-05, | |
| "loss": 0.0079, | |
| "step": 868 | |
| }, | |
| { | |
| "epoch": 0.1200442050006907, | |
| "grad_norm": 0.004664016887545586, | |
| "learning_rate": 8.489863657895262e-05, | |
| "loss": 0.006, | |
| "step": 869 | |
| }, | |
| { | |
| "epoch": 0.12018234562784914, | |
| "grad_norm": 0.01343533769249916, | |
| "learning_rate": 8.491306480640874e-05, | |
| "loss": 0.0146, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.1203204862550076, | |
| "grad_norm": 0.019741838797926903, | |
| "learning_rate": 8.492747645921737e-05, | |
| "loss": 0.0166, | |
| "step": 871 | |
| }, | |
| { | |
| "epoch": 0.12045862688216605, | |
| "grad_norm": 0.01905057206749916, | |
| "learning_rate": 8.494187157541558e-05, | |
| "loss": 0.0118, | |
| "step": 872 | |
| }, | |
| { | |
| "epoch": 0.12059676750932449, | |
| "grad_norm": 0.02122565358877182, | |
| "learning_rate": 8.495625019290968e-05, | |
| "loss": 0.0171, | |
| "step": 873 | |
| }, | |
| { | |
| "epoch": 0.12073490813648294, | |
| "grad_norm": 0.006704521831125021, | |
| "learning_rate": 8.497061234947573e-05, | |
| "loss": 0.0061, | |
| "step": 874 | |
| }, | |
| { | |
| "epoch": 0.12087304876364138, | |
| "grad_norm": 0.0038077067583799362, | |
| "learning_rate": 8.498495808276026e-05, | |
| "loss": 0.005, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 0.12101118939079983, | |
| "grad_norm": 0.036112312227487564, | |
| "learning_rate": 8.499928743028077e-05, | |
| "loss": 0.0333, | |
| "step": 876 | |
| }, | |
| { | |
| "epoch": 0.12114933001795829, | |
| "grad_norm": 0.005154281854629517, | |
| "learning_rate": 8.501360042942633e-05, | |
| "loss": 0.0075, | |
| "step": 877 | |
| }, | |
| { | |
| "epoch": 0.12128747064511673, | |
| "grad_norm": 0.0023703796323388815, | |
| "learning_rate": 8.502789711745823e-05, | |
| "loss": 0.0042, | |
| "step": 878 | |
| }, | |
| { | |
| "epoch": 0.12142561127227518, | |
| "grad_norm": 0.03203393518924713, | |
| "learning_rate": 8.504217753151046e-05, | |
| "loss": 0.0235, | |
| "step": 879 | |
| }, | |
| { | |
| "epoch": 0.12156375189943362, | |
| "grad_norm": 0.02325097844004631, | |
| "learning_rate": 8.505644170859043e-05, | |
| "loss": 0.0168, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.12170189252659207, | |
| "grad_norm": 0.04458438232541084, | |
| "learning_rate": 8.507068968557937e-05, | |
| "loss": 0.0298, | |
| "step": 881 | |
| }, | |
| { | |
| "epoch": 0.12184003315375051, | |
| "grad_norm": 0.022350402548909187, | |
| "learning_rate": 8.508492149923301e-05, | |
| "loss": 0.0173, | |
| "step": 882 | |
| }, | |
| { | |
| "epoch": 0.12197817378090897, | |
| "grad_norm": 0.0038661733269691467, | |
| "learning_rate": 8.509913718618219e-05, | |
| "loss": 0.0058, | |
| "step": 883 | |
| }, | |
| { | |
| "epoch": 0.12211631440806742, | |
| "grad_norm": 0.03452976420521736, | |
| "learning_rate": 8.511333678293331e-05, | |
| "loss": 0.0153, | |
| "step": 884 | |
| }, | |
| { | |
| "epoch": 0.12225445503522586, | |
| "grad_norm": 0.00885214377194643, | |
| "learning_rate": 8.512752032586895e-05, | |
| "loss": 0.0092, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 0.12239259566238431, | |
| "grad_norm": 0.01287474762648344, | |
| "learning_rate": 8.514168785124848e-05, | |
| "loss": 0.0101, | |
| "step": 886 | |
| }, | |
| { | |
| "epoch": 0.12253073628954275, | |
| "grad_norm": 0.0044240327551960945, | |
| "learning_rate": 8.51558393952085e-05, | |
| "loss": 0.004, | |
| "step": 887 | |
| }, | |
| { | |
| "epoch": 0.1226688769167012, | |
| "grad_norm": 0.0011661675525829196, | |
| "learning_rate": 8.516997499376351e-05, | |
| "loss": 0.0023, | |
| "step": 888 | |
| }, | |
| { | |
| "epoch": 0.12280701754385964, | |
| "grad_norm": 0.014828033745288849, | |
| "learning_rate": 8.518409468280637e-05, | |
| "loss": 0.0093, | |
| "step": 889 | |
| }, | |
| { | |
| "epoch": 0.1229451581710181, | |
| "grad_norm": 0.006530273240059614, | |
| "learning_rate": 8.51981984981089e-05, | |
| "loss": 0.0072, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.12308329879817655, | |
| "grad_norm": 0.010840199887752533, | |
| "learning_rate": 8.521228647532241e-05, | |
| "loss": 0.0099, | |
| "step": 891 | |
| }, | |
| { | |
| "epoch": 0.12322143942533499, | |
| "grad_norm": 0.012677641585469246, | |
| "learning_rate": 8.522635864997826e-05, | |
| "loss": 0.0098, | |
| "step": 892 | |
| }, | |
| { | |
| "epoch": 0.12335958005249344, | |
| "grad_norm": 0.011761125177145004, | |
| "learning_rate": 8.524041505748831e-05, | |
| "loss": 0.0112, | |
| "step": 893 | |
| }, | |
| { | |
| "epoch": 0.12349772067965188, | |
| "grad_norm": 0.05986681953072548, | |
| "learning_rate": 8.525445573314562e-05, | |
| "loss": 0.0406, | |
| "step": 894 | |
| }, | |
| { | |
| "epoch": 0.12363586130681033, | |
| "grad_norm": 0.015386831946671009, | |
| "learning_rate": 8.52684807121248e-05, | |
| "loss": 0.0148, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 0.12377400193396879, | |
| "grad_norm": 0.011633604764938354, | |
| "learning_rate": 8.528249002948265e-05, | |
| "loss": 0.0101, | |
| "step": 896 | |
| }, | |
| { | |
| "epoch": 0.12391214256112723, | |
| "grad_norm": 0.008506370708346367, | |
| "learning_rate": 8.529648372015867e-05, | |
| "loss": 0.0066, | |
| "step": 897 | |
| }, | |
| { | |
| "epoch": 0.12405028318828568, | |
| "grad_norm": 0.05086902156472206, | |
| "learning_rate": 8.531046181897555e-05, | |
| "loss": 0.023, | |
| "step": 898 | |
| }, | |
| { | |
| "epoch": 0.12418842381544412, | |
| "grad_norm": 0.001341451657935977, | |
| "learning_rate": 8.532442436063973e-05, | |
| "loss": 0.0021, | |
| "step": 899 | |
| }, | |
| { | |
| "epoch": 0.12432656444260257, | |
| "grad_norm": 0.05210532248020172, | |
| "learning_rate": 8.533837137974189e-05, | |
| "loss": 0.0323, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.12446470506976101, | |
| "grad_norm": 0.007349935360252857, | |
| "learning_rate": 8.535230291075745e-05, | |
| "loss": 0.0087, | |
| "step": 901 | |
| }, | |
| { | |
| "epoch": 0.12460284569691947, | |
| "grad_norm": 0.004401295445859432, | |
| "learning_rate": 8.536621898804715e-05, | |
| "loss": 0.0056, | |
| "step": 902 | |
| }, | |
| { | |
| "epoch": 0.12474098632407792, | |
| "grad_norm": 0.016419509425759315, | |
| "learning_rate": 8.538011964585747e-05, | |
| "loss": 0.0146, | |
| "step": 903 | |
| }, | |
| { | |
| "epoch": 0.12487912695123636, | |
| "grad_norm": 0.012662366963922977, | |
| "learning_rate": 8.539400491832118e-05, | |
| "loss": 0.0102, | |
| "step": 904 | |
| }, | |
| { | |
| "epoch": 0.12501726757839482, | |
| "grad_norm": 0.0041525522246956825, | |
| "learning_rate": 8.540787483945788e-05, | |
| "loss": 0.0056, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 0.12515540820555326, | |
| "grad_norm": 0.001923445495776832, | |
| "learning_rate": 8.542172944317442e-05, | |
| "loss": 0.0033, | |
| "step": 906 | |
| }, | |
| { | |
| "epoch": 0.1252935488327117, | |
| "grad_norm": 0.039593107998371124, | |
| "learning_rate": 8.543556876326543e-05, | |
| "loss": 0.0238, | |
| "step": 907 | |
| }, | |
| { | |
| "epoch": 0.12543168945987015, | |
| "grad_norm": 0.004473451524972916, | |
| "learning_rate": 8.544939283341387e-05, | |
| "loss": 0.0058, | |
| "step": 908 | |
| }, | |
| { | |
| "epoch": 0.1255698300870286, | |
| "grad_norm": 0.004258911591023207, | |
| "learning_rate": 8.546320168719143e-05, | |
| "loss": 0.005, | |
| "step": 909 | |
| }, | |
| { | |
| "epoch": 0.12570797071418705, | |
| "grad_norm": 0.026720302179455757, | |
| "learning_rate": 8.547699535805908e-05, | |
| "loss": 0.022, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.1258461113413455, | |
| "grad_norm": 0.005456454586237669, | |
| "learning_rate": 8.549077387936755e-05, | |
| "loss": 0.0052, | |
| "step": 911 | |
| }, | |
| { | |
| "epoch": 0.12598425196850394, | |
| "grad_norm": 0.01937148906290531, | |
| "learning_rate": 8.550453728435777e-05, | |
| "loss": 0.0169, | |
| "step": 912 | |
| }, | |
| { | |
| "epoch": 0.12612239259566238, | |
| "grad_norm": 0.14971759915351868, | |
| "learning_rate": 8.551828560616142e-05, | |
| "loss": 0.2231, | |
| "step": 913 | |
| }, | |
| { | |
| "epoch": 0.12626053322282083, | |
| "grad_norm": 0.003774692304432392, | |
| "learning_rate": 8.553201887780136e-05, | |
| "loss": 0.0051, | |
| "step": 914 | |
| }, | |
| { | |
| "epoch": 0.12639867384997927, | |
| "grad_norm": 0.012158801779150963, | |
| "learning_rate": 8.554573713219211e-05, | |
| "loss": 0.0086, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 0.12653681447713772, | |
| "grad_norm": 0.0024195557925850153, | |
| "learning_rate": 8.555944040214037e-05, | |
| "loss": 0.004, | |
| "step": 916 | |
| }, | |
| { | |
| "epoch": 0.12667495510429616, | |
| "grad_norm": 0.012309069745242596, | |
| "learning_rate": 8.557312872034538e-05, | |
| "loss": 0.0113, | |
| "step": 917 | |
| }, | |
| { | |
| "epoch": 0.12681309573145463, | |
| "grad_norm": 0.017295846715569496, | |
| "learning_rate": 8.558680211939955e-05, | |
| "loss": 0.0169, | |
| "step": 918 | |
| }, | |
| { | |
| "epoch": 0.12695123635861308, | |
| "grad_norm": 0.013580954633653164, | |
| "learning_rate": 8.560046063178874e-05, | |
| "loss": 0.014, | |
| "step": 919 | |
| }, | |
| { | |
| "epoch": 0.12708937698577152, | |
| "grad_norm": 0.03825406730175018, | |
| "learning_rate": 8.561410428989292e-05, | |
| "loss": 0.0345, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.12722751761292997, | |
| "grad_norm": 0.010689822025597095, | |
| "learning_rate": 8.562773312598646e-05, | |
| "loss": 0.0128, | |
| "step": 921 | |
| }, | |
| { | |
| "epoch": 0.12736565824008841, | |
| "grad_norm": 0.005025971680879593, | |
| "learning_rate": 8.564134717223869e-05, | |
| "loss": 0.0053, | |
| "step": 922 | |
| }, | |
| { | |
| "epoch": 0.12750379886724686, | |
| "grad_norm": 0.005272683221846819, | |
| "learning_rate": 8.56549464607143e-05, | |
| "loss": 0.0048, | |
| "step": 923 | |
| }, | |
| { | |
| "epoch": 0.1276419394944053, | |
| "grad_norm": 0.006757293362170458, | |
| "learning_rate": 8.566853102337387e-05, | |
| "loss": 0.0071, | |
| "step": 924 | |
| }, | |
| { | |
| "epoch": 0.12778008012156375, | |
| "grad_norm": 0.013046073727309704, | |
| "learning_rate": 8.568210089207418e-05, | |
| "loss": 0.0075, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 0.1279182207487222, | |
| "grad_norm": 0.06407853960990906, | |
| "learning_rate": 8.569565609856885e-05, | |
| "loss": 0.0422, | |
| "step": 926 | |
| }, | |
| { | |
| "epoch": 0.12805636137588064, | |
| "grad_norm": 0.02519175596535206, | |
| "learning_rate": 8.570919667450859e-05, | |
| "loss": 0.0159, | |
| "step": 927 | |
| }, | |
| { | |
| "epoch": 0.12819450200303908, | |
| "grad_norm": 0.0054557062685489655, | |
| "learning_rate": 8.57227226514418e-05, | |
| "loss": 0.0065, | |
| "step": 928 | |
| }, | |
| { | |
| "epoch": 0.12833264263019753, | |
| "grad_norm": 0.023621072992682457, | |
| "learning_rate": 8.573623406081491e-05, | |
| "loss": 0.0177, | |
| "step": 929 | |
| }, | |
| { | |
| "epoch": 0.12847078325735597, | |
| "grad_norm": 0.0037180378567427397, | |
| "learning_rate": 8.574973093397288e-05, | |
| "loss": 0.0041, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.12860892388451445, | |
| "grad_norm": 0.020301776006817818, | |
| "learning_rate": 8.576321330215958e-05, | |
| "loss": 0.0146, | |
| "step": 931 | |
| }, | |
| { | |
| "epoch": 0.1287470645116729, | |
| "grad_norm": 0.004708379507064819, | |
| "learning_rate": 8.577668119651829e-05, | |
| "loss": 0.0063, | |
| "step": 932 | |
| }, | |
| { | |
| "epoch": 0.12888520513883134, | |
| "grad_norm": 0.0036832920741289854, | |
| "learning_rate": 8.579013464809206e-05, | |
| "loss": 0.0043, | |
| "step": 933 | |
| }, | |
| { | |
| "epoch": 0.12902334576598978, | |
| "grad_norm": 0.016041239723563194, | |
| "learning_rate": 8.580357368782425e-05, | |
| "loss": 0.0126, | |
| "step": 934 | |
| }, | |
| { | |
| "epoch": 0.12916148639314823, | |
| "grad_norm": 0.07763788849115372, | |
| "learning_rate": 8.581699834655876e-05, | |
| "loss": 0.0569, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 0.12929962702030667, | |
| "grad_norm": 0.04748247563838959, | |
| "learning_rate": 8.58304086550407e-05, | |
| "loss": 0.0277, | |
| "step": 936 | |
| }, | |
| { | |
| "epoch": 0.12943776764746512, | |
| "grad_norm": 0.006311520934104919, | |
| "learning_rate": 8.584380464391663e-05, | |
| "loss": 0.0056, | |
| "step": 937 | |
| }, | |
| { | |
| "epoch": 0.12957590827462356, | |
| "grad_norm": 0.004489540588110685, | |
| "learning_rate": 8.585718634373504e-05, | |
| "loss": 0.0057, | |
| "step": 938 | |
| }, | |
| { | |
| "epoch": 0.129714048901782, | |
| "grad_norm": 0.01480119675397873, | |
| "learning_rate": 8.587055378494679e-05, | |
| "loss": 0.0148, | |
| "step": 939 | |
| }, | |
| { | |
| "epoch": 0.12985218952894045, | |
| "grad_norm": 0.025000635534524918, | |
| "learning_rate": 8.588390699790551e-05, | |
| "loss": 0.0173, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.1299903301560989, | |
| "grad_norm": 0.0023975621443241835, | |
| "learning_rate": 8.589724601286795e-05, | |
| "loss": 0.0031, | |
| "step": 941 | |
| }, | |
| { | |
| "epoch": 0.13012847078325734, | |
| "grad_norm": 0.036758214235305786, | |
| "learning_rate": 8.591057085999454e-05, | |
| "loss": 0.0375, | |
| "step": 942 | |
| }, | |
| { | |
| "epoch": 0.13026661141041582, | |
| "grad_norm": 0.02752687968313694, | |
| "learning_rate": 8.592388156934964e-05, | |
| "loss": 0.0153, | |
| "step": 943 | |
| }, | |
| { | |
| "epoch": 0.13040475203757426, | |
| "grad_norm": 0.0026282810140401125, | |
| "learning_rate": 8.593717817090202e-05, | |
| "loss": 0.004, | |
| "step": 944 | |
| }, | |
| { | |
| "epoch": 0.1305428926647327, | |
| "grad_norm": 0.00695823086425662, | |
| "learning_rate": 8.595046069452531e-05, | |
| "loss": 0.0067, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 0.13068103329189115, | |
| "grad_norm": 0.005962789058685303, | |
| "learning_rate": 8.596372916999831e-05, | |
| "loss": 0.007, | |
| "step": 946 | |
| }, | |
| { | |
| "epoch": 0.1308191739190496, | |
| "grad_norm": 0.19016483426094055, | |
| "learning_rate": 8.597698362700545e-05, | |
| "loss": 0.1038, | |
| "step": 947 | |
| }, | |
| { | |
| "epoch": 0.13095731454620804, | |
| "grad_norm": 0.009827204048633575, | |
| "learning_rate": 8.599022409513714e-05, | |
| "loss": 0.0071, | |
| "step": 948 | |
| }, | |
| { | |
| "epoch": 0.1310954551733665, | |
| "grad_norm": 0.0029148852918297052, | |
| "learning_rate": 8.600345060389027e-05, | |
| "loss": 0.0041, | |
| "step": 949 | |
| }, | |
| { | |
| "epoch": 0.13123359580052493, | |
| "grad_norm": 0.04867846518754959, | |
| "learning_rate": 8.601666318266846e-05, | |
| "loss": 0.0244, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.13137173642768338, | |
| "grad_norm": 0.005912700667977333, | |
| "learning_rate": 8.602986186078254e-05, | |
| "loss": 0.0086, | |
| "step": 951 | |
| }, | |
| { | |
| "epoch": 0.13150987705484182, | |
| "grad_norm": 0.006628013215959072, | |
| "learning_rate": 8.604304666745098e-05, | |
| "loss": 0.0081, | |
| "step": 952 | |
| }, | |
| { | |
| "epoch": 0.13164801768200027, | |
| "grad_norm": 0.008250492624938488, | |
| "learning_rate": 8.605621763180016e-05, | |
| "loss": 0.0074, | |
| "step": 953 | |
| }, | |
| { | |
| "epoch": 0.1317861583091587, | |
| "grad_norm": 0.013309494592249393, | |
| "learning_rate": 8.606937478286485e-05, | |
| "loss": 0.0143, | |
| "step": 954 | |
| }, | |
| { | |
| "epoch": 0.13192429893631716, | |
| "grad_norm": 0.0022110934369266033, | |
| "learning_rate": 8.608251814958855e-05, | |
| "loss": 0.0041, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 0.13206243956347563, | |
| "grad_norm": 0.009347507730126381, | |
| "learning_rate": 8.609564776082388e-05, | |
| "loss": 0.0117, | |
| "step": 956 | |
| }, | |
| { | |
| "epoch": 0.13220058019063408, | |
| "grad_norm": 0.007068824488669634, | |
| "learning_rate": 8.610876364533301e-05, | |
| "loss": 0.0107, | |
| "step": 957 | |
| }, | |
| { | |
| "epoch": 0.13233872081779252, | |
| "grad_norm": 0.02662699855864048, | |
| "learning_rate": 8.612186583178794e-05, | |
| "loss": 0.0315, | |
| "step": 958 | |
| }, | |
| { | |
| "epoch": 0.13247686144495097, | |
| "grad_norm": 0.015548110008239746, | |
| "learning_rate": 8.613495434877095e-05, | |
| "loss": 0.016, | |
| "step": 959 | |
| }, | |
| { | |
| "epoch": 0.1326150020721094, | |
| "grad_norm": 0.0028302946593612432, | |
| "learning_rate": 8.614802922477495e-05, | |
| "loss": 0.0037, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.13275314269926786, | |
| "grad_norm": 0.0047238729894161224, | |
| "learning_rate": 8.616109048820388e-05, | |
| "loss": 0.006, | |
| "step": 961 | |
| }, | |
| { | |
| "epoch": 0.1328912833264263, | |
| "grad_norm": 0.016504082828760147, | |
| "learning_rate": 8.617413816737302e-05, | |
| "loss": 0.014, | |
| "step": 962 | |
| }, | |
| { | |
| "epoch": 0.13302942395358475, | |
| "grad_norm": 0.03382395952939987, | |
| "learning_rate": 8.61871722905094e-05, | |
| "loss": 0.0266, | |
| "step": 963 | |
| }, | |
| { | |
| "epoch": 0.1331675645807432, | |
| "grad_norm": 0.008775300346314907, | |
| "learning_rate": 8.620019288575218e-05, | |
| "loss": 0.01, | |
| "step": 964 | |
| }, | |
| { | |
| "epoch": 0.13330570520790164, | |
| "grad_norm": 0.0034813378006219864, | |
| "learning_rate": 8.621319998115303e-05, | |
| "loss": 0.0044, | |
| "step": 965 | |
| }, | |
| { | |
| "epoch": 0.13344384583506008, | |
| "grad_norm": 0.007844419218599796, | |
| "learning_rate": 8.622619360467634e-05, | |
| "loss": 0.0125, | |
| "step": 966 | |
| }, | |
| { | |
| "epoch": 0.13358198646221853, | |
| "grad_norm": 0.0025214802008122206, | |
| "learning_rate": 8.623917378419984e-05, | |
| "loss": 0.0046, | |
| "step": 967 | |
| }, | |
| { | |
| "epoch": 0.13372012708937697, | |
| "grad_norm": 0.02048412896692753, | |
| "learning_rate": 8.625214054751471e-05, | |
| "loss": 0.0141, | |
| "step": 968 | |
| }, | |
| { | |
| "epoch": 0.13385826771653545, | |
| "grad_norm": 0.008773239329457283, | |
| "learning_rate": 8.62650939223261e-05, | |
| "loss": 0.0126, | |
| "step": 969 | |
| }, | |
| { | |
| "epoch": 0.1339964083436939, | |
| "grad_norm": 0.07702983915805817, | |
| "learning_rate": 8.627803393625343e-05, | |
| "loss": 0.0595, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.13413454897085234, | |
| "grad_norm": 0.012115233577787876, | |
| "learning_rate": 8.629096061683071e-05, | |
| "loss": 0.0109, | |
| "step": 971 | |
| }, | |
| { | |
| "epoch": 0.13427268959801078, | |
| "grad_norm": 0.013093401677906513, | |
| "learning_rate": 8.630387399150694e-05, | |
| "loss": 0.0094, | |
| "step": 972 | |
| }, | |
| { | |
| "epoch": 0.13441083022516923, | |
| "grad_norm": 0.006639427505433559, | |
| "learning_rate": 8.631677408764646e-05, | |
| "loss": 0.0076, | |
| "step": 973 | |
| }, | |
| { | |
| "epoch": 0.13454897085232767, | |
| "grad_norm": 0.04810924455523491, | |
| "learning_rate": 8.632966093252927e-05, | |
| "loss": 0.0368, | |
| "step": 974 | |
| }, | |
| { | |
| "epoch": 0.13468711147948612, | |
| "grad_norm": 0.007319958880543709, | |
| "learning_rate": 8.634253455335138e-05, | |
| "loss": 0.0072, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 0.13482525210664456, | |
| "grad_norm": 0.05134734511375427, | |
| "learning_rate": 8.635539497722518e-05, | |
| "loss": 0.0367, | |
| "step": 976 | |
| }, | |
| { | |
| "epoch": 0.134963392733803, | |
| "grad_norm": 0.010004381649196148, | |
| "learning_rate": 8.636824223117976e-05, | |
| "loss": 0.0104, | |
| "step": 977 | |
| }, | |
| { | |
| "epoch": 0.13510153336096145, | |
| "grad_norm": 0.00346842291764915, | |
| "learning_rate": 8.638107634216126e-05, | |
| "loss": 0.0052, | |
| "step": 978 | |
| }, | |
| { | |
| "epoch": 0.1352396739881199, | |
| "grad_norm": 0.01166492234915495, | |
| "learning_rate": 8.639389733703317e-05, | |
| "loss": 0.0085, | |
| "step": 979 | |
| }, | |
| { | |
| "epoch": 0.13537781461527834, | |
| "grad_norm": 0.02950691431760788, | |
| "learning_rate": 8.640670524257675e-05, | |
| "loss": 0.0222, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.13551595524243681, | |
| "grad_norm": 0.010350341908633709, | |
| "learning_rate": 8.641950008549133e-05, | |
| "loss": 0.0077, | |
| "step": 981 | |
| }, | |
| { | |
| "epoch": 0.13565409586959526, | |
| "grad_norm": 0.0018771348986774683, | |
| "learning_rate": 8.643228189239453e-05, | |
| "loss": 0.003, | |
| "step": 982 | |
| }, | |
| { | |
| "epoch": 0.1357922364967537, | |
| "grad_norm": 0.012851797975599766, | |
| "learning_rate": 8.644505068982283e-05, | |
| "loss": 0.011, | |
| "step": 983 | |
| }, | |
| { | |
| "epoch": 0.13593037712391215, | |
| "grad_norm": 0.004215499386191368, | |
| "learning_rate": 8.645780650423166e-05, | |
| "loss": 0.0054, | |
| "step": 984 | |
| }, | |
| { | |
| "epoch": 0.1360685177510706, | |
| "grad_norm": 0.13906902074813843, | |
| "learning_rate": 8.647054936199591e-05, | |
| "loss": 0.0696, | |
| "step": 985 | |
| }, | |
| { | |
| "epoch": 0.13620665837822904, | |
| "grad_norm": 0.007836158387362957, | |
| "learning_rate": 8.648327928941014e-05, | |
| "loss": 0.0095, | |
| "step": 986 | |
| }, | |
| { | |
| "epoch": 0.13634479900538748, | |
| "grad_norm": 0.0017098481766879559, | |
| "learning_rate": 8.649599631268893e-05, | |
| "loss": 0.0028, | |
| "step": 987 | |
| }, | |
| { | |
| "epoch": 0.13648293963254593, | |
| "grad_norm": 0.004099471494555473, | |
| "learning_rate": 8.650870045796727e-05, | |
| "loss": 0.0046, | |
| "step": 988 | |
| }, | |
| { | |
| "epoch": 0.13662108025970437, | |
| "grad_norm": 0.0024917221162468195, | |
| "learning_rate": 8.65213917513008e-05, | |
| "loss": 0.0032, | |
| "step": 989 | |
| }, | |
| { | |
| "epoch": 0.13675922088686282, | |
| "grad_norm": 0.006787384394556284, | |
| "learning_rate": 8.653407021866616e-05, | |
| "loss": 0.0052, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.13689736151402127, | |
| "grad_norm": 0.07070998102426529, | |
| "learning_rate": 8.654673588596133e-05, | |
| "loss": 0.0427, | |
| "step": 991 | |
| }, | |
| { | |
| "epoch": 0.1370355021411797, | |
| "grad_norm": 0.02854343131184578, | |
| "learning_rate": 8.655938877900593e-05, | |
| "loss": 0.0225, | |
| "step": 992 | |
| }, | |
| { | |
| "epoch": 0.13717364276833816, | |
| "grad_norm": 0.014134407043457031, | |
| "learning_rate": 8.657202892354155e-05, | |
| "loss": 0.0097, | |
| "step": 993 | |
| }, | |
| { | |
| "epoch": 0.13731178339549663, | |
| "grad_norm": 0.006957308389246464, | |
| "learning_rate": 8.658465634523199e-05, | |
| "loss": 0.0075, | |
| "step": 994 | |
| }, | |
| { | |
| "epoch": 0.13744992402265507, | |
| "grad_norm": 0.008567973971366882, | |
| "learning_rate": 8.659727106966367e-05, | |
| "loss": 0.0095, | |
| "step": 995 | |
| }, | |
| { | |
| "epoch": 0.13758806464981352, | |
| "grad_norm": 0.010392943397164345, | |
| "learning_rate": 8.660987312234595e-05, | |
| "loss": 0.0079, | |
| "step": 996 | |
| }, | |
| { | |
| "epoch": 0.13772620527697196, | |
| "grad_norm": 0.09674349427223206, | |
| "learning_rate": 8.662246252871131e-05, | |
| "loss": 0.0385, | |
| "step": 997 | |
| }, | |
| { | |
| "epoch": 0.1378643459041304, | |
| "grad_norm": 0.002456868067383766, | |
| "learning_rate": 8.663503931411578e-05, | |
| "loss": 0.0032, | |
| "step": 998 | |
| }, | |
| { | |
| "epoch": 0.13800248653128885, | |
| "grad_norm": 0.07229039818048477, | |
| "learning_rate": 8.664760350383924e-05, | |
| "loss": 0.0491, | |
| "step": 999 | |
| }, | |
| { | |
| "epoch": 0.1381406271584473, | |
| "grad_norm": 0.00896090641617775, | |
| "learning_rate": 8.666015512308563e-05, | |
| "loss": 0.0083, | |
| "step": 1000 | |
| } | |
| ], | |
| "logging_steps": 1.0, | |
| "max_steps": 28956, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 4, | |
| "save_steps": 1000, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 0.0, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |