{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 4.972704714640199, "eval_steps": 500, "global_step": 755, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.006617038875103391, "grad_norm": 34.37208740591342, "learning_rate": 4.999978357111178e-05, "loss": 2.4934, "num_input_tokens_seen": 262144, "step": 1 }, { "epoch": 0.013234077750206782, "grad_norm": 45.9821682566152, "learning_rate": 4.9999134288194436e-05, "loss": 3.8516, "num_input_tokens_seen": 524288, "step": 2 }, { "epoch": 0.019851116625310174, "grad_norm": 39.177129424978055, "learning_rate": 4.9998052162489854e-05, "loss": 3.5826, "num_input_tokens_seen": 786432, "step": 3 }, { "epoch": 0.026468155500413565, "grad_norm": 240.13346858978267, "learning_rate": 4.999653721273429e-05, "loss": 9.4935, "num_input_tokens_seen": 1048576, "step": 4 }, { "epoch": 0.033085194375516956, "grad_norm": 32.940086480387635, "learning_rate": 4.999458946515808e-05, "loss": 3.9766, "num_input_tokens_seen": 1310720, "step": 5 }, { "epoch": 0.03970223325062035, "grad_norm": 133.3545454549638, "learning_rate": 4.99922089534851e-05, "loss": 4.7751, "num_input_tokens_seen": 1572864, "step": 6 }, { "epoch": 0.04631927212572374, "grad_norm": 27.309262699099058, "learning_rate": 4.998939571893228e-05, "loss": 3.1784, "num_input_tokens_seen": 1835008, "step": 7 }, { "epoch": 0.05293631100082713, "grad_norm": 7.830626860343412, "learning_rate": 4.998614981020884e-05, "loss": 2.5119, "num_input_tokens_seen": 2097152, "step": 8 }, { "epoch": 0.05955334987593052, "grad_norm": 8.048304569875182, "learning_rate": 4.998247128351545e-05, "loss": 2.2489, "num_input_tokens_seen": 2359296, "step": 9 }, { "epoch": 0.06617038875103391, "grad_norm": 8.196172984809154, "learning_rate": 4.997836020254328e-05, "loss": 2.226, "num_input_tokens_seen": 2621440, "step": 10 }, { "epoch": 0.07278742762613731, "grad_norm": 5.02594769824259, "learning_rate": 4.9973816638472846e-05, "loss": 2.0551, "num_input_tokens_seen": 2883584, "step": 11 }, { "epoch": 0.0794044665012407, "grad_norm": 5.423582402121407, "learning_rate": 4.996884066997284e-05, "loss": 1.9894, "num_input_tokens_seen": 3145728, "step": 12 }, { "epoch": 0.08602150537634409, "grad_norm": 3.9697486334940897, "learning_rate": 4.9963432383198726e-05, "loss": 1.9212, "num_input_tokens_seen": 3407872, "step": 13 }, { "epoch": 0.09263854425144748, "grad_norm": 3.184049663288423, "learning_rate": 4.995759187179126e-05, "loss": 1.8776, "num_input_tokens_seen": 3670016, "step": 14 }, { "epoch": 0.09925558312655088, "grad_norm": 2.1877083301154907, "learning_rate": 4.995131923687488e-05, "loss": 1.8142, "num_input_tokens_seen": 3932160, "step": 15 }, { "epoch": 0.10587262200165426, "grad_norm": 3.193160483646996, "learning_rate": 4.9944614587055925e-05, "loss": 1.7548, "num_input_tokens_seen": 4194304, "step": 16 }, { "epoch": 0.11248966087675766, "grad_norm": 4.417433823363654, "learning_rate": 4.993747803842081e-05, "loss": 1.8321, "num_input_tokens_seen": 4456448, "step": 17 }, { "epoch": 0.11910669975186104, "grad_norm": 2.6131027548528114, "learning_rate": 4.992990971453394e-05, "loss": 1.7912, "num_input_tokens_seen": 4718592, "step": 18 }, { "epoch": 0.12572373862696443, "grad_norm": 2.5605507836540236, "learning_rate": 4.9921909746435633e-05, "loss": 1.7313, "num_input_tokens_seen": 4980736, "step": 19 }, { "epoch": 0.13234077750206782, "grad_norm": 1.8403308592664107, "learning_rate": 4.991347827263982e-05, "loss": 1.7471, "num_input_tokens_seen": 5242880, "step": 20 }, { "epoch": 0.13895781637717122, "grad_norm": 1.9498430221428151, "learning_rate": 4.990461543913168e-05, "loss": 1.6706, "num_input_tokens_seen": 5505024, "step": 21 }, { "epoch": 0.14557485525227462, "grad_norm": 1.8088575603375883, "learning_rate": 4.9895321399365044e-05, "loss": 1.6593, "num_input_tokens_seen": 5767168, "step": 22 }, { "epoch": 0.152191894127378, "grad_norm": 1.7965239443436583, "learning_rate": 4.988559631425983e-05, "loss": 1.6705, "num_input_tokens_seen": 6029312, "step": 23 }, { "epoch": 0.1588089330024814, "grad_norm": 3.13574184340746, "learning_rate": 4.987544035219917e-05, "loss": 1.6509, "num_input_tokens_seen": 6291456, "step": 24 }, { "epoch": 0.1654259718775848, "grad_norm": 2.048821572892012, "learning_rate": 4.9864853689026556e-05, "loss": 1.6473, "num_input_tokens_seen": 6553600, "step": 25 }, { "epoch": 0.17204301075268819, "grad_norm": 1.6289600767650514, "learning_rate": 4.985383650804277e-05, "loss": 1.6402, "num_input_tokens_seen": 6815744, "step": 26 }, { "epoch": 0.17866004962779156, "grad_norm": 1.846599816589036, "learning_rate": 4.984238900000271e-05, "loss": 1.5957, "num_input_tokens_seen": 7077888, "step": 27 }, { "epoch": 0.18527708850289495, "grad_norm": 2.49311076745654, "learning_rate": 4.983051136311209e-05, "loss": 1.6361, "num_input_tokens_seen": 7340032, "step": 28 }, { "epoch": 0.19189412737799835, "grad_norm": 1.7775011755473744, "learning_rate": 4.9818203803024e-05, "loss": 1.6234, "num_input_tokens_seen": 7602176, "step": 29 }, { "epoch": 0.19851116625310175, "grad_norm": 2.3264416263522576, "learning_rate": 4.9805466532835376e-05, "loss": 1.6288, "num_input_tokens_seen": 7864320, "step": 30 }, { "epoch": 0.20512820512820512, "grad_norm": 1.8036645148064498, "learning_rate": 4.9792299773083276e-05, "loss": 1.6263, "num_input_tokens_seen": 8126464, "step": 31 }, { "epoch": 0.21174524400330852, "grad_norm": 1.951872109802615, "learning_rate": 4.9778703751741076e-05, "loss": 1.5944, "num_input_tokens_seen": 8388608, "step": 32 }, { "epoch": 0.21836228287841192, "grad_norm": 1.9171459318694717, "learning_rate": 4.9764678704214506e-05, "loss": 1.5671, "num_input_tokens_seen": 8650752, "step": 33 }, { "epoch": 0.22497932175351532, "grad_norm": 1.8596046998788465, "learning_rate": 4.9750224873337605e-05, "loss": 1.5907, "num_input_tokens_seen": 8912896, "step": 34 }, { "epoch": 0.23159636062861869, "grad_norm": 1.9357133650068838, "learning_rate": 4.973534250936851e-05, "loss": 1.5631, "num_input_tokens_seen": 9175040, "step": 35 }, { "epoch": 0.23821339950372208, "grad_norm": 1.9002383695379652, "learning_rate": 4.9720031869985084e-05, "loss": 1.5915, "num_input_tokens_seen": 9437184, "step": 36 }, { "epoch": 0.24483043837882548, "grad_norm": 1.6528688694936653, "learning_rate": 4.970429322028051e-05, "loss": 1.544, "num_input_tokens_seen": 9699328, "step": 37 }, { "epoch": 0.25144747725392885, "grad_norm": 1.7280201764860956, "learning_rate": 4.968812683275866e-05, "loss": 1.5263, "num_input_tokens_seen": 9961472, "step": 38 }, { "epoch": 0.25806451612903225, "grad_norm": 1.3933242558940806, "learning_rate": 4.96715329873294e-05, "loss": 1.5549, "num_input_tokens_seen": 10223616, "step": 39 }, { "epoch": 0.26468155500413565, "grad_norm": 1.8069938406301924, "learning_rate": 4.965451197130373e-05, "loss": 1.5227, "num_input_tokens_seen": 10485760, "step": 40 }, { "epoch": 0.27129859387923905, "grad_norm": 1.520117712776801, "learning_rate": 4.963706407938881e-05, "loss": 1.5338, "num_input_tokens_seen": 10747904, "step": 41 }, { "epoch": 0.27791563275434245, "grad_norm": 1.9100553957972657, "learning_rate": 4.961918961368287e-05, "loss": 1.5387, "num_input_tokens_seen": 11010048, "step": 42 }, { "epoch": 0.28453267162944584, "grad_norm": 1.7226630358900272, "learning_rate": 4.960088888366998e-05, "loss": 1.5439, "num_input_tokens_seen": 11272192, "step": 43 }, { "epoch": 0.29114971050454924, "grad_norm": 1.3593181279067512, "learning_rate": 4.9582162206214654e-05, "loss": 1.5352, "num_input_tokens_seen": 11534336, "step": 44 }, { "epoch": 0.2977667493796526, "grad_norm": 1.3032108941431264, "learning_rate": 4.956300990555643e-05, "loss": 1.5087, "num_input_tokens_seen": 11796480, "step": 45 }, { "epoch": 0.304383788254756, "grad_norm": 1.6080688788534845, "learning_rate": 4.9543432313304184e-05, "loss": 1.4949, "num_input_tokens_seen": 12058624, "step": 46 }, { "epoch": 0.3110008271298594, "grad_norm": 1.8537446327987028, "learning_rate": 4.9523429768430445e-05, "loss": 1.5097, "num_input_tokens_seen": 12320768, "step": 47 }, { "epoch": 0.3176178660049628, "grad_norm": 1.5807168324755116, "learning_rate": 4.950300261726549e-05, "loss": 1.5391, "num_input_tokens_seen": 12582912, "step": 48 }, { "epoch": 0.3242349048800662, "grad_norm": 1.4720318553915235, "learning_rate": 4.9482151213491386e-05, "loss": 1.5192, "num_input_tokens_seen": 12845056, "step": 49 }, { "epoch": 0.3308519437551696, "grad_norm": 1.4230626865941303, "learning_rate": 4.9460875918135804e-05, "loss": 1.5129, "num_input_tokens_seen": 13107200, "step": 50 }, { "epoch": 0.337468982630273, "grad_norm": 1.995697274136008, "learning_rate": 4.943917709956584e-05, "loss": 1.5243, "num_input_tokens_seen": 13369344, "step": 51 }, { "epoch": 0.34408602150537637, "grad_norm": 1.2626400291619349, "learning_rate": 4.941705513348157e-05, "loss": 1.5031, "num_input_tokens_seen": 13631488, "step": 52 }, { "epoch": 0.3507030603804797, "grad_norm": 1.3497492422606214, "learning_rate": 4.939451040290961e-05, "loss": 1.4925, "num_input_tokens_seen": 13893632, "step": 53 }, { "epoch": 0.3573200992555831, "grad_norm": 1.593030123983225, "learning_rate": 4.937154329819644e-05, "loss": 1.474, "num_input_tokens_seen": 14155776, "step": 54 }, { "epoch": 0.3639371381306865, "grad_norm": 1.45845743040763, "learning_rate": 4.934815421700165e-05, "loss": 1.4776, "num_input_tokens_seen": 14417920, "step": 55 }, { "epoch": 0.3705541770057899, "grad_norm": 1.3404864487553547, "learning_rate": 4.932434356429106e-05, "loss": 1.4824, "num_input_tokens_seen": 14680064, "step": 56 }, { "epoch": 0.3771712158808933, "grad_norm": 1.2551629875891903, "learning_rate": 4.930011175232973e-05, "loss": 1.5041, "num_input_tokens_seen": 14942208, "step": 57 }, { "epoch": 0.3837882547559967, "grad_norm": 1.609822533873946, "learning_rate": 4.927545920067479e-05, "loss": 1.4685, "num_input_tokens_seen": 15204352, "step": 58 }, { "epoch": 0.3904052936311001, "grad_norm": 1.4469457354277884, "learning_rate": 4.925038633616818e-05, "loss": 1.4517, "num_input_tokens_seen": 15466496, "step": 59 }, { "epoch": 0.3970223325062035, "grad_norm": 1.6525439629778043, "learning_rate": 4.9224893592929275e-05, "loss": 1.4878, "num_input_tokens_seen": 15728640, "step": 60 }, { "epoch": 0.40363937138130684, "grad_norm": 1.254692246634876, "learning_rate": 4.9198981412347364e-05, "loss": 1.4882, "num_input_tokens_seen": 15990784, "step": 61 }, { "epoch": 0.41025641025641024, "grad_norm": 1.4543664783049992, "learning_rate": 4.917265024307401e-05, "loss": 1.452, "num_input_tokens_seen": 16252928, "step": 62 }, { "epoch": 0.41687344913151364, "grad_norm": 1.660807543545703, "learning_rate": 4.914590054101526e-05, "loss": 1.4566, "num_input_tokens_seen": 16515072, "step": 63 }, { "epoch": 0.42349048800661704, "grad_norm": 1.417113851349832, "learning_rate": 4.911873276932376e-05, "loss": 1.4624, "num_input_tokens_seen": 16777216, "step": 64 }, { "epoch": 0.43010752688172044, "grad_norm": 1.2145243782914297, "learning_rate": 4.909114739839079e-05, "loss": 1.4775, "num_input_tokens_seen": 17039360, "step": 65 }, { "epoch": 0.43672456575682383, "grad_norm": 1.3977545670249865, "learning_rate": 4.906314490583802e-05, "loss": 1.437, "num_input_tokens_seen": 17301504, "step": 66 }, { "epoch": 0.44334160463192723, "grad_norm": 1.2151379240046773, "learning_rate": 4.903472577650934e-05, "loss": 1.4492, "num_input_tokens_seen": 17563648, "step": 67 }, { "epoch": 0.44995864350703063, "grad_norm": 1.4902123461239007, "learning_rate": 4.900589050246237e-05, "loss": 1.4503, "num_input_tokens_seen": 17825792, "step": 68 }, { "epoch": 0.456575682382134, "grad_norm": 1.2602633654476592, "learning_rate": 4.897663958296002e-05, "loss": 1.4489, "num_input_tokens_seen": 18087936, "step": 69 }, { "epoch": 0.46319272125723737, "grad_norm": 1.3183809307307202, "learning_rate": 4.894697352446182e-05, "loss": 1.4577, "num_input_tokens_seen": 18350080, "step": 70 }, { "epoch": 0.46980976013234077, "grad_norm": 1.527801009411038, "learning_rate": 4.891689284061513e-05, "loss": 1.4678, "num_input_tokens_seen": 18612224, "step": 71 }, { "epoch": 0.47642679900744417, "grad_norm": 1.6249022564808302, "learning_rate": 4.888639805224626e-05, "loss": 1.4618, "num_input_tokens_seen": 18874368, "step": 72 }, { "epoch": 0.48304383788254757, "grad_norm": 1.267461796577937, "learning_rate": 4.885548968735147e-05, "loss": 1.4576, "num_input_tokens_seen": 19136512, "step": 73 }, { "epoch": 0.48966087675765096, "grad_norm": 1.7705838242670227, "learning_rate": 4.882416828108781e-05, "loss": 1.4191, "num_input_tokens_seen": 19398656, "step": 74 }, { "epoch": 0.49627791563275436, "grad_norm": 1.5430488224612218, "learning_rate": 4.879243437576383e-05, "loss": 1.4178, "num_input_tokens_seen": 19660800, "step": 75 }, { "epoch": 0.5028949545078577, "grad_norm": 1.1859905942576958, "learning_rate": 4.8760288520830254e-05, "loss": 1.4184, "num_input_tokens_seen": 19922944, "step": 76 }, { "epoch": 0.5095119933829612, "grad_norm": 1.6712647868129495, "learning_rate": 4.8727731272870406e-05, "loss": 1.4604, "num_input_tokens_seen": 20185088, "step": 77 }, { "epoch": 0.5161290322580645, "grad_norm": 1.2289801478873525, "learning_rate": 4.8694763195590606e-05, "loss": 1.4326, "num_input_tokens_seen": 20447232, "step": 78 }, { "epoch": 0.522746071133168, "grad_norm": 1.4590483227044124, "learning_rate": 4.866138485981041e-05, "loss": 1.4106, "num_input_tokens_seen": 20709376, "step": 79 }, { "epoch": 0.5293631100082713, "grad_norm": 1.2028857677920306, "learning_rate": 4.862759684345269e-05, "loss": 1.4125, "num_input_tokens_seen": 20971520, "step": 80 }, { "epoch": 0.5359801488833746, "grad_norm": 1.0904585549604011, "learning_rate": 4.859339973153368e-05, "loss": 1.4316, "num_input_tokens_seen": 21233664, "step": 81 }, { "epoch": 0.5425971877584781, "grad_norm": 1.3835096213926883, "learning_rate": 4.855879411615282e-05, "loss": 1.3964, "num_input_tokens_seen": 21495808, "step": 82 }, { "epoch": 0.5492142266335814, "grad_norm": 1.3130396739637342, "learning_rate": 4.8523780596482475e-05, "loss": 1.3856, "num_input_tokens_seen": 21757952, "step": 83 }, { "epoch": 0.5558312655086849, "grad_norm": 1.6210339400595701, "learning_rate": 4.848835977875764e-05, "loss": 1.443, "num_input_tokens_seen": 22020096, "step": 84 }, { "epoch": 0.5624483043837882, "grad_norm": 1.3048900511501402, "learning_rate": 4.8452532276265364e-05, "loss": 1.3976, "num_input_tokens_seen": 22282240, "step": 85 }, { "epoch": 0.5690653432588917, "grad_norm": 1.5253749617527992, "learning_rate": 4.8416298709334156e-05, "loss": 1.3844, "num_input_tokens_seen": 22544384, "step": 86 }, { "epoch": 0.575682382133995, "grad_norm": 1.3133909248429987, "learning_rate": 4.837965970532328e-05, "loss": 1.4196, "num_input_tokens_seen": 22806528, "step": 87 }, { "epoch": 0.5822994210090985, "grad_norm": 1.5372189546773116, "learning_rate": 4.8342615898611854e-05, "loss": 1.3837, "num_input_tokens_seen": 23068672, "step": 88 }, { "epoch": 0.5889164598842018, "grad_norm": 1.6875441852021882, "learning_rate": 4.8305167930587844e-05, "loss": 1.3941, "num_input_tokens_seen": 23330816, "step": 89 }, { "epoch": 0.5955334987593052, "grad_norm": 1.49140049772374, "learning_rate": 4.8267316449637054e-05, "loss": 1.4027, "num_input_tokens_seen": 23592960, "step": 90 }, { "epoch": 0.6021505376344086, "grad_norm": 1.437082448499028, "learning_rate": 4.8229062111131764e-05, "loss": 1.4069, "num_input_tokens_seen": 23855104, "step": 91 }, { "epoch": 0.608767576509512, "grad_norm": 1.4617250864457985, "learning_rate": 4.8190405577419506e-05, "loss": 1.3939, "num_input_tokens_seen": 24117248, "step": 92 }, { "epoch": 0.6153846153846154, "grad_norm": 1.6286516280162249, "learning_rate": 4.815134751781153e-05, "loss": 1.3755, "num_input_tokens_seen": 24379392, "step": 93 }, { "epoch": 0.6220016542597188, "grad_norm": 1.4334609780513905, "learning_rate": 4.8111888608571234e-05, "loss": 1.4233, "num_input_tokens_seen": 24641536, "step": 94 }, { "epoch": 0.6286186931348222, "grad_norm": 1.2470313746006922, "learning_rate": 4.8072029532902426e-05, "loss": 1.3776, "num_input_tokens_seen": 24903680, "step": 95 }, { "epoch": 0.6352357320099256, "grad_norm": 1.6301987833809242, "learning_rate": 4.803177098093757e-05, "loss": 1.3677, "num_input_tokens_seen": 25165824, "step": 96 }, { "epoch": 0.6418527708850289, "grad_norm": 1.283546564124238, "learning_rate": 4.7991113649725734e-05, "loss": 1.3608, "num_input_tokens_seen": 25427968, "step": 97 }, { "epoch": 0.6484698097601324, "grad_norm": 1.7514307612371371, "learning_rate": 4.795005824322061e-05, "loss": 1.3782, "num_input_tokens_seen": 25690112, "step": 98 }, { "epoch": 0.6550868486352357, "grad_norm": 1.3349872986236062, "learning_rate": 4.7908605472268266e-05, "loss": 1.3891, "num_input_tokens_seen": 25952256, "step": 99 }, { "epoch": 0.6617038875103392, "grad_norm": 1.6887963390015819, "learning_rate": 4.786675605459487e-05, "loss": 1.3962, "num_input_tokens_seen": 26214400, "step": 100 }, { "epoch": 0.6683209263854425, "grad_norm": 1.3690085291694667, "learning_rate": 4.782451071479428e-05, "loss": 1.3793, "num_input_tokens_seen": 26476544, "step": 101 }, { "epoch": 0.674937965260546, "grad_norm": 1.3438890649631892, "learning_rate": 4.7781870184315435e-05, "loss": 1.3795, "num_input_tokens_seen": 26738688, "step": 102 }, { "epoch": 0.6815550041356493, "grad_norm": 1.4251453112744144, "learning_rate": 4.773883520144974e-05, "loss": 1.3813, "num_input_tokens_seen": 27000832, "step": 103 }, { "epoch": 0.6881720430107527, "grad_norm": 1.3732536405102582, "learning_rate": 4.769540651131828e-05, "loss": 1.3582, "num_input_tokens_seen": 27262976, "step": 104 }, { "epoch": 0.6947890818858561, "grad_norm": 1.1035631693243748, "learning_rate": 4.76515848658589e-05, "loss": 1.3708, "num_input_tokens_seen": 27525120, "step": 105 }, { "epoch": 0.7014061207609594, "grad_norm": 1.2754803303745712, "learning_rate": 4.760737102381321e-05, "loss": 1.3467, "num_input_tokens_seen": 27787264, "step": 106 }, { "epoch": 0.7080231596360629, "grad_norm": 1.1917616813939031, "learning_rate": 4.756276575071342e-05, "loss": 1.3691, "num_input_tokens_seen": 28049408, "step": 107 }, { "epoch": 0.7146401985111662, "grad_norm": 1.2971389049692688, "learning_rate": 4.75177698188691e-05, "loss": 1.3573, "num_input_tokens_seen": 28311552, "step": 108 }, { "epoch": 0.7212572373862697, "grad_norm": 1.1796083016863366, "learning_rate": 4.7472384007353804e-05, "loss": 1.3456, "num_input_tokens_seen": 28573696, "step": 109 }, { "epoch": 0.727874276261373, "grad_norm": 1.393759938768685, "learning_rate": 4.7426609101991605e-05, "loss": 1.3741, "num_input_tokens_seen": 28835840, "step": 110 }, { "epoch": 0.7344913151364765, "grad_norm": 1.239317627457199, "learning_rate": 4.7380445895343445e-05, "loss": 1.3871, "num_input_tokens_seen": 29097984, "step": 111 }, { "epoch": 0.7411083540115798, "grad_norm": 1.2292939808383032, "learning_rate": 4.7333895186693445e-05, "loss": 1.401, "num_input_tokens_seen": 29360128, "step": 112 }, { "epoch": 0.7477253928866832, "grad_norm": 1.0821916096299973, "learning_rate": 4.728695778203505e-05, "loss": 1.3394, "num_input_tokens_seen": 29622272, "step": 113 }, { "epoch": 0.7543424317617866, "grad_norm": 1.2136622411407614, "learning_rate": 4.723963449405709e-05, "loss": 1.3368, "num_input_tokens_seen": 29884416, "step": 114 }, { "epoch": 0.76095947063689, "grad_norm": 1.0757650949421957, "learning_rate": 4.719192614212969e-05, "loss": 1.3414, "num_input_tokens_seen": 30146560, "step": 115 }, { "epoch": 0.7675765095119934, "grad_norm": 1.2630564279613965, "learning_rate": 4.7143833552290104e-05, "loss": 1.352, "num_input_tokens_seen": 30408704, "step": 116 }, { "epoch": 0.7741935483870968, "grad_norm": 1.1325930683405476, "learning_rate": 4.709535755722839e-05, "loss": 1.3444, "num_input_tokens_seen": 30670848, "step": 117 }, { "epoch": 0.7808105872622002, "grad_norm": 1.2210863008246031, "learning_rate": 4.704649899627299e-05, "loss": 1.3464, "num_input_tokens_seen": 30932992, "step": 118 }, { "epoch": 0.7874276261373035, "grad_norm": 1.1397504462999077, "learning_rate": 4.6997258715376234e-05, "loss": 1.3521, "num_input_tokens_seen": 31195136, "step": 119 }, { "epoch": 0.794044665012407, "grad_norm": 1.2115843784948899, "learning_rate": 4.694763756709967e-05, "loss": 1.3534, "num_input_tokens_seen": 31457280, "step": 120 }, { "epoch": 0.8006617038875103, "grad_norm": 1.399225473111164, "learning_rate": 4.689763641059929e-05, "loss": 1.3368, "num_input_tokens_seen": 31719424, "step": 121 }, { "epoch": 0.8072787427626137, "grad_norm": 1.0957293580137375, "learning_rate": 4.684725611161067e-05, "loss": 1.3561, "num_input_tokens_seen": 31981568, "step": 122 }, { "epoch": 0.8138957816377171, "grad_norm": 1.1475976465568731, "learning_rate": 4.679649754243398e-05, "loss": 1.335, "num_input_tokens_seen": 32243712, "step": 123 }, { "epoch": 0.8205128205128205, "grad_norm": 1.0767288647528503, "learning_rate": 4.6745361581918866e-05, "loss": 1.3462, "num_input_tokens_seen": 32505856, "step": 124 }, { "epoch": 0.8271298593879239, "grad_norm": 1.1053947900183125, "learning_rate": 4.669384911544927e-05, "loss": 1.3474, "num_input_tokens_seen": 32768000, "step": 125 }, { "epoch": 0.8337468982630273, "grad_norm": 0.9746263899049297, "learning_rate": 4.664196103492805e-05, "loss": 1.3464, "num_input_tokens_seen": 33030144, "step": 126 }, { "epoch": 0.8403639371381307, "grad_norm": 1.1843673694125876, "learning_rate": 4.658969823876157e-05, "loss": 1.3218, "num_input_tokens_seen": 33292288, "step": 127 }, { "epoch": 0.8469809760132341, "grad_norm": 1.0675619106578007, "learning_rate": 4.6537061631844144e-05, "loss": 1.3615, "num_input_tokens_seen": 33554432, "step": 128 }, { "epoch": 0.8535980148883374, "grad_norm": 1.1082771918099024, "learning_rate": 4.648405212554236e-05, "loss": 1.3234, "num_input_tokens_seen": 33816576, "step": 129 }, { "epoch": 0.8602150537634409, "grad_norm": 1.0533705567853635, "learning_rate": 4.6430670637679295e-05, "loss": 1.3131, "num_input_tokens_seen": 34078720, "step": 130 }, { "epoch": 0.8668320926385442, "grad_norm": 1.3250710828705743, "learning_rate": 4.637691809251863e-05, "loss": 1.3449, "num_input_tokens_seen": 34340864, "step": 131 }, { "epoch": 0.8734491315136477, "grad_norm": 1.0631490766176317, "learning_rate": 4.6322795420748664e-05, "loss": 1.3372, "num_input_tokens_seen": 34603008, "step": 132 }, { "epoch": 0.880066170388751, "grad_norm": 1.1978298622584882, "learning_rate": 4.626830355946616e-05, "loss": 1.3545, "num_input_tokens_seen": 34865152, "step": 133 }, { "epoch": 0.8866832092638545, "grad_norm": 1.2388970968450599, "learning_rate": 4.621344345216017e-05, "loss": 1.3345, "num_input_tokens_seen": 35127296, "step": 134 }, { "epoch": 0.8933002481389578, "grad_norm": 1.1244455375131996, "learning_rate": 4.615821604869564e-05, "loss": 1.3314, "num_input_tokens_seen": 35389440, "step": 135 }, { "epoch": 0.8999172870140613, "grad_norm": 1.0226285228765448, "learning_rate": 4.6102622305297015e-05, "loss": 1.3437, "num_input_tokens_seen": 35651584, "step": 136 }, { "epoch": 0.9065343258891646, "grad_norm": 1.179215839596238, "learning_rate": 4.604666318453167e-05, "loss": 1.3109, "num_input_tokens_seen": 35913728, "step": 137 }, { "epoch": 0.913151364764268, "grad_norm": 1.0205108272754437, "learning_rate": 4.5990339655293215e-05, "loss": 1.3061, "num_input_tokens_seen": 36175872, "step": 138 }, { "epoch": 0.9197684036393714, "grad_norm": 1.1432719000110645, "learning_rate": 4.593365269278477e-05, "loss": 1.3036, "num_input_tokens_seen": 36438016, "step": 139 }, { "epoch": 0.9263854425144747, "grad_norm": 1.0822518315588558, "learning_rate": 4.587660327850203e-05, "loss": 1.3078, "num_input_tokens_seen": 36700160, "step": 140 }, { "epoch": 0.9330024813895782, "grad_norm": 1.2554416669389759, "learning_rate": 4.581919240021629e-05, "loss": 1.3116, "num_input_tokens_seen": 36962304, "step": 141 }, { "epoch": 0.9396195202646815, "grad_norm": 1.2559908169183205, "learning_rate": 4.576142105195737e-05, "loss": 1.3215, "num_input_tokens_seen": 37224448, "step": 142 }, { "epoch": 0.946236559139785, "grad_norm": 1.1886853683206, "learning_rate": 4.570329023399636e-05, "loss": 1.325, "num_input_tokens_seen": 37486592, "step": 143 }, { "epoch": 0.9528535980148883, "grad_norm": 1.1590288482384423, "learning_rate": 4.564480095282832e-05, "loss": 1.3397, "num_input_tokens_seen": 37748736, "step": 144 }, { "epoch": 0.9594706368899917, "grad_norm": 1.2963117712374421, "learning_rate": 4.5585954221154856e-05, "loss": 1.3371, "num_input_tokens_seen": 38010880, "step": 145 }, { "epoch": 0.9660876757650951, "grad_norm": 1.1183668865567917, "learning_rate": 4.552675105786659e-05, "loss": 1.2819, "num_input_tokens_seen": 38273024, "step": 146 }, { "epoch": 0.9727047146401985, "grad_norm": 1.4680624474377164, "learning_rate": 4.546719248802551e-05, "loss": 1.2962, "num_input_tokens_seen": 38535168, "step": 147 }, { "epoch": 0.9793217535153019, "grad_norm": 1.0568661810526727, "learning_rate": 4.540727954284721e-05, "loss": 1.3416, "num_input_tokens_seen": 38797312, "step": 148 }, { "epoch": 0.9859387923904053, "grad_norm": 1.3165517826941278, "learning_rate": 4.534701325968308e-05, "loss": 1.3301, "num_input_tokens_seen": 39059456, "step": 149 }, { "epoch": 0.9925558312655087, "grad_norm": 1.0620702266856872, "learning_rate": 4.528639468200226e-05, "loss": 1.3065, "num_input_tokens_seen": 39321600, "step": 150 }, { "epoch": 0.9991728701406121, "grad_norm": 1.0129869486885617, "learning_rate": 4.522542485937369e-05, "loss": 1.2729, "num_input_tokens_seen": 39583744, "step": 151 }, { "epoch": 1.0, "grad_norm": 1.0129869486885617, "learning_rate": 4.5164104847447825e-05, "loss": 1.2598, "num_input_tokens_seen": 39616512, "step": 152 }, { "epoch": 1.0066170388751035, "grad_norm": 3.2796835203295367, "learning_rate": 4.5102435707938434e-05, "loss": 0.929, "num_input_tokens_seen": 39878656, "step": 153 }, { "epoch": 1.0132340777502067, "grad_norm": 1.6915018216454805, "learning_rate": 4.5040418508604185e-05, "loss": 0.9083, "num_input_tokens_seen": 40140800, "step": 154 }, { "epoch": 1.0198511166253101, "grad_norm": 1.6354997706408312, "learning_rate": 4.497805432323015e-05, "loss": 0.9349, "num_input_tokens_seen": 40402944, "step": 155 }, { "epoch": 1.0264681555004136, "grad_norm": 1.7763647403929517, "learning_rate": 4.491534423160923e-05, "loss": 0.9363, "num_input_tokens_seen": 40665088, "step": 156 }, { "epoch": 1.033085194375517, "grad_norm": 1.574116875627686, "learning_rate": 4.485228931952347e-05, "loss": 0.9249, "num_input_tokens_seen": 40927232, "step": 157 }, { "epoch": 1.0397022332506203, "grad_norm": 1.4461856174287244, "learning_rate": 4.4788890678725224e-05, "loss": 0.8876, "num_input_tokens_seen": 41189376, "step": 158 }, { "epoch": 1.0463192721257237, "grad_norm": 1.43842705918855, "learning_rate": 4.472514940691828e-05, "loss": 0.8882, "num_input_tokens_seen": 41451520, "step": 159 }, { "epoch": 1.0529363110008272, "grad_norm": 1.2624518837851444, "learning_rate": 4.466106660773885e-05, "loss": 0.8947, "num_input_tokens_seen": 41713664, "step": 160 }, { "epoch": 1.0595533498759304, "grad_norm": 1.3665729188837292, "learning_rate": 4.4596643390736444e-05, "loss": 0.8734, "num_input_tokens_seen": 41975808, "step": 161 }, { "epoch": 1.0661703887510339, "grad_norm": 1.3002771069796804, "learning_rate": 4.4531880871354683e-05, "loss": 0.8626, "num_input_tokens_seen": 42237952, "step": 162 }, { "epoch": 1.0727874276261373, "grad_norm": 1.2690537305241547, "learning_rate": 4.446678017091198e-05, "loss": 0.8404, "num_input_tokens_seen": 42500096, "step": 163 }, { "epoch": 1.0794044665012408, "grad_norm": 1.330267070475486, "learning_rate": 4.4401342416582106e-05, "loss": 0.8663, "num_input_tokens_seen": 42762240, "step": 164 }, { "epoch": 1.086021505376344, "grad_norm": 1.2652249101149649, "learning_rate": 4.43355687413747e-05, "loss": 0.8591, "num_input_tokens_seen": 43024384, "step": 165 }, { "epoch": 1.0926385442514475, "grad_norm": 1.303585645944488, "learning_rate": 4.4269460284115624e-05, "loss": 0.8794, "num_input_tokens_seen": 43286528, "step": 166 }, { "epoch": 1.099255583126551, "grad_norm": 1.1211778159345078, "learning_rate": 4.420301818942728e-05, "loss": 0.8565, "num_input_tokens_seen": 43548672, "step": 167 }, { "epoch": 1.1058726220016544, "grad_norm": 1.258025084841763, "learning_rate": 4.413624360770876e-05, "loss": 0.8905, "num_input_tokens_seen": 43810816, "step": 168 }, { "epoch": 1.1124896608767576, "grad_norm": 1.125309033988079, "learning_rate": 4.406913769511594e-05, "loss": 0.8448, "num_input_tokens_seen": 44072960, "step": 169 }, { "epoch": 1.119106699751861, "grad_norm": 1.1784006185919231, "learning_rate": 4.4001701613541456e-05, "loss": 0.8488, "num_input_tokens_seen": 44335104, "step": 170 }, { "epoch": 1.1257237386269645, "grad_norm": 1.2604324137156369, "learning_rate": 4.393393653059462e-05, "loss": 0.8803, "num_input_tokens_seen": 44597248, "step": 171 }, { "epoch": 1.1323407775020677, "grad_norm": 1.4523529165761135, "learning_rate": 4.386584361958115e-05, "loss": 0.856, "num_input_tokens_seen": 44859392, "step": 172 }, { "epoch": 1.1389578163771712, "grad_norm": 1.2839516251773848, "learning_rate": 4.379742405948288e-05, "loss": 0.8392, "num_input_tokens_seen": 45121536, "step": 173 }, { "epoch": 1.1455748552522746, "grad_norm": 1.1643605452939796, "learning_rate": 4.372867903493737e-05, "loss": 0.8649, "num_input_tokens_seen": 45383680, "step": 174 }, { "epoch": 1.152191894127378, "grad_norm": 1.2437437966014462, "learning_rate": 4.3659609736217344e-05, "loss": 0.8392, "num_input_tokens_seen": 45645824, "step": 175 }, { "epoch": 1.1588089330024813, "grad_norm": 1.258435176144592, "learning_rate": 4.359021735921013e-05, "loss": 0.8563, "num_input_tokens_seen": 45907968, "step": 176 }, { "epoch": 1.1654259718775848, "grad_norm": 1.9852111216718347, "learning_rate": 4.352050310539694e-05, "loss": 0.859, "num_input_tokens_seen": 46170112, "step": 177 }, { "epoch": 1.1720430107526882, "grad_norm": 1.3877250659239455, "learning_rate": 4.345046818183203e-05, "loss": 0.8804, "num_input_tokens_seen": 46432256, "step": 178 }, { "epoch": 1.1786600496277915, "grad_norm": 1.1698140955270617, "learning_rate": 4.3380113801121854e-05, "loss": 0.8729, "num_input_tokens_seen": 46694400, "step": 179 }, { "epoch": 1.185277088502895, "grad_norm": 1.3931561318052186, "learning_rate": 4.330944118140407e-05, "loss": 0.8741, "num_input_tokens_seen": 46956544, "step": 180 }, { "epoch": 1.1918941273779984, "grad_norm": 1.4096016502064548, "learning_rate": 4.3238451546326367e-05, "loss": 0.8826, "num_input_tokens_seen": 47218688, "step": 181 }, { "epoch": 1.1985111662531018, "grad_norm": 1.071563835965369, "learning_rate": 4.3167146125025385e-05, "loss": 0.8786, "num_input_tokens_seen": 47480832, "step": 182 }, { "epoch": 1.205128205128205, "grad_norm": 1.4535954622349747, "learning_rate": 4.309552615210536e-05, "loss": 0.844, "num_input_tokens_seen": 47742976, "step": 183 }, { "epoch": 1.2117452440033085, "grad_norm": 1.1956562696985422, "learning_rate": 4.302359286761679e-05, "loss": 0.8603, "num_input_tokens_seen": 48005120, "step": 184 }, { "epoch": 1.218362282878412, "grad_norm": 1.4990950568465007, "learning_rate": 4.295134751703493e-05, "loss": 0.8361, "num_input_tokens_seen": 48267264, "step": 185 }, { "epoch": 1.2249793217535152, "grad_norm": 1.3042279096165876, "learning_rate": 4.287879135123826e-05, "loss": 0.8707, "num_input_tokens_seen": 48529408, "step": 186 }, { "epoch": 1.2315963606286187, "grad_norm": 1.4236645671572632, "learning_rate": 4.2805925626486796e-05, "loss": 0.8599, "num_input_tokens_seen": 48791552, "step": 187 }, { "epoch": 1.2382133995037221, "grad_norm": 1.1684645721401956, "learning_rate": 4.2732751604400364e-05, "loss": 0.8643, "num_input_tokens_seen": 49053696, "step": 188 }, { "epoch": 1.2448304383788256, "grad_norm": 1.3489623217080666, "learning_rate": 4.2659270551936756e-05, "loss": 0.8214, "num_input_tokens_seen": 49315840, "step": 189 }, { "epoch": 1.2514474772539288, "grad_norm": 1.1325213775166751, "learning_rate": 4.2585483741369755e-05, "loss": 0.8434, "num_input_tokens_seen": 49577984, "step": 190 }, { "epoch": 1.2580645161290323, "grad_norm": 1.6126949606880288, "learning_rate": 4.251139245026716e-05, "loss": 0.8657, "num_input_tokens_seen": 49840128, "step": 191 }, { "epoch": 1.2646815550041357, "grad_norm": 1.3428951116543568, "learning_rate": 4.243699796146863e-05, "loss": 0.8482, "num_input_tokens_seen": 50102272, "step": 192 }, { "epoch": 1.271298593879239, "grad_norm": 1.4845069023405626, "learning_rate": 4.236230156306348e-05, "loss": 0.8668, "num_input_tokens_seen": 50364416, "step": 193 }, { "epoch": 1.2779156327543424, "grad_norm": 1.3390183725385267, "learning_rate": 4.2287304548368386e-05, "loss": 0.8463, "num_input_tokens_seen": 50626560, "step": 194 }, { "epoch": 1.2845326716294458, "grad_norm": 1.3393923999288566, "learning_rate": 4.2212008215905e-05, "loss": 0.8861, "num_input_tokens_seen": 50888704, "step": 195 }, { "epoch": 1.2911497105045493, "grad_norm": 1.3594637430884577, "learning_rate": 4.213641386937743e-05, "loss": 0.8411, "num_input_tokens_seen": 51150848, "step": 196 }, { "epoch": 1.2977667493796525, "grad_norm": 1.354872266010071, "learning_rate": 4.206052281764973e-05, "loss": 0.8913, "num_input_tokens_seen": 51412992, "step": 197 }, { "epoch": 1.304383788254756, "grad_norm": 1.480084514475744, "learning_rate": 4.198433637472314e-05, "loss": 0.8932, "num_input_tokens_seen": 51675136, "step": 198 }, { "epoch": 1.3110008271298594, "grad_norm": 1.2179508694377172, "learning_rate": 4.190785585971346e-05, "loss": 0.8565, "num_input_tokens_seen": 51937280, "step": 199 }, { "epoch": 1.3176178660049627, "grad_norm": 1.2588919604551994, "learning_rate": 4.1831082596828106e-05, "loss": 0.8268, "num_input_tokens_seen": 52199424, "step": 200 }, { "epoch": 1.3242349048800661, "grad_norm": 1.2893099525717986, "learning_rate": 4.1754017915343234e-05, "loss": 0.8533, "num_input_tokens_seen": 52461568, "step": 201 }, { "epoch": 1.3308519437551696, "grad_norm": 1.2570685698759807, "learning_rate": 4.167666314958071e-05, "loss": 0.8529, "num_input_tokens_seen": 52723712, "step": 202 }, { "epoch": 1.337468982630273, "grad_norm": 1.3910303184986648, "learning_rate": 4.159901963888502e-05, "loss": 0.8777, "num_input_tokens_seen": 52985856, "step": 203 }, { "epoch": 1.3440860215053765, "grad_norm": 1.222611343425846, "learning_rate": 4.152108872760004e-05, "loss": 0.8855, "num_input_tokens_seen": 53248000, "step": 204 }, { "epoch": 1.3507030603804797, "grad_norm": 1.1899088358775345, "learning_rate": 4.144287176504582e-05, "loss": 0.8781, "num_input_tokens_seen": 53510144, "step": 205 }, { "epoch": 1.3573200992555832, "grad_norm": 1.117237186516348, "learning_rate": 4.136437010549518e-05, "loss": 0.8227, "num_input_tokens_seen": 53772288, "step": 206 }, { "epoch": 1.3639371381306864, "grad_norm": 1.212096841752928, "learning_rate": 4.1285585108150273e-05, "loss": 0.8606, "num_input_tokens_seen": 54034432, "step": 207 }, { "epoch": 1.3705541770057899, "grad_norm": 1.2208761945851812, "learning_rate": 4.120651813711905e-05, "loss": 0.8681, "num_input_tokens_seen": 54296576, "step": 208 }, { "epoch": 1.3771712158808933, "grad_norm": 1.2204263612725177, "learning_rate": 4.112717056139164e-05, "loss": 0.8655, "num_input_tokens_seen": 54558720, "step": 209 }, { "epoch": 1.3837882547559968, "grad_norm": 1.3841625088026035, "learning_rate": 4.104754375481664e-05, "loss": 0.8752, "num_input_tokens_seen": 54820864, "step": 210 }, { "epoch": 1.3904052936311002, "grad_norm": 1.291256777563656, "learning_rate": 4.096763909607737e-05, "loss": 0.8605, "num_input_tokens_seen": 55083008, "step": 211 }, { "epoch": 1.3970223325062034, "grad_norm": 1.1816906512953695, "learning_rate": 4.088745796866793e-05, "loss": 0.9004, "num_input_tokens_seen": 55345152, "step": 212 }, { "epoch": 1.403639371381307, "grad_norm": 1.0937124460607777, "learning_rate": 4.08070017608693e-05, "loss": 0.8701, "num_input_tokens_seen": 55607296, "step": 213 }, { "epoch": 1.4102564102564101, "grad_norm": 1.1192173919656867, "learning_rate": 4.072627186572531e-05, "loss": 0.8665, "num_input_tokens_seen": 55869440, "step": 214 }, { "epoch": 1.4168734491315136, "grad_norm": 1.2027417701165573, "learning_rate": 4.064526968101844e-05, "loss": 0.8508, "num_input_tokens_seen": 56131584, "step": 215 }, { "epoch": 1.423490488006617, "grad_norm": 1.167715206363498, "learning_rate": 4.056399660924575e-05, "loss": 0.8425, "num_input_tokens_seen": 56393728, "step": 216 }, { "epoch": 1.4301075268817205, "grad_norm": 1.1346595160468362, "learning_rate": 4.048245405759446e-05, "loss": 0.8638, "num_input_tokens_seen": 56655872, "step": 217 }, { "epoch": 1.436724565756824, "grad_norm": 1.197370470624576, "learning_rate": 4.040064343791767e-05, "loss": 0.8424, "num_input_tokens_seen": 56918016, "step": 218 }, { "epoch": 1.4433416046319272, "grad_norm": 1.111136180713314, "learning_rate": 4.0318566166709925e-05, "loss": 0.8734, "num_input_tokens_seen": 57180160, "step": 219 }, { "epoch": 1.4499586435070306, "grad_norm": 1.125826256539522, "learning_rate": 4.023622366508261e-05, "loss": 0.8303, "num_input_tokens_seen": 57442304, "step": 220 }, { "epoch": 1.4565756823821339, "grad_norm": 1.1310355381266508, "learning_rate": 4.0153617358739406e-05, "loss": 0.8845, "num_input_tokens_seen": 57704448, "step": 221 }, { "epoch": 1.4631927212572373, "grad_norm": 1.1677707737228877, "learning_rate": 4.0070748677951605e-05, "loss": 0.8556, "num_input_tokens_seen": 57966592, "step": 222 }, { "epoch": 1.4698097601323408, "grad_norm": 1.1528540778585532, "learning_rate": 3.998761905753333e-05, "loss": 0.8533, "num_input_tokens_seen": 58228736, "step": 223 }, { "epoch": 1.4764267990074442, "grad_norm": 1.253460257564942, "learning_rate": 3.9904229936816674e-05, "loss": 0.8773, "num_input_tokens_seen": 58490880, "step": 224 }, { "epoch": 1.4830438378825477, "grad_norm": 1.1496232622442173, "learning_rate": 3.9820582759626825e-05, "loss": 0.8899, "num_input_tokens_seen": 58753024, "step": 225 }, { "epoch": 1.489660876757651, "grad_norm": 1.2995455269062046, "learning_rate": 3.973667897425701e-05, "loss": 0.8717, "num_input_tokens_seen": 59015168, "step": 226 }, { "epoch": 1.4962779156327544, "grad_norm": 1.144101297289898, "learning_rate": 3.9652520033443485e-05, "loss": 0.849, "num_input_tokens_seen": 59277312, "step": 227 }, { "epoch": 1.5028949545078576, "grad_norm": 1.5210835767609536, "learning_rate": 3.956810739434032e-05, "loss": 0.8796, "num_input_tokens_seen": 59539456, "step": 228 }, { "epoch": 1.509511993382961, "grad_norm": 1.313822814756041, "learning_rate": 3.948344251849421e-05, "loss": 0.8912, "num_input_tokens_seen": 59801600, "step": 229 }, { "epoch": 1.5161290322580645, "grad_norm": 1.1203755838084284, "learning_rate": 3.9398526871819154e-05, "loss": 0.86, "num_input_tokens_seen": 60063744, "step": 230 }, { "epoch": 1.522746071133168, "grad_norm": 1.2939167541224808, "learning_rate": 3.931336192457106e-05, "loss": 0.8617, "num_input_tokens_seen": 60325888, "step": 231 }, { "epoch": 1.5293631100082714, "grad_norm": 1.0601911330047828, "learning_rate": 3.9227949151322326e-05, "loss": 0.8474, "num_input_tokens_seen": 60588032, "step": 232 }, { "epoch": 1.5359801488833746, "grad_norm": 1.3013451522634765, "learning_rate": 3.914229003093627e-05, "loss": 0.8394, "num_input_tokens_seen": 60850176, "step": 233 }, { "epoch": 1.542597187758478, "grad_norm": 1.0264196280817957, "learning_rate": 3.905638604654156e-05, "loss": 0.8639, "num_input_tokens_seen": 61112320, "step": 234 }, { "epoch": 1.5492142266335813, "grad_norm": 1.170686578102918, "learning_rate": 3.897023868550649e-05, "loss": 0.8656, "num_input_tokens_seen": 61374464, "step": 235 }, { "epoch": 1.5558312655086848, "grad_norm": 1.1035633097086974, "learning_rate": 3.8883849439413265e-05, "loss": 0.8839, "num_input_tokens_seen": 61636608, "step": 236 }, { "epoch": 1.5624483043837882, "grad_norm": 1.0804813512718974, "learning_rate": 3.879721980403217e-05, "loss": 0.8833, "num_input_tokens_seen": 61898752, "step": 237 }, { "epoch": 1.5690653432588917, "grad_norm": 1.1077709384095282, "learning_rate": 3.871035127929566e-05, "loss": 0.8818, "num_input_tokens_seen": 62160896, "step": 238 }, { "epoch": 1.5756823821339951, "grad_norm": 1.182231681118449, "learning_rate": 3.8623245369272385e-05, "loss": 0.8514, "num_input_tokens_seen": 62423040, "step": 239 }, { "epoch": 1.5822994210090986, "grad_norm": 1.2617220805970901, "learning_rate": 3.853590358214119e-05, "loss": 0.8688, "num_input_tokens_seen": 62685184, "step": 240 }, { "epoch": 1.5889164598842018, "grad_norm": 1.1349928239204545, "learning_rate": 3.844832743016491e-05, "loss": 0.8636, "num_input_tokens_seen": 62947328, "step": 241 }, { "epoch": 1.595533498759305, "grad_norm": 1.2926042995530984, "learning_rate": 3.83605184296643e-05, "loss": 0.8556, "num_input_tokens_seen": 63209472, "step": 242 }, { "epoch": 1.6021505376344085, "grad_norm": 1.111484999792133, "learning_rate": 3.8272478100991714e-05, "loss": 0.8908, "num_input_tokens_seen": 63471616, "step": 243 }, { "epoch": 1.608767576509512, "grad_norm": 1.2640736812779294, "learning_rate": 3.818420796850478e-05, "loss": 0.8654, "num_input_tokens_seen": 63733760, "step": 244 }, { "epoch": 1.6153846153846154, "grad_norm": 1.005896566039546, "learning_rate": 3.809570956054004e-05, "loss": 0.8693, "num_input_tokens_seen": 63995904, "step": 245 }, { "epoch": 1.6220016542597189, "grad_norm": 1.3829300626223842, "learning_rate": 3.800698440938644e-05, "loss": 0.8818, "num_input_tokens_seen": 64258048, "step": 246 }, { "epoch": 1.6286186931348223, "grad_norm": 1.0255361852704923, "learning_rate": 3.791803405125885e-05, "loss": 0.8915, "num_input_tokens_seen": 64520192, "step": 247 }, { "epoch": 1.6352357320099256, "grad_norm": 1.1092615011275497, "learning_rate": 3.782886002627145e-05, "loss": 0.8551, "num_input_tokens_seen": 64782336, "step": 248 }, { "epoch": 1.6418527708850288, "grad_norm": 1.1366146935597203, "learning_rate": 3.773946387841103e-05, "loss": 0.8436, "num_input_tokens_seen": 65044480, "step": 249 }, { "epoch": 1.6484698097601322, "grad_norm": 1.1579215508983878, "learning_rate": 3.764984715551032e-05, "loss": 0.8664, "num_input_tokens_seen": 65306624, "step": 250 }, { "epoch": 1.6550868486352357, "grad_norm": 1.1104612323014593, "learning_rate": 3.756001140922112e-05, "loss": 0.8453, "num_input_tokens_seen": 65568768, "step": 251 }, { "epoch": 1.6617038875103392, "grad_norm": 1.1393363090562745, "learning_rate": 3.74699581949875e-05, "loss": 0.8932, "num_input_tokens_seen": 65830912, "step": 252 }, { "epoch": 1.6683209263854426, "grad_norm": 1.1318034968552904, "learning_rate": 3.737968907201882e-05, "loss": 0.8922, "num_input_tokens_seen": 66093056, "step": 253 }, { "epoch": 1.674937965260546, "grad_norm": 1.072676962267901, "learning_rate": 3.728920560326275e-05, "loss": 0.8708, "num_input_tokens_seen": 66355200, "step": 254 }, { "epoch": 1.6815550041356493, "grad_norm": 1.0759090742363742, "learning_rate": 3.719850935537821e-05, "loss": 0.8904, "num_input_tokens_seen": 66617344, "step": 255 }, { "epoch": 1.6881720430107527, "grad_norm": 1.1591455124390417, "learning_rate": 3.710760189870825e-05, "loss": 0.8846, "num_input_tokens_seen": 66879488, "step": 256 }, { "epoch": 1.694789081885856, "grad_norm": 1.048864507790183, "learning_rate": 3.701648480725286e-05, "loss": 0.8906, "num_input_tokens_seen": 67141632, "step": 257 }, { "epoch": 1.7014061207609594, "grad_norm": 1.0992102136364237, "learning_rate": 3.692515965864169e-05, "loss": 0.8653, "num_input_tokens_seen": 67403776, "step": 258 }, { "epoch": 1.7080231596360629, "grad_norm": 1.1084880653140272, "learning_rate": 3.683362803410678e-05, "loss": 0.9049, "num_input_tokens_seen": 67665920, "step": 259 }, { "epoch": 1.7146401985111663, "grad_norm": 1.0193571027301067, "learning_rate": 3.674189151845515e-05, "loss": 0.8668, "num_input_tokens_seen": 67928064, "step": 260 }, { "epoch": 1.7212572373862698, "grad_norm": 1.1435915369787901, "learning_rate": 3.6649951700041366e-05, "loss": 0.85, "num_input_tokens_seen": 68190208, "step": 261 }, { "epoch": 1.727874276261373, "grad_norm": 1.0624020069945779, "learning_rate": 3.6557810170740034e-05, "loss": 0.8833, "num_input_tokens_seen": 68452352, "step": 262 }, { "epoch": 1.7344913151364765, "grad_norm": 1.1397221255920662, "learning_rate": 3.646546852591827e-05, "loss": 0.8979, "num_input_tokens_seen": 68714496, "step": 263 }, { "epoch": 1.7411083540115797, "grad_norm": 1.0934989352817812, "learning_rate": 3.637292836440802e-05, "loss": 0.8789, "num_input_tokens_seen": 68976640, "step": 264 }, { "epoch": 1.7477253928866832, "grad_norm": 1.1270519325330435, "learning_rate": 3.6280191288478436e-05, "loss": 0.8738, "num_input_tokens_seen": 69238784, "step": 265 }, { "epoch": 1.7543424317617866, "grad_norm": 1.1064043030347466, "learning_rate": 3.61872589038081e-05, "loss": 0.9082, "num_input_tokens_seen": 69500928, "step": 266 }, { "epoch": 1.76095947063689, "grad_norm": 1.0016528471988886, "learning_rate": 3.6094132819457205e-05, "loss": 0.8879, "num_input_tokens_seen": 69763072, "step": 267 }, { "epoch": 1.7675765095119935, "grad_norm": 1.1490777889864239, "learning_rate": 3.600081464783977e-05, "loss": 0.9091, "num_input_tokens_seen": 70025216, "step": 268 }, { "epoch": 1.7741935483870968, "grad_norm": 0.9851189833580496, "learning_rate": 3.5907306004695636e-05, "loss": 0.8406, "num_input_tokens_seen": 70287360, "step": 269 }, { "epoch": 1.7808105872622002, "grad_norm": 1.0441019161167733, "learning_rate": 3.5813608509062526e-05, "loss": 0.8573, "num_input_tokens_seen": 70549504, "step": 270 }, { "epoch": 1.7874276261373034, "grad_norm": 1.0021575087386156, "learning_rate": 3.5719723783248045e-05, "loss": 0.8901, "num_input_tokens_seen": 70811648, "step": 271 }, { "epoch": 1.794044665012407, "grad_norm": 1.0004983205114817, "learning_rate": 3.5625653452801525e-05, "loss": 0.8643, "num_input_tokens_seen": 71073792, "step": 272 }, { "epoch": 1.8006617038875103, "grad_norm": 1.0020219633694651, "learning_rate": 3.553139914648593e-05, "loss": 0.8372, "num_input_tokens_seen": 71335936, "step": 273 }, { "epoch": 1.8072787427626138, "grad_norm": 1.089710937917061, "learning_rate": 3.543696249624965e-05, "loss": 0.8566, "num_input_tokens_seen": 71598080, "step": 274 }, { "epoch": 1.8138957816377173, "grad_norm": 0.9963288164384833, "learning_rate": 3.534234513719821e-05, "loss": 0.867, "num_input_tokens_seen": 71860224, "step": 275 }, { "epoch": 1.8205128205128205, "grad_norm": 1.1734651632685695, "learning_rate": 3.5247548707565986e-05, "loss": 0.8618, "num_input_tokens_seen": 72122368, "step": 276 }, { "epoch": 1.827129859387924, "grad_norm": 1.0888558238501158, "learning_rate": 3.5152574848687875e-05, "loss": 0.8733, "num_input_tokens_seen": 72384512, "step": 277 }, { "epoch": 1.8337468982630272, "grad_norm": 1.0461245836544661, "learning_rate": 3.505742520497079e-05, "loss": 0.8677, "num_input_tokens_seen": 72646656, "step": 278 }, { "epoch": 1.8403639371381306, "grad_norm": 1.086689279678754, "learning_rate": 3.496210142386527e-05, "loss": 0.8389, "num_input_tokens_seen": 72908800, "step": 279 }, { "epoch": 1.846980976013234, "grad_norm": 1.1718920032890414, "learning_rate": 3.4866605155836915e-05, "loss": 0.8909, "num_input_tokens_seen": 73170944, "step": 280 }, { "epoch": 1.8535980148883375, "grad_norm": 1.1057752723633063, "learning_rate": 3.47709380543378e-05, "loss": 0.8844, "num_input_tokens_seen": 73433088, "step": 281 }, { "epoch": 1.860215053763441, "grad_norm": 1.098019821064288, "learning_rate": 3.46751017757779e-05, "loss": 0.8717, "num_input_tokens_seen": 73695232, "step": 282 }, { "epoch": 1.8668320926385442, "grad_norm": 1.0393178044085931, "learning_rate": 3.457909797949633e-05, "loss": 0.8801, "num_input_tokens_seen": 73957376, "step": 283 }, { "epoch": 1.8734491315136477, "grad_norm": 1.2431012129822678, "learning_rate": 3.448292832773269e-05, "loss": 0.8604, "num_input_tokens_seen": 74219520, "step": 284 }, { "epoch": 1.880066170388751, "grad_norm": 1.019422889343279, "learning_rate": 3.438659448559825e-05, "loss": 0.8681, "num_input_tokens_seen": 74481664, "step": 285 }, { "epoch": 1.8866832092638544, "grad_norm": 1.2811321592728038, "learning_rate": 3.4290098121047114e-05, "loss": 0.8734, "num_input_tokens_seen": 74743808, "step": 286 }, { "epoch": 1.8933002481389578, "grad_norm": 1.070768836122241, "learning_rate": 3.419344090484736e-05, "loss": 0.8605, "num_input_tokens_seen": 75005952, "step": 287 }, { "epoch": 1.8999172870140613, "grad_norm": 1.1946653702822796, "learning_rate": 3.409662451055208e-05, "loss": 0.8714, "num_input_tokens_seen": 75268096, "step": 288 }, { "epoch": 1.9065343258891647, "grad_norm": 1.140200266566108, "learning_rate": 3.3999650614470445e-05, "loss": 0.8573, "num_input_tokens_seen": 75530240, "step": 289 }, { "epoch": 1.913151364764268, "grad_norm": 1.121817062197686, "learning_rate": 3.390252089563867e-05, "loss": 0.8889, "num_input_tokens_seen": 75792384, "step": 290 }, { "epoch": 1.9197684036393714, "grad_norm": 1.1795614581286045, "learning_rate": 3.3805237035790924e-05, "loss": 0.8636, "num_input_tokens_seen": 76054528, "step": 291 }, { "epoch": 1.9263854425144746, "grad_norm": 1.1839809443148863, "learning_rate": 3.3707800719330184e-05, "loss": 0.8729, "num_input_tokens_seen": 76316672, "step": 292 }, { "epoch": 1.933002481389578, "grad_norm": 1.1302681193686077, "learning_rate": 3.361021363329917e-05, "loss": 0.8474, "num_input_tokens_seen": 76578816, "step": 293 }, { "epoch": 1.9396195202646815, "grad_norm": 0.9877234338487746, "learning_rate": 3.351247746735103e-05, "loss": 0.8728, "num_input_tokens_seen": 76840960, "step": 294 }, { "epoch": 1.946236559139785, "grad_norm": 1.112763917244249, "learning_rate": 3.341459391372016e-05, "loss": 0.8552, "num_input_tokens_seen": 77103104, "step": 295 }, { "epoch": 1.9528535980148884, "grad_norm": 1.0135909963895178, "learning_rate": 3.331656466719284e-05, "loss": 0.8591, "num_input_tokens_seen": 77365248, "step": 296 }, { "epoch": 1.9594706368899917, "grad_norm": 1.057962809664509, "learning_rate": 3.321839142507794e-05, "loss": 0.8703, "num_input_tokens_seen": 77627392, "step": 297 }, { "epoch": 1.9660876757650951, "grad_norm": 1.024489249713761, "learning_rate": 3.312007588717751e-05, "loss": 0.8433, "num_input_tokens_seen": 77889536, "step": 298 }, { "epoch": 1.9727047146401984, "grad_norm": 1.1626693600678064, "learning_rate": 3.302161975575736e-05, "loss": 0.8727, "num_input_tokens_seen": 78151680, "step": 299 }, { "epoch": 1.9793217535153018, "grad_norm": 0.9962455067400268, "learning_rate": 3.292302473551757e-05, "loss": 0.8487, "num_input_tokens_seen": 78413824, "step": 300 }, { "epoch": 1.9859387923904053, "grad_norm": 1.1129572616806531, "learning_rate": 3.2824292533562996e-05, "loss": 0.8782, "num_input_tokens_seen": 78675968, "step": 301 }, { "epoch": 1.9925558312655087, "grad_norm": 1.1359145638143653, "learning_rate": 3.272542485937369e-05, "loss": 0.8671, "num_input_tokens_seen": 78938112, "step": 302 }, { "epoch": 1.9991728701406122, "grad_norm": 1.0937608862202273, "learning_rate": 3.262642342477531e-05, "loss": 0.8609, "num_input_tokens_seen": 79200256, "step": 303 }, { "epoch": 2.0, "grad_norm": 1.0937608862202273, "learning_rate": 3.252728994390951e-05, "loss": 0.8026, "num_input_tokens_seen": 79233024, "step": 304 }, { "epoch": 2.0066170388751035, "grad_norm": 3.398274310113878, "learning_rate": 3.2428026133204184e-05, "loss": 0.4556, "num_input_tokens_seen": 79495168, "step": 305 }, { "epoch": 2.013234077750207, "grad_norm": 2.5271586821507293, "learning_rate": 3.232863371134385e-05, "loss": 0.4076, "num_input_tokens_seen": 79757312, "step": 306 }, { "epoch": 2.0198511166253104, "grad_norm": 1.783345206375513, "learning_rate": 3.2229114399239816e-05, "loss": 0.4099, "num_input_tokens_seen": 80019456, "step": 307 }, { "epoch": 2.0264681555004134, "grad_norm": 2.5054864860618133, "learning_rate": 3.21294699200004e-05, "loss": 0.3909, "num_input_tokens_seen": 80281600, "step": 308 }, { "epoch": 2.033085194375517, "grad_norm": 2.376970275167304, "learning_rate": 3.202970199890111e-05, "loss": 0.3925, "num_input_tokens_seen": 80543744, "step": 309 }, { "epoch": 2.0397022332506203, "grad_norm": 2.1738898999313485, "learning_rate": 3.1929812363354766e-05, "loss": 0.3738, "num_input_tokens_seen": 80805888, "step": 310 }, { "epoch": 2.0463192721257237, "grad_norm": 1.688931111731337, "learning_rate": 3.18298027428816e-05, "loss": 0.3376, "num_input_tokens_seen": 81068032, "step": 311 }, { "epoch": 2.052936311000827, "grad_norm": 1.4853385671984687, "learning_rate": 3.172967486907928e-05, "loss": 0.3595, "num_input_tokens_seen": 81330176, "step": 312 }, { "epoch": 2.0595533498759306, "grad_norm": 1.3248855510062758, "learning_rate": 3.1629430475592955e-05, "loss": 0.3749, "num_input_tokens_seen": 81592320, "step": 313 }, { "epoch": 2.066170388751034, "grad_norm": 1.4372594674243309, "learning_rate": 3.1529071298085236e-05, "loss": 0.361, "num_input_tokens_seen": 81854464, "step": 314 }, { "epoch": 2.072787427626137, "grad_norm": 1.1608848796138636, "learning_rate": 3.142859907420615e-05, "loss": 0.3521, "num_input_tokens_seen": 82116608, "step": 315 }, { "epoch": 2.0794044665012406, "grad_norm": 1.2207815470871421, "learning_rate": 3.132801554356303e-05, "loss": 0.329, "num_input_tokens_seen": 82378752, "step": 316 }, { "epoch": 2.086021505376344, "grad_norm": 1.3338094436906303, "learning_rate": 3.122732244769041e-05, "loss": 0.3486, "num_input_tokens_seen": 82640896, "step": 317 }, { "epoch": 2.0926385442514475, "grad_norm": 1.2172300343622242, "learning_rate": 3.1126521530019874e-05, "loss": 0.3487, "num_input_tokens_seen": 82903040, "step": 318 }, { "epoch": 2.099255583126551, "grad_norm": 1.4126997538766228, "learning_rate": 3.102561453584987e-05, "loss": 0.3437, "num_input_tokens_seen": 83165184, "step": 319 }, { "epoch": 2.1058726220016544, "grad_norm": 1.1853200893362168, "learning_rate": 3.092460321231547e-05, "loss": 0.3278, "num_input_tokens_seen": 83427328, "step": 320 }, { "epoch": 2.112489660876758, "grad_norm": 1.1715327320798274, "learning_rate": 3.0823489308358174e-05, "loss": 0.3086, "num_input_tokens_seen": 83689472, "step": 321 }, { "epoch": 2.119106699751861, "grad_norm": 1.0852203225918198, "learning_rate": 3.072227457469554e-05, "loss": 0.3321, "num_input_tokens_seen": 83951616, "step": 322 }, { "epoch": 2.1257237386269643, "grad_norm": 1.1506537447974439, "learning_rate": 3.062096076379097e-05, "loss": 0.3396, "num_input_tokens_seen": 84213760, "step": 323 }, { "epoch": 2.1323407775020677, "grad_norm": 1.2893107469955971, "learning_rate": 3.0519549629823286e-05, "loss": 0.3243, "num_input_tokens_seen": 84475904, "step": 324 }, { "epoch": 2.138957816377171, "grad_norm": 1.112482245858434, "learning_rate": 3.0418042928656414e-05, "loss": 0.3304, "num_input_tokens_seen": 84738048, "step": 325 }, { "epoch": 2.1455748552522746, "grad_norm": 1.0505114063288397, "learning_rate": 3.0316442417808954e-05, "loss": 0.3289, "num_input_tokens_seen": 85000192, "step": 326 }, { "epoch": 2.152191894127378, "grad_norm": 1.0702654196144072, "learning_rate": 3.0214749856423745e-05, "loss": 0.3114, "num_input_tokens_seen": 85262336, "step": 327 }, { "epoch": 2.1588089330024816, "grad_norm": 1.050296411553259, "learning_rate": 3.0112967005237443e-05, "loss": 0.3066, "num_input_tokens_seen": 85524480, "step": 328 }, { "epoch": 2.1654259718775846, "grad_norm": 1.0882638825076911, "learning_rate": 3.0011095626549977e-05, "loss": 0.3193, "num_input_tokens_seen": 85786624, "step": 329 }, { "epoch": 2.172043010752688, "grad_norm": 1.1623059515415315, "learning_rate": 2.990913748419411e-05, "loss": 0.3526, "num_input_tokens_seen": 86048768, "step": 330 }, { "epoch": 2.1786600496277915, "grad_norm": 1.1908398814663688, "learning_rate": 2.9807094343504804e-05, "loss": 0.3518, "num_input_tokens_seen": 86310912, "step": 331 }, { "epoch": 2.185277088502895, "grad_norm": 1.1808598497115437, "learning_rate": 2.9704967971288755e-05, "loss": 0.3385, "num_input_tokens_seen": 86573056, "step": 332 }, { "epoch": 2.1918941273779984, "grad_norm": 1.1384860728455102, "learning_rate": 2.9602760135793735e-05, "loss": 0.3228, "num_input_tokens_seen": 86835200, "step": 333 }, { "epoch": 2.198511166253102, "grad_norm": 1.0916481479985516, "learning_rate": 2.9500472606677994e-05, "loss": 0.3118, "num_input_tokens_seen": 87097344, "step": 334 }, { "epoch": 2.2051282051282053, "grad_norm": 0.9878895674767743, "learning_rate": 2.9398107154979638e-05, "loss": 0.3293, "num_input_tokens_seen": 87359488, "step": 335 }, { "epoch": 2.2117452440033087, "grad_norm": 1.1884030862907187, "learning_rate": 2.9295665553085937e-05, "loss": 0.3331, "num_input_tokens_seen": 87621632, "step": 336 }, { "epoch": 2.2183622828784118, "grad_norm": 1.1094941724975913, "learning_rate": 2.919314957470265e-05, "loss": 0.322, "num_input_tokens_seen": 87883776, "step": 337 }, { "epoch": 2.224979321753515, "grad_norm": 0.9881499762640955, "learning_rate": 2.909056099482332e-05, "loss": 0.3239, "num_input_tokens_seen": 88145920, "step": 338 }, { "epoch": 2.2315963606286187, "grad_norm": 1.0840977567322172, "learning_rate": 2.8987901589698517e-05, "loss": 0.3163, "num_input_tokens_seen": 88408064, "step": 339 }, { "epoch": 2.238213399503722, "grad_norm": 1.0548483105248472, "learning_rate": 2.8885173136805127e-05, "loss": 0.3303, "num_input_tokens_seen": 88670208, "step": 340 }, { "epoch": 2.2448304383788256, "grad_norm": 1.1561588854960063, "learning_rate": 2.8782377414815532e-05, "loss": 0.3327, "num_input_tokens_seen": 88932352, "step": 341 }, { "epoch": 2.251447477253929, "grad_norm": 1.1826177082260094, "learning_rate": 2.867951620356684e-05, "loss": 0.3161, "num_input_tokens_seen": 89194496, "step": 342 }, { "epoch": 2.258064516129032, "grad_norm": 1.0982559371878697, "learning_rate": 2.8576591284030058e-05, "loss": 0.3238, "num_input_tokens_seen": 89456640, "step": 343 }, { "epoch": 2.2646815550041355, "grad_norm": 1.1788800916865765, "learning_rate": 2.847360443827926e-05, "loss": 0.3423, "num_input_tokens_seen": 89718784, "step": 344 }, { "epoch": 2.271298593879239, "grad_norm": 1.1681881241245002, "learning_rate": 2.837055744946072e-05, "loss": 0.3479, "num_input_tokens_seen": 89980928, "step": 345 }, { "epoch": 2.2779156327543424, "grad_norm": 1.150709547987284, "learning_rate": 2.826745210176207e-05, "loss": 0.3252, "num_input_tokens_seen": 90243072, "step": 346 }, { "epoch": 2.284532671629446, "grad_norm": 1.0763555223019872, "learning_rate": 2.8164290180381365e-05, "loss": 0.3398, "num_input_tokens_seen": 90505216, "step": 347 }, { "epoch": 2.2911497105045493, "grad_norm": 1.1937992265302106, "learning_rate": 2.8061073471496195e-05, "loss": 0.3397, "num_input_tokens_seen": 90767360, "step": 348 }, { "epoch": 2.2977667493796528, "grad_norm": 0.9959579454351639, "learning_rate": 2.795780376223277e-05, "loss": 0.3506, "num_input_tokens_seen": 91029504, "step": 349 }, { "epoch": 2.304383788254756, "grad_norm": 1.0702429896944252, "learning_rate": 2.7854482840634965e-05, "loss": 0.3208, "num_input_tokens_seen": 91291648, "step": 350 }, { "epoch": 2.311000827129859, "grad_norm": 1.1450863009989223, "learning_rate": 2.7751112495633345e-05, "loss": 0.3313, "num_input_tokens_seen": 91553792, "step": 351 }, { "epoch": 2.3176178660049627, "grad_norm": 1.0477187731328461, "learning_rate": 2.764769451701421e-05, "loss": 0.3298, "num_input_tokens_seen": 91815936, "step": 352 }, { "epoch": 2.324234904880066, "grad_norm": 1.1549329031215105, "learning_rate": 2.7544230695388634e-05, "loss": 0.341, "num_input_tokens_seen": 92078080, "step": 353 }, { "epoch": 2.3308519437551696, "grad_norm": 1.0366387991827954, "learning_rate": 2.744072282216139e-05, "loss": 0.3425, "num_input_tokens_seen": 92340224, "step": 354 }, { "epoch": 2.337468982630273, "grad_norm": 1.1817424289487461, "learning_rate": 2.73371726895e-05, "loss": 0.3409, "num_input_tokens_seen": 92602368, "step": 355 }, { "epoch": 2.3440860215053765, "grad_norm": 1.0301874196493457, "learning_rate": 2.7233582090303674e-05, "loss": 0.3229, "num_input_tokens_seen": 92864512, "step": 356 }, { "epoch": 2.3507030603804795, "grad_norm": 1.0646554418765033, "learning_rate": 2.7129952818172256e-05, "loss": 0.3344, "num_input_tokens_seen": 93126656, "step": 357 }, { "epoch": 2.357320099255583, "grad_norm": 1.084404941345572, "learning_rate": 2.702628666737521e-05, "loss": 0.3254, "num_input_tokens_seen": 93388800, "step": 358 }, { "epoch": 2.3639371381306864, "grad_norm": 1.065831371281375, "learning_rate": 2.692258543282052e-05, "loss": 0.3211, "num_input_tokens_seen": 93650944, "step": 359 }, { "epoch": 2.37055417700579, "grad_norm": 1.153320273204117, "learning_rate": 2.68188509100236e-05, "loss": 0.3374, "num_input_tokens_seen": 93913088, "step": 360 }, { "epoch": 2.3771712158808933, "grad_norm": 1.170047331616093, "learning_rate": 2.671508489507626e-05, "loss": 0.3478, "num_input_tokens_seen": 94175232, "step": 361 }, { "epoch": 2.3837882547559968, "grad_norm": 1.1329869776410735, "learning_rate": 2.6611289184615558e-05, "loss": 0.3347, "num_input_tokens_seen": 94437376, "step": 362 }, { "epoch": 2.3904052936311, "grad_norm": 1.0908955007085368, "learning_rate": 2.6507465575792707e-05, "loss": 0.3341, "num_input_tokens_seen": 94699520, "step": 363 }, { "epoch": 2.3970223325062037, "grad_norm": 1.0929346197114131, "learning_rate": 2.6403615866241964e-05, "loss": 0.3492, "num_input_tokens_seen": 94961664, "step": 364 }, { "epoch": 2.4036393713813067, "grad_norm": 1.142990047990031, "learning_rate": 2.6299741854049508e-05, "loss": 0.334, "num_input_tokens_seen": 95223808, "step": 365 }, { "epoch": 2.41025641025641, "grad_norm": 1.1552852832233795, "learning_rate": 2.6195845337722303e-05, "loss": 0.3299, "num_input_tokens_seen": 95485952, "step": 366 }, { "epoch": 2.4168734491315136, "grad_norm": 1.111643861242986, "learning_rate": 2.6091928116156938e-05, "loss": 0.3278, "num_input_tokens_seen": 95748096, "step": 367 }, { "epoch": 2.423490488006617, "grad_norm": 1.1676117715035266, "learning_rate": 2.598799198860853e-05, "loss": 0.3355, "num_input_tokens_seen": 96010240, "step": 368 }, { "epoch": 2.4301075268817205, "grad_norm": 1.1044739988434005, "learning_rate": 2.5884038754659497e-05, "loss": 0.3475, "num_input_tokens_seen": 96272384, "step": 369 }, { "epoch": 2.436724565756824, "grad_norm": 1.0768350723016058, "learning_rate": 2.5780070214188478e-05, "loss": 0.3257, "num_input_tokens_seen": 96534528, "step": 370 }, { "epoch": 2.4433416046319274, "grad_norm": 1.0381793723404524, "learning_rate": 2.5676088167339128e-05, "loss": 0.3318, "num_input_tokens_seen": 96796672, "step": 371 }, { "epoch": 2.4499586435070304, "grad_norm": 1.2240143879709446, "learning_rate": 2.5572094414488944e-05, "loss": 0.3424, "num_input_tokens_seen": 97058816, "step": 372 }, { "epoch": 2.456575682382134, "grad_norm": 1.0385118889958669, "learning_rate": 2.5468090756218117e-05, "loss": 0.3353, "num_input_tokens_seen": 97320960, "step": 373 }, { "epoch": 2.4631927212572373, "grad_norm": 1.1995150658571436, "learning_rate": 2.5364078993278335e-05, "loss": 0.3345, "num_input_tokens_seen": 97583104, "step": 374 }, { "epoch": 2.4698097601323408, "grad_norm": 1.1417927976011224, "learning_rate": 2.526006092656161e-05, "loss": 0.3467, "num_input_tokens_seen": 97845248, "step": 375 }, { "epoch": 2.4764267990074442, "grad_norm": 1.11091502135598, "learning_rate": 2.5156038357069106e-05, "loss": 0.3378, "num_input_tokens_seen": 98107392, "step": 376 }, { "epoch": 2.4830438378825477, "grad_norm": 1.1198492983330697, "learning_rate": 2.5052013085879955e-05, "loss": 0.3411, "num_input_tokens_seen": 98369536, "step": 377 }, { "epoch": 2.489660876757651, "grad_norm": 1.0507799282706884, "learning_rate": 2.494798691412005e-05, "loss": 0.3356, "num_input_tokens_seen": 98631680, "step": 378 }, { "epoch": 2.4962779156327546, "grad_norm": 1.104596832922824, "learning_rate": 2.4843961642930904e-05, "loss": 0.3303, "num_input_tokens_seen": 98893824, "step": 379 }, { "epoch": 2.5028949545078576, "grad_norm": 1.0982025410563003, "learning_rate": 2.4739939073438397e-05, "loss": 0.3212, "num_input_tokens_seen": 99155968, "step": 380 }, { "epoch": 2.509511993382961, "grad_norm": 1.0765071778760118, "learning_rate": 2.463592100672168e-05, "loss": 0.3335, "num_input_tokens_seen": 99418112, "step": 381 }, { "epoch": 2.5161290322580645, "grad_norm": 1.0693943579256004, "learning_rate": 2.4531909243781885e-05, "loss": 0.3263, "num_input_tokens_seen": 99680256, "step": 382 }, { "epoch": 2.522746071133168, "grad_norm": 1.078185931899051, "learning_rate": 2.442790558551106e-05, "loss": 0.3363, "num_input_tokens_seen": 99942400, "step": 383 }, { "epoch": 2.5293631100082714, "grad_norm": 1.0519087719201188, "learning_rate": 2.4323911832660874e-05, "loss": 0.339, "num_input_tokens_seen": 100204544, "step": 384 }, { "epoch": 2.5359801488833744, "grad_norm": 1.0506382318529084, "learning_rate": 2.4219929785811518e-05, "loss": 0.3359, "num_input_tokens_seen": 100466688, "step": 385 }, { "epoch": 2.542597187758478, "grad_norm": 1.0113516063929346, "learning_rate": 2.4115961245340505e-05, "loss": 0.3379, "num_input_tokens_seen": 100728832, "step": 386 }, { "epoch": 2.5492142266335813, "grad_norm": 1.0637623036996438, "learning_rate": 2.401200801139148e-05, "loss": 0.3507, "num_input_tokens_seen": 100990976, "step": 387 }, { "epoch": 2.555831265508685, "grad_norm": 1.1501956408026655, "learning_rate": 2.3908071883843068e-05, "loss": 0.3411, "num_input_tokens_seen": 101253120, "step": 388 }, { "epoch": 2.5624483043837882, "grad_norm": 1.1150126958290971, "learning_rate": 2.38041546622777e-05, "loss": 0.3366, "num_input_tokens_seen": 101515264, "step": 389 }, { "epoch": 2.5690653432588917, "grad_norm": 1.073213710305425, "learning_rate": 2.3700258145950495e-05, "loss": 0.3656, "num_input_tokens_seen": 101777408, "step": 390 }, { "epoch": 2.575682382133995, "grad_norm": 1.078732276630299, "learning_rate": 2.359638413375804e-05, "loss": 0.3423, "num_input_tokens_seen": 102039552, "step": 391 }, { "epoch": 2.5822994210090986, "grad_norm": 1.0226452813564892, "learning_rate": 2.349253442420731e-05, "loss": 0.3217, "num_input_tokens_seen": 102301696, "step": 392 }, { "epoch": 2.588916459884202, "grad_norm": 1.0479991409872327, "learning_rate": 2.338871081538445e-05, "loss": 0.3174, "num_input_tokens_seen": 102563840, "step": 393 }, { "epoch": 2.595533498759305, "grad_norm": 1.0103696849600996, "learning_rate": 2.3284915104923752e-05, "loss": 0.3346, "num_input_tokens_seen": 102825984, "step": 394 }, { "epoch": 2.6021505376344085, "grad_norm": 1.161868760445816, "learning_rate": 2.3181149089976405e-05, "loss": 0.3444, "num_input_tokens_seen": 103088128, "step": 395 }, { "epoch": 2.608767576509512, "grad_norm": 1.0926396656803976, "learning_rate": 2.3077414567179494e-05, "loss": 0.3348, "num_input_tokens_seen": 103350272, "step": 396 }, { "epoch": 2.6153846153846154, "grad_norm": 1.1576399191571771, "learning_rate": 2.2973713332624796e-05, "loss": 0.3288, "num_input_tokens_seen": 103612416, "step": 397 }, { "epoch": 2.622001654259719, "grad_norm": 0.9917185104236572, "learning_rate": 2.2870047181827743e-05, "loss": 0.3328, "num_input_tokens_seen": 103874560, "step": 398 }, { "epoch": 2.6286186931348223, "grad_norm": 1.037992774395136, "learning_rate": 2.2766417909696332e-05, "loss": 0.3554, "num_input_tokens_seen": 104136704, "step": 399 }, { "epoch": 2.6352357320099253, "grad_norm": 1.0486140714606227, "learning_rate": 2.2662827310499995e-05, "loss": 0.3296, "num_input_tokens_seen": 104398848, "step": 400 }, { "epoch": 2.641852770885029, "grad_norm": 1.0447686801537046, "learning_rate": 2.2559277177838612e-05, "loss": 0.3262, "num_input_tokens_seen": 104660992, "step": 401 }, { "epoch": 2.6484698097601322, "grad_norm": 1.0165801431956982, "learning_rate": 2.2455769304611365e-05, "loss": 0.3453, "num_input_tokens_seen": 104923136, "step": 402 }, { "epoch": 2.6550868486352357, "grad_norm": 0.9788796226467195, "learning_rate": 2.2352305482985793e-05, "loss": 0.3343, "num_input_tokens_seen": 105185280, "step": 403 }, { "epoch": 2.661703887510339, "grad_norm": 1.1239336553275057, "learning_rate": 2.224888750436666e-05, "loss": 0.3231, "num_input_tokens_seen": 105447424, "step": 404 }, { "epoch": 2.6683209263854426, "grad_norm": 1.0832025698438201, "learning_rate": 2.2145517159365044e-05, "loss": 0.3412, "num_input_tokens_seen": 105709568, "step": 405 }, { "epoch": 2.674937965260546, "grad_norm": 1.1361074407484597, "learning_rate": 2.2042196237767233e-05, "loss": 0.3411, "num_input_tokens_seen": 105971712, "step": 406 }, { "epoch": 2.6815550041356495, "grad_norm": 1.1514517127780615, "learning_rate": 2.1938926528503807e-05, "loss": 0.3538, "num_input_tokens_seen": 106233856, "step": 407 }, { "epoch": 2.688172043010753, "grad_norm": 1.1204288030113607, "learning_rate": 2.183570981961864e-05, "loss": 0.3286, "num_input_tokens_seen": 106496000, "step": 408 }, { "epoch": 2.694789081885856, "grad_norm": 1.073945290517408, "learning_rate": 2.173254789823794e-05, "loss": 0.3388, "num_input_tokens_seen": 106758144, "step": 409 }, { "epoch": 2.7014061207609594, "grad_norm": 1.1488472446305815, "learning_rate": 2.162944255053928e-05, "loss": 0.3408, "num_input_tokens_seen": 107020288, "step": 410 }, { "epoch": 2.708023159636063, "grad_norm": 1.0097780396990714, "learning_rate": 2.1526395561720742e-05, "loss": 0.3385, "num_input_tokens_seen": 107282432, "step": 411 }, { "epoch": 2.7146401985111663, "grad_norm": 1.159742866337566, "learning_rate": 2.1423408715969948e-05, "loss": 0.353, "num_input_tokens_seen": 107544576, "step": 412 }, { "epoch": 2.72125723738627, "grad_norm": 1.0399730947948516, "learning_rate": 2.132048379643316e-05, "loss": 0.3249, "num_input_tokens_seen": 107806720, "step": 413 }, { "epoch": 2.727874276261373, "grad_norm": 1.0290808670104754, "learning_rate": 2.1217622585184474e-05, "loss": 0.3175, "num_input_tokens_seen": 108068864, "step": 414 }, { "epoch": 2.7344913151364763, "grad_norm": 1.0570208529977188, "learning_rate": 2.1114826863194882e-05, "loss": 0.3425, "num_input_tokens_seen": 108331008, "step": 415 }, { "epoch": 2.7411083540115797, "grad_norm": 1.071729013559576, "learning_rate": 2.101209841030149e-05, "loss": 0.3336, "num_input_tokens_seen": 108593152, "step": 416 }, { "epoch": 2.747725392886683, "grad_norm": 1.114457915850528, "learning_rate": 2.090943900517669e-05, "loss": 0.3207, "num_input_tokens_seen": 108855296, "step": 417 }, { "epoch": 2.7543424317617866, "grad_norm": 1.077218628216003, "learning_rate": 2.0806850425297363e-05, "loss": 0.3393, "num_input_tokens_seen": 109117440, "step": 418 }, { "epoch": 2.76095947063689, "grad_norm": 1.1199612602141782, "learning_rate": 2.070433444691407e-05, "loss": 0.3207, "num_input_tokens_seen": 109379584, "step": 419 }, { "epoch": 2.7675765095119935, "grad_norm": 1.1174362860816687, "learning_rate": 2.060189284502037e-05, "loss": 0.3374, "num_input_tokens_seen": 109641728, "step": 420 }, { "epoch": 2.774193548387097, "grad_norm": 1.1287911544921245, "learning_rate": 2.049952739332201e-05, "loss": 0.3191, "num_input_tokens_seen": 109903872, "step": 421 }, { "epoch": 2.7808105872622004, "grad_norm": 1.0450302695675435, "learning_rate": 2.039723986420628e-05, "loss": 0.3559, "num_input_tokens_seen": 110166016, "step": 422 }, { "epoch": 2.7874276261373034, "grad_norm": 1.09368946884029, "learning_rate": 2.029503202871125e-05, "loss": 0.3413, "num_input_tokens_seen": 110428160, "step": 423 }, { "epoch": 2.794044665012407, "grad_norm": 1.022545929344771, "learning_rate": 2.01929056564952e-05, "loss": 0.3448, "num_input_tokens_seen": 110690304, "step": 424 }, { "epoch": 2.8006617038875103, "grad_norm": 1.030074482926079, "learning_rate": 2.0090862515805898e-05, "loss": 0.3623, "num_input_tokens_seen": 110952448, "step": 425 }, { "epoch": 2.807278742762614, "grad_norm": 1.109666478372538, "learning_rate": 1.9988904373450018e-05, "loss": 0.3223, "num_input_tokens_seen": 111214592, "step": 426 }, { "epoch": 2.8138957816377173, "grad_norm": 1.0575561897801167, "learning_rate": 1.9887032994762563e-05, "loss": 0.347, "num_input_tokens_seen": 111476736, "step": 427 }, { "epoch": 2.8205128205128203, "grad_norm": 1.0648123386835917, "learning_rate": 1.9785250143576254e-05, "loss": 0.3375, "num_input_tokens_seen": 111738880, "step": 428 }, { "epoch": 2.8271298593879237, "grad_norm": 1.1050336868555022, "learning_rate": 1.9683557582191055e-05, "loss": 0.3239, "num_input_tokens_seen": 112001024, "step": 429 }, { "epoch": 2.833746898263027, "grad_norm": 1.0359983663485797, "learning_rate": 1.9581957071343592e-05, "loss": 0.3335, "num_input_tokens_seen": 112263168, "step": 430 }, { "epoch": 2.8403639371381306, "grad_norm": 1.0414734239060488, "learning_rate": 1.9480450370176726e-05, "loss": 0.333, "num_input_tokens_seen": 112525312, "step": 431 }, { "epoch": 2.846980976013234, "grad_norm": 1.0382890141371446, "learning_rate": 1.9379039236209037e-05, "loss": 0.3319, "num_input_tokens_seen": 112787456, "step": 432 }, { "epoch": 2.8535980148883375, "grad_norm": 1.0013368189447842, "learning_rate": 1.9277725425304467e-05, "loss": 0.331, "num_input_tokens_seen": 113049600, "step": 433 }, { "epoch": 2.860215053763441, "grad_norm": 1.0179396934960452, "learning_rate": 1.917651069164183e-05, "loss": 0.3245, "num_input_tokens_seen": 113311744, "step": 434 }, { "epoch": 2.8668320926385444, "grad_norm": 0.9928696786268584, "learning_rate": 1.9075396787684533e-05, "loss": 0.3347, "num_input_tokens_seen": 113573888, "step": 435 }, { "epoch": 2.873449131513648, "grad_norm": 1.046777848590959, "learning_rate": 1.8974385464150136e-05, "loss": 0.3238, "num_input_tokens_seen": 113836032, "step": 436 }, { "epoch": 2.880066170388751, "grad_norm": 1.0405751720587308, "learning_rate": 1.8873478469980125e-05, "loss": 0.3457, "num_input_tokens_seen": 114098176, "step": 437 }, { "epoch": 2.8866832092638544, "grad_norm": 1.122608330877757, "learning_rate": 1.8772677552309596e-05, "loss": 0.3412, "num_input_tokens_seen": 114360320, "step": 438 }, { "epoch": 2.893300248138958, "grad_norm": 1.0442286190838361, "learning_rate": 1.8671984456436968e-05, "loss": 0.3496, "num_input_tokens_seen": 114622464, "step": 439 }, { "epoch": 2.8999172870140613, "grad_norm": 1.0227002785782815, "learning_rate": 1.8571400925793855e-05, "loss": 0.3397, "num_input_tokens_seen": 114884608, "step": 440 }, { "epoch": 2.9065343258891647, "grad_norm": 1.0156032484754163, "learning_rate": 1.8470928701914763e-05, "loss": 0.3191, "num_input_tokens_seen": 115146752, "step": 441 }, { "epoch": 2.9131513647642677, "grad_norm": 1.1068895653021453, "learning_rate": 1.8370569524407054e-05, "loss": 0.323, "num_input_tokens_seen": 115408896, "step": 442 }, { "epoch": 2.919768403639371, "grad_norm": 0.9691703910653604, "learning_rate": 1.8270325130920728e-05, "loss": 0.3116, "num_input_tokens_seen": 115671040, "step": 443 }, { "epoch": 2.9263854425144746, "grad_norm": 0.9826661649635426, "learning_rate": 1.817019725711841e-05, "loss": 0.3324, "num_input_tokens_seen": 115933184, "step": 444 }, { "epoch": 2.933002481389578, "grad_norm": 1.0159017429748878, "learning_rate": 1.807018763664524e-05, "loss": 0.328, "num_input_tokens_seen": 116195328, "step": 445 }, { "epoch": 2.9396195202646815, "grad_norm": 1.0737382354639178, "learning_rate": 1.7970298001098897e-05, "loss": 0.3114, "num_input_tokens_seen": 116457472, "step": 446 }, { "epoch": 2.946236559139785, "grad_norm": 1.0066108381286578, "learning_rate": 1.787053007999961e-05, "loss": 0.3332, "num_input_tokens_seen": 116719616, "step": 447 }, { "epoch": 2.9528535980148884, "grad_norm": 1.0680776988030307, "learning_rate": 1.7770885600760183e-05, "loss": 0.3467, "num_input_tokens_seen": 116981760, "step": 448 }, { "epoch": 2.959470636889992, "grad_norm": 1.0349807857034838, "learning_rate": 1.7671366288656152e-05, "loss": 0.3241, "num_input_tokens_seen": 117243904, "step": 449 }, { "epoch": 2.9660876757650954, "grad_norm": 1.0388721618627996, "learning_rate": 1.7571973866795815e-05, "loss": 0.3292, "num_input_tokens_seen": 117506048, "step": 450 }, { "epoch": 2.9727047146401984, "grad_norm": 1.0181955179089524, "learning_rate": 1.7472710056090502e-05, "loss": 0.3499, "num_input_tokens_seen": 117768192, "step": 451 }, { "epoch": 2.979321753515302, "grad_norm": 1.0554960164473584, "learning_rate": 1.7373576575224686e-05, "loss": 0.3331, "num_input_tokens_seen": 118030336, "step": 452 }, { "epoch": 2.9859387923904053, "grad_norm": 1.0744246028347093, "learning_rate": 1.7274575140626318e-05, "loss": 0.3239, "num_input_tokens_seen": 118292480, "step": 453 }, { "epoch": 2.9925558312655087, "grad_norm": 1.0034573899878152, "learning_rate": 1.7175707466437007e-05, "loss": 0.3298, "num_input_tokens_seen": 118554624, "step": 454 }, { "epoch": 2.999172870140612, "grad_norm": 1.0586016785419656, "learning_rate": 1.7076975264482434e-05, "loss": 0.3279, "num_input_tokens_seen": 118816768, "step": 455 }, { "epoch": 3.0, "grad_norm": 1.0586016785419656, "learning_rate": 1.6978380244242647e-05, "loss": 0.3046, "num_input_tokens_seen": 118849536, "step": 456 }, { "epoch": 3.0066170388751035, "grad_norm": 2.6288675198935905, "learning_rate": 1.6879924112822505e-05, "loss": 0.1025, "num_input_tokens_seen": 119111680, "step": 457 }, { "epoch": 3.013234077750207, "grad_norm": 1.4229544470512288, "learning_rate": 1.678160857492207e-05, "loss": 0.1027, "num_input_tokens_seen": 119373824, "step": 458 }, { "epoch": 3.0198511166253104, "grad_norm": 1.2481296003332036, "learning_rate": 1.6683435332807172e-05, "loss": 0.098, "num_input_tokens_seen": 119635968, "step": 459 }, { "epoch": 3.0264681555004134, "grad_norm": 0.8805105822487416, "learning_rate": 1.658540608627985e-05, "loss": 0.0876, "num_input_tokens_seen": 119898112, "step": 460 }, { "epoch": 3.033085194375517, "grad_norm": 0.828690671159354, "learning_rate": 1.6487522532648962e-05, "loss": 0.0931, "num_input_tokens_seen": 120160256, "step": 461 }, { "epoch": 3.0397022332506203, "grad_norm": 1.0491253577091149, "learning_rate": 1.6389786366700836e-05, "loss": 0.0841, "num_input_tokens_seen": 120422400, "step": 462 }, { "epoch": 3.0463192721257237, "grad_norm": 1.006633637656072, "learning_rate": 1.629219928066982e-05, "loss": 0.0856, "num_input_tokens_seen": 120684544, "step": 463 }, { "epoch": 3.052936311000827, "grad_norm": 1.3229784618102733, "learning_rate": 1.619476296420909e-05, "loss": 0.0999, "num_input_tokens_seen": 120946688, "step": 464 }, { "epoch": 3.0595533498759306, "grad_norm": 1.3217791698144548, "learning_rate": 1.6097479104361326e-05, "loss": 0.0913, "num_input_tokens_seen": 121208832, "step": 465 }, { "epoch": 3.066170388751034, "grad_norm": 1.0018438706002273, "learning_rate": 1.6000349385529557e-05, "loss": 0.0839, "num_input_tokens_seen": 121470976, "step": 466 }, { "epoch": 3.072787427626137, "grad_norm": 0.8813830494513232, "learning_rate": 1.5903375489447925e-05, "loss": 0.0928, "num_input_tokens_seen": 121733120, "step": 467 }, { "epoch": 3.0794044665012406, "grad_norm": 0.7791960697591414, "learning_rate": 1.5806559095152652e-05, "loss": 0.0805, "num_input_tokens_seen": 121995264, "step": 468 }, { "epoch": 3.086021505376344, "grad_norm": 0.7326718702522317, "learning_rate": 1.570990187895289e-05, "loss": 0.0823, "num_input_tokens_seen": 122257408, "step": 469 }, { "epoch": 3.0926385442514475, "grad_norm": 0.8396403661003932, "learning_rate": 1.561340551440176e-05, "loss": 0.0862, "num_input_tokens_seen": 122519552, "step": 470 }, { "epoch": 3.099255583126551, "grad_norm": 0.813187861447653, "learning_rate": 1.5517071672267314e-05, "loss": 0.0881, "num_input_tokens_seen": 122781696, "step": 471 }, { "epoch": 3.1058726220016544, "grad_norm": 0.7573337922842879, "learning_rate": 1.542090202050368e-05, "loss": 0.0835, "num_input_tokens_seen": 123043840, "step": 472 }, { "epoch": 3.112489660876758, "grad_norm": 0.6833827588765713, "learning_rate": 1.532489822422211e-05, "loss": 0.084, "num_input_tokens_seen": 123305984, "step": 473 }, { "epoch": 3.119106699751861, "grad_norm": 0.7934273146364325, "learning_rate": 1.5229061945662198e-05, "loss": 0.079, "num_input_tokens_seen": 123568128, "step": 474 }, { "epoch": 3.1257237386269643, "grad_norm": 0.7555918786299218, "learning_rate": 1.5133394844163093e-05, "loss": 0.0767, "num_input_tokens_seen": 123830272, "step": 475 }, { "epoch": 3.1323407775020677, "grad_norm": 0.7883685737277187, "learning_rate": 1.503789857613473e-05, "loss": 0.0854, "num_input_tokens_seen": 124092416, "step": 476 }, { "epoch": 3.138957816377171, "grad_norm": 0.7705225890710093, "learning_rate": 1.4942574795029213e-05, "loss": 0.0748, "num_input_tokens_seen": 124354560, "step": 477 }, { "epoch": 3.1455748552522746, "grad_norm": 0.7409548352905246, "learning_rate": 1.4847425151312127e-05, "loss": 0.0728, "num_input_tokens_seen": 124616704, "step": 478 }, { "epoch": 3.152191894127378, "grad_norm": 0.732934471932393, "learning_rate": 1.4752451292434016e-05, "loss": 0.0782, "num_input_tokens_seen": 124878848, "step": 479 }, { "epoch": 3.1588089330024816, "grad_norm": 0.7587629849721698, "learning_rate": 1.4657654862801798e-05, "loss": 0.0758, "num_input_tokens_seen": 125140992, "step": 480 }, { "epoch": 3.1654259718775846, "grad_norm": 0.7303711757854935, "learning_rate": 1.4563037503750366e-05, "loss": 0.0757, "num_input_tokens_seen": 125403136, "step": 481 }, { "epoch": 3.172043010752688, "grad_norm": 0.7477005165930065, "learning_rate": 1.446860085351407e-05, "loss": 0.0798, "num_input_tokens_seen": 125665280, "step": 482 }, { "epoch": 3.1786600496277915, "grad_norm": 0.6628074364177822, "learning_rate": 1.4374346547198486e-05, "loss": 0.0741, "num_input_tokens_seen": 125927424, "step": 483 }, { "epoch": 3.185277088502895, "grad_norm": 0.6904627228112755, "learning_rate": 1.4280276216751956e-05, "loss": 0.0767, "num_input_tokens_seen": 126189568, "step": 484 }, { "epoch": 3.1918941273779984, "grad_norm": 0.7078389751667725, "learning_rate": 1.4186391490937481e-05, "loss": 0.0738, "num_input_tokens_seen": 126451712, "step": 485 }, { "epoch": 3.198511166253102, "grad_norm": 0.688986245366018, "learning_rate": 1.4092693995304368e-05, "loss": 0.0765, "num_input_tokens_seen": 126713856, "step": 486 }, { "epoch": 3.2051282051282053, "grad_norm": 0.7409087917501026, "learning_rate": 1.3999185352160231e-05, "loss": 0.0768, "num_input_tokens_seen": 126976000, "step": 487 }, { "epoch": 3.2117452440033087, "grad_norm": 0.712307910459613, "learning_rate": 1.3905867180542803e-05, "loss": 0.0795, "num_input_tokens_seen": 127238144, "step": 488 }, { "epoch": 3.2183622828784118, "grad_norm": 0.700878195423877, "learning_rate": 1.3812741096191905e-05, "loss": 0.0822, "num_input_tokens_seen": 127500288, "step": 489 }, { "epoch": 3.224979321753515, "grad_norm": 0.7268693202788183, "learning_rate": 1.3719808711521573e-05, "loss": 0.0732, "num_input_tokens_seen": 127762432, "step": 490 }, { "epoch": 3.2315963606286187, "grad_norm": 0.7051720495475905, "learning_rate": 1.3627071635591976e-05, "loss": 0.0693, "num_input_tokens_seen": 128024576, "step": 491 }, { "epoch": 3.238213399503722, "grad_norm": 0.7075601778394868, "learning_rate": 1.353453147408174e-05, "loss": 0.0752, "num_input_tokens_seen": 128286720, "step": 492 }, { "epoch": 3.2448304383788256, "grad_norm": 0.6983282572573932, "learning_rate": 1.344218982925996e-05, "loss": 0.0743, "num_input_tokens_seen": 128548864, "step": 493 }, { "epoch": 3.251447477253929, "grad_norm": 0.6787490400858531, "learning_rate": 1.3350048299958645e-05, "loss": 0.0723, "num_input_tokens_seen": 128811008, "step": 494 }, { "epoch": 3.258064516129032, "grad_norm": 0.6780173788911718, "learning_rate": 1.3258108481544849e-05, "loss": 0.07, "num_input_tokens_seen": 129073152, "step": 495 }, { "epoch": 3.2646815550041355, "grad_norm": 0.6978698177594762, "learning_rate": 1.3166371965893226e-05, "loss": 0.0746, "num_input_tokens_seen": 129335296, "step": 496 }, { "epoch": 3.271298593879239, "grad_norm": 0.693822546631228, "learning_rate": 1.3074840341358318e-05, "loss": 0.0754, "num_input_tokens_seen": 129597440, "step": 497 }, { "epoch": 3.2779156327543424, "grad_norm": 0.6691956665865821, "learning_rate": 1.2983515192747153e-05, "loss": 0.0743, "num_input_tokens_seen": 129859584, "step": 498 }, { "epoch": 3.284532671629446, "grad_norm": 0.6867164097827277, "learning_rate": 1.2892398101291759e-05, "loss": 0.0679, "num_input_tokens_seen": 130121728, "step": 499 }, { "epoch": 3.2911497105045493, "grad_norm": 0.6431526093398369, "learning_rate": 1.2801490644621789e-05, "loss": 0.0668, "num_input_tokens_seen": 130383872, "step": 500 }, { "epoch": 3.2977667493796528, "grad_norm": 0.6685555901467412, "learning_rate": 1.271079439673726e-05, "loss": 0.0718, "num_input_tokens_seen": 130646016, "step": 501 }, { "epoch": 3.304383788254756, "grad_norm": 0.682140794722891, "learning_rate": 1.2620310927981176e-05, "loss": 0.0721, "num_input_tokens_seen": 130908160, "step": 502 }, { "epoch": 3.311000827129859, "grad_norm": 0.6387878739981501, "learning_rate": 1.2530041805012504e-05, "loss": 0.068, "num_input_tokens_seen": 131170304, "step": 503 }, { "epoch": 3.3176178660049627, "grad_norm": 0.6757177577212953, "learning_rate": 1.2439988590778872e-05, "loss": 0.0763, "num_input_tokens_seen": 131432448, "step": 504 }, { "epoch": 3.324234904880066, "grad_norm": 0.6907225144982705, "learning_rate": 1.235015284448969e-05, "loss": 0.0743, "num_input_tokens_seen": 131694592, "step": 505 }, { "epoch": 3.3308519437551696, "grad_norm": 0.6852742550824547, "learning_rate": 1.2260536121588978e-05, "loss": 0.071, "num_input_tokens_seen": 131956736, "step": 506 }, { "epoch": 3.337468982630273, "grad_norm": 0.6514568739578507, "learning_rate": 1.2171139973728566e-05, "loss": 0.0703, "num_input_tokens_seen": 132218880, "step": 507 }, { "epoch": 3.3440860215053765, "grad_norm": 0.676377694947883, "learning_rate": 1.2081965948741162e-05, "loss": 0.0701, "num_input_tokens_seen": 132481024, "step": 508 }, { "epoch": 3.3507030603804795, "grad_norm": 0.6239946346262029, "learning_rate": 1.1993015590613573e-05, "loss": 0.0745, "num_input_tokens_seen": 132743168, "step": 509 }, { "epoch": 3.357320099255583, "grad_norm": 0.6668058973586733, "learning_rate": 1.1904290439459973e-05, "loss": 0.0684, "num_input_tokens_seen": 133005312, "step": 510 }, { "epoch": 3.3639371381306864, "grad_norm": 0.6386467362434296, "learning_rate": 1.1815792031495224e-05, "loss": 0.0689, "num_input_tokens_seen": 133267456, "step": 511 }, { "epoch": 3.37055417700579, "grad_norm": 0.6136816329646979, "learning_rate": 1.172752189900829e-05, "loss": 0.0696, "num_input_tokens_seen": 133529600, "step": 512 }, { "epoch": 3.3771712158808933, "grad_norm": 0.6315422013733348, "learning_rate": 1.1639481570335692e-05, "loss": 0.0689, "num_input_tokens_seen": 133791744, "step": 513 }, { "epoch": 3.3837882547559968, "grad_norm": 0.6957057963725773, "learning_rate": 1.1551672569835095e-05, "loss": 0.0715, "num_input_tokens_seen": 134053888, "step": 514 }, { "epoch": 3.3904052936311, "grad_norm": 0.7124213473576223, "learning_rate": 1.1464096417858822e-05, "loss": 0.0782, "num_input_tokens_seen": 134316032, "step": 515 }, { "epoch": 3.3970223325062037, "grad_norm": 0.7081989598962511, "learning_rate": 1.1376754630727616e-05, "loss": 0.0718, "num_input_tokens_seen": 134578176, "step": 516 }, { "epoch": 3.4036393713813067, "grad_norm": 0.6916080492642819, "learning_rate": 1.1289648720704347e-05, "loss": 0.073, "num_input_tokens_seen": 134840320, "step": 517 }, { "epoch": 3.41025641025641, "grad_norm": 0.7300988532144203, "learning_rate": 1.1202780195967836e-05, "loss": 0.0698, "num_input_tokens_seen": 135102464, "step": 518 }, { "epoch": 3.4168734491315136, "grad_norm": 0.6836195672441547, "learning_rate": 1.111615056058674e-05, "loss": 0.0734, "num_input_tokens_seen": 135364608, "step": 519 }, { "epoch": 3.423490488006617, "grad_norm": 0.7042584712844966, "learning_rate": 1.102976131449352e-05, "loss": 0.0732, "num_input_tokens_seen": 135626752, "step": 520 }, { "epoch": 3.4301075268817205, "grad_norm": 0.6592443491715577, "learning_rate": 1.0943613953458448e-05, "loss": 0.0702, "num_input_tokens_seen": 135888896, "step": 521 }, { "epoch": 3.436724565756824, "grad_norm": 0.6737665976311692, "learning_rate": 1.0857709969063734e-05, "loss": 0.0711, "num_input_tokens_seen": 136151040, "step": 522 }, { "epoch": 3.4433416046319274, "grad_norm": 0.6674839459806772, "learning_rate": 1.0772050848677682e-05, "loss": 0.0718, "num_input_tokens_seen": 136413184, "step": 523 }, { "epoch": 3.4499586435070304, "grad_norm": 0.6719895509682765, "learning_rate": 1.0686638075428947e-05, "loss": 0.0672, "num_input_tokens_seen": 136675328, "step": 524 }, { "epoch": 3.456575682382134, "grad_norm": 0.6756296196553735, "learning_rate": 1.0601473128180855e-05, "loss": 0.071, "num_input_tokens_seen": 136937472, "step": 525 }, { "epoch": 3.4631927212572373, "grad_norm": 0.6590609477211611, "learning_rate": 1.0516557481505795e-05, "loss": 0.0702, "num_input_tokens_seen": 137199616, "step": 526 }, { "epoch": 3.4698097601323408, "grad_norm": 0.6494091226770315, "learning_rate": 1.0431892605659685e-05, "loss": 0.0816, "num_input_tokens_seen": 137461760, "step": 527 }, { "epoch": 3.4764267990074442, "grad_norm": 0.7117194125057636, "learning_rate": 1.034747996655652e-05, "loss": 0.0753, "num_input_tokens_seen": 137723904, "step": 528 }, { "epoch": 3.4830438378825477, "grad_norm": 0.6790594971542001, "learning_rate": 1.0263321025742991e-05, "loss": 0.0703, "num_input_tokens_seen": 137986048, "step": 529 }, { "epoch": 3.489660876757651, "grad_norm": 0.6495417825384646, "learning_rate": 1.0179417240373183e-05, "loss": 0.0742, "num_input_tokens_seen": 138248192, "step": 530 }, { "epoch": 3.4962779156327546, "grad_norm": 0.6860713447018258, "learning_rate": 1.009577006318333e-05, "loss": 0.0706, "num_input_tokens_seen": 138510336, "step": 531 }, { "epoch": 3.5028949545078576, "grad_norm": 0.6084319486167035, "learning_rate": 1.0012380942466673e-05, "loss": 0.0682, "num_input_tokens_seen": 138772480, "step": 532 }, { "epoch": 3.509511993382961, "grad_norm": 0.6710562393229845, "learning_rate": 9.929251322048397e-06, "loss": 0.0697, "num_input_tokens_seen": 139034624, "step": 533 }, { "epoch": 3.5161290322580645, "grad_norm": 0.6662367171728351, "learning_rate": 9.8463826412606e-06, "loss": 0.0697, "num_input_tokens_seen": 139296768, "step": 534 }, { "epoch": 3.522746071133168, "grad_norm": 0.6609857197579974, "learning_rate": 9.763776334917399e-06, "loss": 0.0693, "num_input_tokens_seen": 139558912, "step": 535 }, { "epoch": 3.5293631100082714, "grad_norm": 0.6480323453400129, "learning_rate": 9.681433833290079e-06, "loss": 0.0695, "num_input_tokens_seen": 139821056, "step": 536 }, { "epoch": 3.5359801488833744, "grad_norm": 0.6496784196977948, "learning_rate": 9.599356562082329e-06, "loss": 0.064, "num_input_tokens_seen": 140083200, "step": 537 }, { "epoch": 3.542597187758478, "grad_norm": 0.6610199288253872, "learning_rate": 9.517545942405548e-06, "loss": 0.0699, "num_input_tokens_seen": 140345344, "step": 538 }, { "epoch": 3.5492142266335813, "grad_norm": 0.6626720993294117, "learning_rate": 9.436003390754258e-06, "loss": 0.0671, "num_input_tokens_seen": 140607488, "step": 539 }, { "epoch": 3.555831265508685, "grad_norm": 0.6594895617735759, "learning_rate": 9.354730318981562e-06, "loss": 0.0732, "num_input_tokens_seen": 140869632, "step": 540 }, { "epoch": 3.5624483043837882, "grad_norm": 0.65561962599018, "learning_rate": 9.273728134274701e-06, "loss": 0.067, "num_input_tokens_seen": 141131776, "step": 541 }, { "epoch": 3.5690653432588917, "grad_norm": 0.6434061253218011, "learning_rate": 9.192998239130699e-06, "loss": 0.0696, "num_input_tokens_seen": 141393920, "step": 542 }, { "epoch": 3.575682382133995, "grad_norm": 0.6235071777837268, "learning_rate": 9.112542031332075e-06, "loss": 0.067, "num_input_tokens_seen": 141656064, "step": 543 }, { "epoch": 3.5822994210090986, "grad_norm": 0.6916008151430184, "learning_rate": 9.032360903922635e-06, "loss": 0.064, "num_input_tokens_seen": 141918208, "step": 544 }, { "epoch": 3.588916459884202, "grad_norm": 0.6316220441846853, "learning_rate": 8.95245624518336e-06, "loss": 0.0685, "num_input_tokens_seen": 142180352, "step": 545 }, { "epoch": 3.595533498759305, "grad_norm": 0.6520805302384783, "learning_rate": 8.872829438608368e-06, "loss": 0.0673, "num_input_tokens_seen": 142442496, "step": 546 }, { "epoch": 3.6021505376344085, "grad_norm": 0.6414106319420477, "learning_rate": 8.793481862880953e-06, "loss": 0.0645, "num_input_tokens_seen": 142704640, "step": 547 }, { "epoch": 3.608767576509512, "grad_norm": 0.6353643058896132, "learning_rate": 8.714414891849737e-06, "loss": 0.0681, "num_input_tokens_seen": 142966784, "step": 548 }, { "epoch": 3.6153846153846154, "grad_norm": 0.6513946954232326, "learning_rate": 8.63562989450482e-06, "loss": 0.0736, "num_input_tokens_seen": 143228928, "step": 549 }, { "epoch": 3.622001654259719, "grad_norm": 0.6801727390096496, "learning_rate": 8.55712823495419e-06, "loss": 0.068, "num_input_tokens_seen": 143491072, "step": 550 }, { "epoch": 3.6286186931348223, "grad_norm": 0.6778450196456566, "learning_rate": 8.478911272399964e-06, "loss": 0.067, "num_input_tokens_seen": 143753216, "step": 551 }, { "epoch": 3.6352357320099253, "grad_norm": 0.6519049335263876, "learning_rate": 8.400980361114985e-06, "loss": 0.0689, "num_input_tokens_seen": 144015360, "step": 552 }, { "epoch": 3.641852770885029, "grad_norm": 0.6425369950760212, "learning_rate": 8.323336850419289e-06, "loss": 0.0645, "num_input_tokens_seen": 144277504, "step": 553 }, { "epoch": 3.6484698097601322, "grad_norm": 0.6472046415210123, "learning_rate": 8.245982084656765e-06, "loss": 0.0625, "num_input_tokens_seen": 144539648, "step": 554 }, { "epoch": 3.6550868486352357, "grad_norm": 0.6638130847765896, "learning_rate": 8.168917403171891e-06, "loss": 0.0647, "num_input_tokens_seen": 144801792, "step": 555 }, { "epoch": 3.661703887510339, "grad_norm": 0.7189252678002049, "learning_rate": 8.092144140286539e-06, "loss": 0.0684, "num_input_tokens_seen": 145063936, "step": 556 }, { "epoch": 3.6683209263854426, "grad_norm": 0.6024962011757397, "learning_rate": 8.015663625276864e-06, "loss": 0.0638, "num_input_tokens_seen": 145326080, "step": 557 }, { "epoch": 3.674937965260546, "grad_norm": 0.6380399282992304, "learning_rate": 7.939477182350277e-06, "loss": 0.0695, "num_input_tokens_seen": 145588224, "step": 558 }, { "epoch": 3.6815550041356495, "grad_norm": 0.6789043850107335, "learning_rate": 7.863586130622574e-06, "loss": 0.0668, "num_input_tokens_seen": 145850368, "step": 559 }, { "epoch": 3.688172043010753, "grad_norm": 0.6555026556001212, "learning_rate": 7.787991784095e-06, "loss": 0.0641, "num_input_tokens_seen": 146112512, "step": 560 }, { "epoch": 3.694789081885856, "grad_norm": 0.6418080447317104, "learning_rate": 7.712695451631621e-06, "loss": 0.0659, "num_input_tokens_seen": 146374656, "step": 561 }, { "epoch": 3.7014061207609594, "grad_norm": 0.6637076291897438, "learning_rate": 7.637698436936524e-06, "loss": 0.0659, "num_input_tokens_seen": 146636800, "step": 562 }, { "epoch": 3.708023159636063, "grad_norm": 0.6535273139092588, "learning_rate": 7.563002038531383e-06, "loss": 0.0678, "num_input_tokens_seen": 146898944, "step": 563 }, { "epoch": 3.7146401985111663, "grad_norm": 0.6436806576666099, "learning_rate": 7.488607549732843e-06, "loss": 0.0681, "num_input_tokens_seen": 147161088, "step": 564 }, { "epoch": 3.72125723738627, "grad_norm": 0.6432691211156978, "learning_rate": 7.414516258630244e-06, "loss": 0.0664, "num_input_tokens_seen": 147423232, "step": 565 }, { "epoch": 3.727874276261373, "grad_norm": 0.6425863914086812, "learning_rate": 7.340729448063252e-06, "loss": 0.069, "num_input_tokens_seen": 147685376, "step": 566 }, { "epoch": 3.7344913151364763, "grad_norm": 0.7237782721921531, "learning_rate": 7.267248395599632e-06, "loss": 0.0705, "num_input_tokens_seen": 147947520, "step": 567 }, { "epoch": 3.7411083540115797, "grad_norm": 0.5899400593495776, "learning_rate": 7.1940743735132126e-06, "loss": 0.0585, "num_input_tokens_seen": 148209664, "step": 568 }, { "epoch": 3.747725392886683, "grad_norm": 0.6298028583777657, "learning_rate": 7.121208648761743e-06, "loss": 0.0693, "num_input_tokens_seen": 148471808, "step": 569 }, { "epoch": 3.7543424317617866, "grad_norm": 0.6182729924844672, "learning_rate": 7.048652482965079e-06, "loss": 0.0644, "num_input_tokens_seen": 148733952, "step": 570 }, { "epoch": 3.76095947063689, "grad_norm": 0.6368990475721128, "learning_rate": 6.9764071323832145e-06, "loss": 0.0646, "num_input_tokens_seen": 148996096, "step": 571 }, { "epoch": 3.7675765095119935, "grad_norm": 0.6253822482714361, "learning_rate": 6.904473847894652e-06, "loss": 0.0625, "num_input_tokens_seen": 149258240, "step": 572 }, { "epoch": 3.774193548387097, "grad_norm": 0.6509773243060786, "learning_rate": 6.832853874974629e-06, "loss": 0.0665, "num_input_tokens_seen": 149520384, "step": 573 }, { "epoch": 3.7808105872622004, "grad_norm": 0.6348506211821107, "learning_rate": 6.761548453673647e-06, "loss": 0.0673, "num_input_tokens_seen": 149782528, "step": 574 }, { "epoch": 3.7874276261373034, "grad_norm": 0.6635181414850976, "learning_rate": 6.690558818595943e-06, "loss": 0.0627, "num_input_tokens_seen": 150044672, "step": 575 }, { "epoch": 3.794044665012407, "grad_norm": 0.6494205642323634, "learning_rate": 6.619886198878142e-06, "loss": 0.0701, "num_input_tokens_seen": 150306816, "step": 576 }, { "epoch": 3.8006617038875103, "grad_norm": 0.6664804797773478, "learning_rate": 6.549531818167981e-06, "loss": 0.0664, "num_input_tokens_seen": 150568960, "step": 577 }, { "epoch": 3.807278742762614, "grad_norm": 0.6170617316337039, "learning_rate": 6.479496894603066e-06, "loss": 0.057, "num_input_tokens_seen": 150831104, "step": 578 }, { "epoch": 3.8138957816377173, "grad_norm": 0.6049691188285755, "learning_rate": 6.409782640789874e-06, "loss": 0.0624, "num_input_tokens_seen": 151093248, "step": 579 }, { "epoch": 3.8205128205128203, "grad_norm": 0.6624674362233367, "learning_rate": 6.340390263782655e-06, "loss": 0.0712, "num_input_tokens_seen": 151355392, "step": 580 }, { "epoch": 3.8271298593879237, "grad_norm": 0.6641352801341951, "learning_rate": 6.271320965062638e-06, "loss": 0.0663, "num_input_tokens_seen": 151617536, "step": 581 }, { "epoch": 3.833746898263027, "grad_norm": 0.6545794002701258, "learning_rate": 6.202575940517122e-06, "loss": 0.0642, "num_input_tokens_seen": 151879680, "step": 582 }, { "epoch": 3.8403639371381306, "grad_norm": 0.6130739092946907, "learning_rate": 6.134156380418857e-06, "loss": 0.0656, "num_input_tokens_seen": 152141824, "step": 583 }, { "epoch": 3.846980976013234, "grad_norm": 0.6280514072168198, "learning_rate": 6.066063469405384e-06, "loss": 0.0605, "num_input_tokens_seen": 152403968, "step": 584 }, { "epoch": 3.8535980148883375, "grad_norm": 0.6275884943541263, "learning_rate": 5.998298386458545e-06, "loss": 0.0614, "num_input_tokens_seen": 152666112, "step": 585 }, { "epoch": 3.860215053763441, "grad_norm": 0.6842304679401343, "learning_rate": 5.9308623048840685e-06, "loss": 0.0667, "num_input_tokens_seen": 152928256, "step": 586 }, { "epoch": 3.8668320926385444, "grad_norm": 0.6503988865659992, "learning_rate": 5.863756392291248e-06, "loss": 0.0668, "num_input_tokens_seen": 153190400, "step": 587 }, { "epoch": 3.873449131513648, "grad_norm": 0.6594850152732623, "learning_rate": 5.796981810572724e-06, "loss": 0.0732, "num_input_tokens_seen": 153452544, "step": 588 }, { "epoch": 3.880066170388751, "grad_norm": 0.6343679688901765, "learning_rate": 5.7305397158843725e-06, "loss": 0.0585, "num_input_tokens_seen": 153714688, "step": 589 }, { "epoch": 3.8866832092638544, "grad_norm": 0.6431836549542845, "learning_rate": 5.664431258625305e-06, "loss": 0.0669, "num_input_tokens_seen": 153976832, "step": 590 }, { "epoch": 3.893300248138958, "grad_norm": 0.5999309364227289, "learning_rate": 5.598657583417896e-06, "loss": 0.0602, "num_input_tokens_seen": 154238976, "step": 591 }, { "epoch": 3.8999172870140613, "grad_norm": 0.6399699282669513, "learning_rate": 5.533219829088021e-06, "loss": 0.0643, "num_input_tokens_seen": 154501120, "step": 592 }, { "epoch": 3.9065343258891647, "grad_norm": 0.6312470036786655, "learning_rate": 5.468119128645319e-06, "loss": 0.0643, "num_input_tokens_seen": 154763264, "step": 593 }, { "epoch": 3.9131513647642677, "grad_norm": 0.5905035092164123, "learning_rate": 5.403356609263563e-06, "loss": 0.0625, "num_input_tokens_seen": 155025408, "step": 594 }, { "epoch": 3.919768403639371, "grad_norm": 0.6161426769945006, "learning_rate": 5.338933392261159e-06, "loss": 0.0665, "num_input_tokens_seen": 155287552, "step": 595 }, { "epoch": 3.9263854425144746, "grad_norm": 0.594382143213924, "learning_rate": 5.274850593081726e-06, "loss": 0.0613, "num_input_tokens_seen": 155549696, "step": 596 }, { "epoch": 3.933002481389578, "grad_norm": 0.6203429561074691, "learning_rate": 5.2111093212747845e-06, "loss": 0.0567, "num_input_tokens_seen": 155811840, "step": 597 }, { "epoch": 3.9396195202646815, "grad_norm": 0.6400006500847638, "learning_rate": 5.147710680476536e-06, "loss": 0.0675, "num_input_tokens_seen": 156073984, "step": 598 }, { "epoch": 3.946236559139785, "grad_norm": 0.6142056516422985, "learning_rate": 5.0846557683907755e-06, "loss": 0.0635, "num_input_tokens_seen": 156336128, "step": 599 }, { "epoch": 3.9528535980148884, "grad_norm": 0.6160208864975072, "learning_rate": 5.02194567676986e-06, "loss": 0.0643, "num_input_tokens_seen": 156598272, "step": 600 }, { "epoch": 3.959470636889992, "grad_norm": 0.6226902309804632, "learning_rate": 4.959581491395823e-06, "loss": 0.0636, "num_input_tokens_seen": 156860416, "step": 601 }, { "epoch": 3.9660876757650954, "grad_norm": 0.6279199663238604, "learning_rate": 4.897564292061568e-06, "loss": 0.0601, "num_input_tokens_seen": 157122560, "step": 602 }, { "epoch": 3.9727047146401984, "grad_norm": 0.612461694031015, "learning_rate": 4.835895152552178e-06, "loss": 0.0691, "num_input_tokens_seen": 157384704, "step": 603 }, { "epoch": 3.979321753515302, "grad_norm": 0.6279699097628595, "learning_rate": 4.7745751406263165e-06, "loss": 0.058, "num_input_tokens_seen": 157646848, "step": 604 }, { "epoch": 3.9859387923904053, "grad_norm": 0.6202131769337398, "learning_rate": 4.713605317997741e-06, "loss": 0.0622, "num_input_tokens_seen": 157908992, "step": 605 }, { "epoch": 3.9925558312655087, "grad_norm": 0.5955424851115548, "learning_rate": 4.652986740316928e-06, "loss": 0.0637, "num_input_tokens_seen": 158171136, "step": 606 }, { "epoch": 3.999172870140612, "grad_norm": 0.615956451068701, "learning_rate": 4.592720457152788e-06, "loss": 0.0662, "num_input_tokens_seen": 158433280, "step": 607 }, { "epoch": 4.0, "grad_norm": 0.615956451068701, "learning_rate": 4.532807511974491e-06, "loss": 0.059, "num_input_tokens_seen": 158466048, "step": 608 }, { "epoch": 4.0066170388751035, "grad_norm": 1.9322051031913439, "learning_rate": 4.47324894213341e-06, "loss": 0.0164, "num_input_tokens_seen": 158728192, "step": 609 }, { "epoch": 4.013234077750207, "grad_norm": 0.4558460196163836, "learning_rate": 4.414045778845144e-06, "loss": 0.0196, "num_input_tokens_seen": 158990336, "step": 610 }, { "epoch": 4.01985111662531, "grad_norm": 0.42853650955768524, "learning_rate": 4.355199047171685e-06, "loss": 0.0155, "num_input_tokens_seen": 159252480, "step": 611 }, { "epoch": 4.026468155500414, "grad_norm": 0.39608647720989215, "learning_rate": 4.2967097660036456e-06, "loss": 0.0169, "num_input_tokens_seen": 159514624, "step": 612 }, { "epoch": 4.033085194375517, "grad_norm": 0.3847681953616679, "learning_rate": 4.238578948042632e-06, "loss": 0.0158, "num_input_tokens_seen": 159776768, "step": 613 }, { "epoch": 4.039702233250621, "grad_norm": 0.3730962283899974, "learning_rate": 4.180807599783712e-06, "loss": 0.0156, "num_input_tokens_seen": 160038912, "step": 614 }, { "epoch": 4.046319272125724, "grad_norm": 0.31708613246483647, "learning_rate": 4.123396721497977e-06, "loss": 0.0137, "num_input_tokens_seen": 160301056, "step": 615 }, { "epoch": 4.052936311000827, "grad_norm": 0.2926072814394024, "learning_rate": 4.066347307215235e-06, "loss": 0.0151, "num_input_tokens_seen": 160563200, "step": 616 }, { "epoch": 4.05955334987593, "grad_norm": 0.3073138304585139, "learning_rate": 4.009660344706786e-06, "loss": 0.0134, "num_input_tokens_seen": 160825344, "step": 617 }, { "epoch": 4.066170388751034, "grad_norm": 0.29258815799018134, "learning_rate": 3.9533368154683365e-06, "loss": 0.013, "num_input_tokens_seen": 161087488, "step": 618 }, { "epoch": 4.072787427626137, "grad_norm": 0.2950143373147824, "learning_rate": 3.8973776947029864e-06, "loss": 0.0132, "num_input_tokens_seen": 161349632, "step": 619 }, { "epoch": 4.079404466501241, "grad_norm": 0.31508955936467287, "learning_rate": 3.8417839513043645e-06, "loss": 0.0127, "num_input_tokens_seen": 161611776, "step": 620 }, { "epoch": 4.086021505376344, "grad_norm": 0.33771585486450106, "learning_rate": 3.7865565478398314e-06, "loss": 0.0131, "num_input_tokens_seen": 161873920, "step": 621 }, { "epoch": 4.0926385442514475, "grad_norm": 0.29025238047918867, "learning_rate": 3.7316964405338357e-06, "loss": 0.0129, "num_input_tokens_seen": 162136064, "step": 622 }, { "epoch": 4.099255583126551, "grad_norm": 0.3337321590239914, "learning_rate": 3.6772045792513336e-06, "loss": 0.0123, "num_input_tokens_seen": 162398208, "step": 623 }, { "epoch": 4.105872622001654, "grad_norm": 0.35876768327894176, "learning_rate": 3.6230819074813737e-06, "loss": 0.0113, "num_input_tokens_seen": 162660352, "step": 624 }, { "epoch": 4.112489660876758, "grad_norm": 0.3292554272437977, "learning_rate": 3.5693293623207086e-06, "loss": 0.0129, "num_input_tokens_seen": 162922496, "step": 625 }, { "epoch": 4.119106699751861, "grad_norm": 0.35617949310152636, "learning_rate": 3.515947874457648e-06, "loss": 0.0135, "num_input_tokens_seen": 163184640, "step": 626 }, { "epoch": 4.125723738626965, "grad_norm": 0.3625371803838258, "learning_rate": 3.4629383681558576e-06, "loss": 0.0132, "num_input_tokens_seen": 163446784, "step": 627 }, { "epoch": 4.132340777502068, "grad_norm": 0.337416041836457, "learning_rate": 3.4103017612384323e-06, "loss": 0.0129, "num_input_tokens_seen": 163708928, "step": 628 }, { "epoch": 4.138957816377172, "grad_norm": 0.36408714339861237, "learning_rate": 3.358038965071955e-06, "loss": 0.0128, "num_input_tokens_seen": 163971072, "step": 629 }, { "epoch": 4.145574855252274, "grad_norm": 0.3530279053340078, "learning_rate": 3.3061508845507323e-06, "loss": 0.0132, "num_input_tokens_seen": 164233216, "step": 630 }, { "epoch": 4.152191894127378, "grad_norm": 0.3646811917338047, "learning_rate": 3.254638418081132e-06, "loss": 0.011, "num_input_tokens_seen": 164495360, "step": 631 }, { "epoch": 4.158808933002481, "grad_norm": 0.32293791810237155, "learning_rate": 3.2035024575660237e-06, "loss": 0.0133, "num_input_tokens_seen": 164757504, "step": 632 }, { "epoch": 4.165425971877585, "grad_norm": 0.37559363453473765, "learning_rate": 3.152743888389334e-06, "loss": 0.0118, "num_input_tokens_seen": 165019648, "step": 633 }, { "epoch": 4.172043010752688, "grad_norm": 0.3279963359002966, "learning_rate": 3.1023635894007085e-06, "loss": 0.0128, "num_input_tokens_seen": 165281792, "step": 634 }, { "epoch": 4.1786600496277915, "grad_norm": 0.3672827713383433, "learning_rate": 3.0523624329003324e-06, "loss": 0.0135, "num_input_tokens_seen": 165543936, "step": 635 }, { "epoch": 4.185277088502895, "grad_norm": 0.3151395869301595, "learning_rate": 3.002741284623764e-06, "loss": 0.0117, "num_input_tokens_seen": 165806080, "step": 636 }, { "epoch": 4.191894127377998, "grad_norm": 0.2861017283610466, "learning_rate": 2.953501003727019e-06, "loss": 0.0124, "num_input_tokens_seen": 166068224, "step": 637 }, { "epoch": 4.198511166253102, "grad_norm": 0.32170006278076657, "learning_rate": 2.904642442771616e-06, "loss": 0.0112, "num_input_tokens_seen": 166330368, "step": 638 }, { "epoch": 4.205128205128205, "grad_norm": 0.2827836397703704, "learning_rate": 2.8561664477098985e-06, "loss": 0.0118, "num_input_tokens_seen": 166592512, "step": 639 }, { "epoch": 4.211745244003309, "grad_norm": 0.3128634727340433, "learning_rate": 2.8080738578703054e-06, "loss": 0.0123, "num_input_tokens_seen": 166854656, "step": 640 }, { "epoch": 4.218362282878412, "grad_norm": 0.3016473217975274, "learning_rate": 2.7603655059429077e-06, "loss": 0.0119, "num_input_tokens_seen": 167116800, "step": 641 }, { "epoch": 4.224979321753516, "grad_norm": 0.2861165129091277, "learning_rate": 2.7130422179649563e-06, "loss": 0.0114, "num_input_tokens_seen": 167378944, "step": 642 }, { "epoch": 4.231596360628619, "grad_norm": 0.2599772487106339, "learning_rate": 2.666104813306558e-06, "loss": 0.0103, "num_input_tokens_seen": 167641088, "step": 643 }, { "epoch": 4.238213399503722, "grad_norm": 0.29133859690520947, "learning_rate": 2.619554104656563e-06, "loss": 0.013, "num_input_tokens_seen": 167903232, "step": 644 }, { "epoch": 4.244830438378825, "grad_norm": 0.26908297741174175, "learning_rate": 2.5733908980083988e-06, "loss": 0.0104, "num_input_tokens_seen": 168165376, "step": 645 }, { "epoch": 4.251447477253929, "grad_norm": 0.2659036178447148, "learning_rate": 2.527615992646201e-06, "loss": 0.0118, "num_input_tokens_seen": 168427520, "step": 646 }, { "epoch": 4.258064516129032, "grad_norm": 0.2901657036601284, "learning_rate": 2.4822301811309066e-06, "loss": 0.0134, "num_input_tokens_seen": 168689664, "step": 647 }, { "epoch": 4.2646815550041355, "grad_norm": 0.2887641626437273, "learning_rate": 2.437234249286588e-06, "loss": 0.0116, "num_input_tokens_seen": 168951808, "step": 648 }, { "epoch": 4.271298593879239, "grad_norm": 0.2972053563257295, "learning_rate": 2.3926289761867892e-06, "loss": 0.0123, "num_input_tokens_seen": 169213952, "step": 649 }, { "epoch": 4.277915632754342, "grad_norm": 0.28699830840097074, "learning_rate": 2.348415134141102e-06, "loss": 0.0111, "num_input_tokens_seen": 169476096, "step": 650 }, { "epoch": 4.284532671629446, "grad_norm": 0.2884445423690383, "learning_rate": 2.304593488681725e-06, "loss": 0.0118, "num_input_tokens_seen": 169738240, "step": 651 }, { "epoch": 4.291149710504549, "grad_norm": 0.2725595053360008, "learning_rate": 2.2611647985502636e-06, "loss": 0.0112, "num_input_tokens_seen": 170000384, "step": 652 }, { "epoch": 4.297766749379653, "grad_norm": 0.27157106563305955, "learning_rate": 2.218129815684572e-06, "loss": 0.0117, "num_input_tokens_seen": 170262528, "step": 653 }, { "epoch": 4.304383788254756, "grad_norm": 0.26577567593613255, "learning_rate": 2.1754892852057174e-06, "loss": 0.0121, "num_input_tokens_seen": 170524672, "step": 654 }, { "epoch": 4.31100082712986, "grad_norm": 0.31771200106763137, "learning_rate": 2.133243945405128e-06, "loss": 0.0129, "num_input_tokens_seen": 170786816, "step": 655 }, { "epoch": 4.317617866004963, "grad_norm": 0.29474813542572154, "learning_rate": 2.0913945277317384e-06, "loss": 0.0111, "num_input_tokens_seen": 171048960, "step": 656 }, { "epoch": 4.324234904880067, "grad_norm": 0.31855204764887335, "learning_rate": 2.0499417567794e-06, "loss": 0.013, "num_input_tokens_seen": 171311104, "step": 657 }, { "epoch": 4.330851943755169, "grad_norm": 0.2848669170044536, "learning_rate": 2.0088863502742665e-06, "loss": 0.011, "num_input_tokens_seen": 171573248, "step": 658 }, { "epoch": 4.337468982630273, "grad_norm": 0.2579942897761665, "learning_rate": 1.968229019062437e-06, "loss": 0.012, "num_input_tokens_seen": 171835392, "step": 659 }, { "epoch": 4.344086021505376, "grad_norm": 0.32880599270669536, "learning_rate": 1.927970467097573e-06, "loss": 0.013, "num_input_tokens_seen": 172097536, "step": 660 }, { "epoch": 4.3507030603804795, "grad_norm": 0.27491400258924964, "learning_rate": 1.8881113914287735e-06, "loss": 0.011, "num_input_tokens_seen": 172359680, "step": 661 }, { "epoch": 4.357320099255583, "grad_norm": 0.2782433259001828, "learning_rate": 1.8486524821884705e-06, "loss": 0.01, "num_input_tokens_seen": 172621824, "step": 662 }, { "epoch": 4.363937138130686, "grad_norm": 0.2813636030144918, "learning_rate": 1.8095944225804961e-06, "loss": 0.012, "num_input_tokens_seen": 172883968, "step": 663 }, { "epoch": 4.37055417700579, "grad_norm": 0.31012917378381333, "learning_rate": 1.7709378888682404e-06, "loss": 0.0112, "num_input_tokens_seen": 173146112, "step": 664 }, { "epoch": 4.377171215880893, "grad_norm": 0.2815088227497734, "learning_rate": 1.732683550362954e-06, "loss": 0.0102, "num_input_tokens_seen": 173408256, "step": 665 }, { "epoch": 4.383788254755997, "grad_norm": 0.28745562870408126, "learning_rate": 1.6948320694121527e-06, "loss": 0.0116, "num_input_tokens_seen": 173670400, "step": 666 }, { "epoch": 4.3904052936311, "grad_norm": 0.26246236882260804, "learning_rate": 1.6573841013881486e-06, "loss": 0.0103, "num_input_tokens_seen": 173932544, "step": 667 }, { "epoch": 4.397022332506204, "grad_norm": 0.25186617226497, "learning_rate": 1.6203402946767198e-06, "loss": 0.0114, "num_input_tokens_seen": 174194688, "step": 668 }, { "epoch": 4.403639371381307, "grad_norm": 0.32281818675069474, "learning_rate": 1.5837012906658484e-06, "loss": 0.0126, "num_input_tokens_seen": 174456832, "step": 669 }, { "epoch": 4.410256410256411, "grad_norm": 0.3347793755605357, "learning_rate": 1.5474677237346468e-06, "loss": 0.0112, "num_input_tokens_seen": 174718976, "step": 670 }, { "epoch": 4.416873449131514, "grad_norm": 0.2831919531375974, "learning_rate": 1.5116402212423664e-06, "loss": 0.0126, "num_input_tokens_seen": 174981120, "step": 671 }, { "epoch": 4.4234904880066175, "grad_norm": 0.33618014730043816, "learning_rate": 1.4762194035175286e-06, "loss": 0.0104, "num_input_tokens_seen": 175243264, "step": 672 }, { "epoch": 4.43010752688172, "grad_norm": 0.29188046505767673, "learning_rate": 1.4412058838471908e-06, "loss": 0.0125, "num_input_tokens_seen": 175505408, "step": 673 }, { "epoch": 4.4367245657568235, "grad_norm": 0.34185968510159637, "learning_rate": 1.406600268466321e-06, "loss": 0.0123, "num_input_tokens_seen": 175767552, "step": 674 }, { "epoch": 4.443341604631927, "grad_norm": 0.2751875387353734, "learning_rate": 1.3724031565473112e-06, "loss": 0.012, "num_input_tokens_seen": 176029696, "step": 675 }, { "epoch": 4.44995864350703, "grad_norm": 0.307753681169333, "learning_rate": 1.3386151401895919e-06, "loss": 0.0132, "num_input_tokens_seen": 176291840, "step": 676 }, { "epoch": 4.456575682382134, "grad_norm": 0.2941773412259061, "learning_rate": 1.3052368044093904e-06, "loss": 0.0122, "num_input_tokens_seen": 176553984, "step": 677 }, { "epoch": 4.463192721257237, "grad_norm": 0.29718218740929087, "learning_rate": 1.272268727129597e-06, "loss": 0.0105, "num_input_tokens_seen": 176816128, "step": 678 }, { "epoch": 4.469809760132341, "grad_norm": 0.28425687102272895, "learning_rate": 1.23971147916975e-06, "loss": 0.0114, "num_input_tokens_seen": 177078272, "step": 679 }, { "epoch": 4.476426799007444, "grad_norm": 0.28375740052067294, "learning_rate": 1.2075656242361732e-06, "loss": 0.0125, "num_input_tokens_seen": 177340416, "step": 680 }, { "epoch": 4.483043837882548, "grad_norm": 0.31804138215647476, "learning_rate": 1.1758317189121987e-06, "loss": 0.0103, "num_input_tokens_seen": 177602560, "step": 681 }, { "epoch": 4.489660876757651, "grad_norm": 0.27189609988366087, "learning_rate": 1.144510312648528e-06, "loss": 0.0112, "num_input_tokens_seen": 177864704, "step": 682 }, { "epoch": 4.496277915632755, "grad_norm": 0.27386745327873196, "learning_rate": 1.1136019477537397e-06, "loss": 0.0117, "num_input_tokens_seen": 178126848, "step": 683 }, { "epoch": 4.502894954507858, "grad_norm": 0.2807660974451893, "learning_rate": 1.0831071593848747e-06, "loss": 0.0102, "num_input_tokens_seen": 178388992, "step": 684 }, { "epoch": 4.5095119933829615, "grad_norm": 0.2807185122191924, "learning_rate": 1.0530264755381824e-06, "loss": 0.0134, "num_input_tokens_seen": 178651136, "step": 685 }, { "epoch": 4.516129032258064, "grad_norm": 0.30476848574607546, "learning_rate": 1.0233604170399813e-06, "loss": 0.0107, "num_input_tokens_seen": 178913280, "step": 686 }, { "epoch": 4.522746071133168, "grad_norm": 0.274310182626815, "learning_rate": 9.941094975376374e-07, "loss": 0.0129, "num_input_tokens_seen": 179175424, "step": 687 }, { "epoch": 4.529363110008271, "grad_norm": 0.2851652151422951, "learning_rate": 9.652742234906698e-07, "loss": 0.0108, "num_input_tokens_seen": 179437568, "step": 688 }, { "epoch": 4.535980148883374, "grad_norm": 0.2927927384652578, "learning_rate": 9.368550941619786e-07, "loss": 0.0124, "num_input_tokens_seen": 179699712, "step": 689 }, { "epoch": 4.542597187758478, "grad_norm": 0.2827600595189759, "learning_rate": 9.088526016092142e-07, "loss": 0.0111, "num_input_tokens_seen": 179961856, "step": 690 }, { "epoch": 4.549214226633581, "grad_norm": 0.2828143968736632, "learning_rate": 8.812672306762415e-07, "loss": 0.0115, "num_input_tokens_seen": 180224000, "step": 691 }, { "epoch": 4.555831265508685, "grad_norm": 0.28022586590651605, "learning_rate": 8.540994589847501e-07, "loss": 0.0104, "num_input_tokens_seen": 180486144, "step": 692 }, { "epoch": 4.562448304383788, "grad_norm": 0.26928602347071406, "learning_rate": 8.273497569259935e-07, "loss": 0.0115, "num_input_tokens_seen": 180748288, "step": 693 }, { "epoch": 4.569065343258892, "grad_norm": 0.2941361327088142, "learning_rate": 8.010185876526328e-07, "loss": 0.012, "num_input_tokens_seen": 181010432, "step": 694 }, { "epoch": 4.575682382133995, "grad_norm": 0.3099598307049961, "learning_rate": 7.751064070707248e-07, "loss": 0.0123, "num_input_tokens_seen": 181272576, "step": 695 }, { "epoch": 4.582299421009099, "grad_norm": 0.32349950320623483, "learning_rate": 7.496136638318218e-07, "loss": 0.0114, "num_input_tokens_seen": 181534720, "step": 696 }, { "epoch": 4.588916459884202, "grad_norm": 0.29772991851883923, "learning_rate": 7.245407993252101e-07, "loss": 0.0105, "num_input_tokens_seen": 181796864, "step": 697 }, { "epoch": 4.5955334987593055, "grad_norm": 0.25034733406750814, "learning_rate": 6.998882476702662e-07, "loss": 0.0115, "num_input_tokens_seen": 182059008, "step": 698 }, { "epoch": 4.602150537634409, "grad_norm": 0.27397283744512196, "learning_rate": 6.756564357089379e-07, "loss": 0.0115, "num_input_tokens_seen": 182321152, "step": 699 }, { "epoch": 4.608767576509512, "grad_norm": 0.27869832899439306, "learning_rate": 6.51845782998356e-07, "loss": 0.0106, "num_input_tokens_seen": 182583296, "step": 700 }, { "epoch": 4.615384615384615, "grad_norm": 0.30104324881793254, "learning_rate": 6.28456701803562e-07, "loss": 0.0131, "num_input_tokens_seen": 182845440, "step": 701 }, { "epoch": 4.622001654259718, "grad_norm": 0.2901161193436187, "learning_rate": 6.054895970903945e-07, "loss": 0.0124, "num_input_tokens_seen": 183107584, "step": 702 }, { "epoch": 4.628618693134822, "grad_norm": 0.30087510691273794, "learning_rate": 5.829448665184339e-07, "loss": 0.0104, "num_input_tokens_seen": 183369728, "step": 703 }, { "epoch": 4.635235732009925, "grad_norm": 0.28826623789313743, "learning_rate": 5.608229004341686e-07, "loss": 0.0107, "num_input_tokens_seen": 183631872, "step": 704 }, { "epoch": 4.641852770885029, "grad_norm": 0.2760868624828502, "learning_rate": 5.391240818642007e-07, "loss": 0.0123, "num_input_tokens_seen": 183894016, "step": 705 }, { "epoch": 4.648469809760132, "grad_norm": 0.30478625382300967, "learning_rate": 5.178487865086179e-07, "loss": 0.0111, "num_input_tokens_seen": 184156160, "step": 706 }, { "epoch": 4.655086848635236, "grad_norm": 0.2764507588923767, "learning_rate": 4.96997382734507e-07, "loss": 0.0112, "num_input_tokens_seen": 184418304, "step": 707 }, { "epoch": 4.661703887510339, "grad_norm": 0.2910917254737324, "learning_rate": 4.7657023156955636e-07, "loss": 0.0106, "num_input_tokens_seen": 184680448, "step": 708 }, { "epoch": 4.668320926385443, "grad_norm": 0.26960031865863143, "learning_rate": 4.5656768669582183e-07, "loss": 0.0105, "num_input_tokens_seen": 184942592, "step": 709 }, { "epoch": 4.674937965260546, "grad_norm": 0.2648467075940531, "learning_rate": 4.3699009444357344e-07, "loss": 0.0112, "num_input_tokens_seen": 185204736, "step": 710 }, { "epoch": 4.6815550041356495, "grad_norm": 0.3059089915950766, "learning_rate": 4.1783779378534727e-07, "loss": 0.0127, "num_input_tokens_seen": 185466880, "step": 711 }, { "epoch": 4.688172043010753, "grad_norm": 0.31042353213566753, "learning_rate": 3.9911111633002543e-07, "loss": 0.0122, "num_input_tokens_seen": 185729024, "step": 712 }, { "epoch": 4.694789081885856, "grad_norm": 0.2767065512527441, "learning_rate": 3.8081038631713195e-07, "loss": 0.0105, "num_input_tokens_seen": 185991168, "step": 713 }, { "epoch": 4.701406120760959, "grad_norm": 0.3001480980848603, "learning_rate": 3.6293592061119596e-07, "loss": 0.0109, "num_input_tokens_seen": 186253312, "step": 714 }, { "epoch": 4.708023159636063, "grad_norm": 0.2659712148033865, "learning_rate": 3.454880286962781e-07, "loss": 0.0116, "num_input_tokens_seen": 186515456, "step": 715 }, { "epoch": 4.714640198511166, "grad_norm": 0.2885929418188894, "learning_rate": 3.2846701267060243e-07, "loss": 0.0135, "num_input_tokens_seen": 186777600, "step": 716 }, { "epoch": 4.721257237386269, "grad_norm": 0.3206907985460206, "learning_rate": 3.1187316724133885e-07, "loss": 0.0108, "num_input_tokens_seen": 187039744, "step": 717 }, { "epoch": 4.727874276261373, "grad_norm": 0.25766634533303956, "learning_rate": 2.957067797194929e-07, "loss": 0.0124, "num_input_tokens_seen": 187301888, "step": 718 }, { "epoch": 4.734491315136476, "grad_norm": 0.2797485860867943, "learning_rate": 2.799681300149154e-07, "loss": 0.0122, "num_input_tokens_seen": 187564032, "step": 719 }, { "epoch": 4.74110835401158, "grad_norm": 0.3004663285838236, "learning_rate": 2.646574906314925e-07, "loss": 0.0107, "num_input_tokens_seen": 187826176, "step": 720 }, { "epoch": 4.747725392886683, "grad_norm": 0.29686494498566196, "learning_rate": 2.497751266623938e-07, "loss": 0.0102, "num_input_tokens_seen": 188088320, "step": 721 }, { "epoch": 4.754342431761787, "grad_norm": 0.28503781761594604, "learning_rate": 2.3532129578549834e-07, "loss": 0.0126, "num_input_tokens_seen": 188350464, "step": 722 }, { "epoch": 4.76095947063689, "grad_norm": 0.2828533817351923, "learning_rate": 2.212962482589287e-07, "loss": 0.0104, "num_input_tokens_seen": 188612608, "step": 723 }, { "epoch": 4.7675765095119935, "grad_norm": 0.28500905365606427, "learning_rate": 2.0770022691672387e-07, "loss": 0.0109, "num_input_tokens_seen": 188874752, "step": 724 }, { "epoch": 4.774193548387097, "grad_norm": 0.2722334517755525, "learning_rate": 1.9453346716462317e-07, "loss": 0.0114, "num_input_tokens_seen": 189136896, "step": 725 }, { "epoch": 4.7808105872622, "grad_norm": 0.32973797361712476, "learning_rate": 1.8179619697600292e-07, "loss": 0.0123, "num_input_tokens_seen": 189399040, "step": 726 }, { "epoch": 4.787427626137304, "grad_norm": 0.28386615493093464, "learning_rate": 1.6948863688791837e-07, "loss": 0.0106, "num_input_tokens_seen": 189661184, "step": 727 }, { "epoch": 4.794044665012407, "grad_norm": 0.26188027672390873, "learning_rate": 1.576109999972958e-07, "loss": 0.0102, "num_input_tokens_seen": 189923328, "step": 728 }, { "epoch": 4.80066170388751, "grad_norm": 0.27818657182928697, "learning_rate": 1.4616349195723245e-07, "loss": 0.0117, "num_input_tokens_seen": 190185472, "step": 729 }, { "epoch": 4.807278742762613, "grad_norm": 0.2713376652927809, "learning_rate": 1.351463109734441e-07, "loss": 0.0115, "num_input_tokens_seen": 190447616, "step": 730 }, { "epoch": 4.813895781637717, "grad_norm": 0.292337124074996, "learning_rate": 1.2455964780083152e-07, "loss": 0.0104, "num_input_tokens_seen": 190709760, "step": 731 }, { "epoch": 4.82051282051282, "grad_norm": 0.2624592360389428, "learning_rate": 1.1440368574017202e-07, "loss": 0.01, "num_input_tokens_seen": 190971904, "step": 732 }, { "epoch": 4.827129859387924, "grad_norm": 0.27915994440217407, "learning_rate": 1.0467860063495538e-07, "loss": 0.012, "num_input_tokens_seen": 191234048, "step": 733 }, { "epoch": 4.833746898263027, "grad_norm": 0.2988899889264722, "learning_rate": 9.538456086832237e-08, "loss": 0.0104, "num_input_tokens_seen": 191496192, "step": 734 }, { "epoch": 4.840363937138131, "grad_norm": 0.27437989333246415, "learning_rate": 8.652172736017816e-08, "loss": 0.0121, "num_input_tokens_seen": 191758336, "step": 735 }, { "epoch": 4.846980976013234, "grad_norm": 0.3475747345133441, "learning_rate": 7.809025356436961e-08, "loss": 0.0125, "num_input_tokens_seen": 192020480, "step": 736 }, { "epoch": 4.8535980148883375, "grad_norm": 0.32205466401651683, "learning_rate": 7.009028546606233e-08, "loss": 0.0111, "num_input_tokens_seen": 192282624, "step": 737 }, { "epoch": 4.860215053763441, "grad_norm": 0.4216447213602033, "learning_rate": 6.252196157919276e-08, "loss": 0.0106, "num_input_tokens_seen": 192544768, "step": 738 }, { "epoch": 4.866832092638544, "grad_norm": 0.26368939069501807, "learning_rate": 5.538541294407285e-08, "loss": 0.0128, "num_input_tokens_seen": 192806912, "step": 739 }, { "epoch": 4.873449131513648, "grad_norm": 0.27372393362295216, "learning_rate": 4.868076312512515e-08, "loss": 0.0123, "num_input_tokens_seen": 193069056, "step": 740 }, { "epoch": 4.880066170388751, "grad_norm": 0.3065832019500159, "learning_rate": 4.240812820874296e-08, "loss": 0.0104, "num_input_tokens_seen": 193331200, "step": 741 }, { "epoch": 4.886683209263855, "grad_norm": 0.2602099220704508, "learning_rate": 3.656761680127796e-08, "loss": 0.0114, "num_input_tokens_seen": 193593344, "step": 742 }, { "epoch": 4.893300248138958, "grad_norm": 0.28036077725532615, "learning_rate": 3.1159330027161204e-08, "loss": 0.0106, "num_input_tokens_seen": 193855488, "step": 743 }, { "epoch": 4.899917287014061, "grad_norm": 0.27157575496127295, "learning_rate": 2.618336152715728e-08, "loss": 0.0105, "num_input_tokens_seen": 194117632, "step": 744 }, { "epoch": 4.906534325889164, "grad_norm": 0.29757307933818905, "learning_rate": 2.1639797456723952e-08, "loss": 0.0111, "num_input_tokens_seen": 194379776, "step": 745 }, { "epoch": 4.913151364764268, "grad_norm": 0.28454058209618394, "learning_rate": 1.752871648454668e-08, "loss": 0.0108, "num_input_tokens_seen": 194641920, "step": 746 }, { "epoch": 4.919768403639371, "grad_norm": 0.22677688916259928, "learning_rate": 1.3850189791161927e-08, "loss": 0.0118, "num_input_tokens_seen": 194904064, "step": 747 }, { "epoch": 4.926385442514475, "grad_norm": 0.2865694178009504, "learning_rate": 1.0604281067724819e-08, "loss": 0.0114, "num_input_tokens_seen": 195166208, "step": 748 }, { "epoch": 4.933002481389578, "grad_norm": 0.29714518565256665, "learning_rate": 7.791046514907252e-09, "loss": 0.0104, "num_input_tokens_seen": 195428352, "step": 749 }, { "epoch": 4.9396195202646815, "grad_norm": 0.29893263843518475, "learning_rate": 5.41053484192644e-09, "loss": 0.0131, "num_input_tokens_seen": 195690496, "step": 750 }, { "epoch": 4.946236559139785, "grad_norm": 0.32165979568623193, "learning_rate": 3.462787265703926e-09, "loss": 0.0109, "num_input_tokens_seen": 195952640, "step": 751 }, { "epoch": 4.9528535980148884, "grad_norm": 0.2467567792729774, "learning_rate": 1.9478375101467104e-09, "loss": 0.0101, "num_input_tokens_seen": 196214784, "step": 752 }, { "epoch": 4.959470636889992, "grad_norm": 0.30153729872087115, "learning_rate": 8.657118055643843e-10, "loss": 0.0105, "num_input_tokens_seen": 196476928, "step": 753 }, { "epoch": 4.966087675765095, "grad_norm": 0.28356258452414257, "learning_rate": 2.164288882194887e-10, "loss": 0.0129, "num_input_tokens_seen": 196739072, "step": 754 }, { "epoch": 4.972704714640199, "grad_norm": 0.30560004687167447, "learning_rate": 0.0, "loss": 0.0115, "num_input_tokens_seen": 197001216, "step": 755 }, { "epoch": 4.972704714640199, "num_input_tokens_seen": 197001216, "step": 755, "total_flos": 314645277573120.0, "train_loss": 0.5845526486795568, "train_runtime": 18622.3068, "train_samples_per_second": 10.386, "train_steps_per_second": 0.041 } ], "logging_steps": 1, "max_steps": 755, "num_input_tokens_seen": 197001216, "num_train_epochs": 5, "save_steps": 5000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 314645277573120.0, "train_batch_size": 8, "trial_name": null, "trial_params": null }