| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.0, | |
| "eval_steps": 500, | |
| "global_step": 1614, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.012397334573066791, | |
| "grad_norm": 0.6915941834449768, | |
| "learning_rate": 0.00018, | |
| "loss": 2.3006, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.024794669146133583, | |
| "grad_norm": 0.6273291707038879, | |
| "learning_rate": 0.0001992534218166736, | |
| "loss": 2.1317, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.037192003719200374, | |
| "grad_norm": 0.45228877663612366, | |
| "learning_rate": 0.00019842389050186645, | |
| "loss": 1.9808, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.049589338292267165, | |
| "grad_norm": 0.3886250853538513, | |
| "learning_rate": 0.00019759435918705932, | |
| "loss": 1.8968, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.06198667286533395, | |
| "grad_norm": 0.3263673484325409, | |
| "learning_rate": 0.00019676482787225219, | |
| "loss": 1.9159, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.07438400743840075, | |
| "grad_norm": 0.3357856273651123, | |
| "learning_rate": 0.00019593529655744505, | |
| "loss": 1.9052, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.08678134201146753, | |
| "grad_norm": 0.3885687291622162, | |
| "learning_rate": 0.00019510576524263792, | |
| "loss": 1.8788, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.09917867658453433, | |
| "grad_norm": 0.33165615797042847, | |
| "learning_rate": 0.00019427623392783078, | |
| "loss": 1.876, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.11157601115760112, | |
| "grad_norm": 0.3510221242904663, | |
| "learning_rate": 0.00019344670261302365, | |
| "loss": 1.7992, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.1239733457306679, | |
| "grad_norm": 0.35817059874534607, | |
| "learning_rate": 0.0001926171712982165, | |
| "loss": 1.7386, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.1363706803037347, | |
| "grad_norm": 0.3222461938858032, | |
| "learning_rate": 0.00019178763998340938, | |
| "loss": 1.8766, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.1487680148768015, | |
| "grad_norm": 0.302379310131073, | |
| "learning_rate": 0.00019095810866860224, | |
| "loss": 1.833, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.16116534944986827, | |
| "grad_norm": 0.3621092736721039, | |
| "learning_rate": 0.00019012857735379514, | |
| "loss": 1.7433, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.17356268402293507, | |
| "grad_norm": 0.35103851556777954, | |
| "learning_rate": 0.00018929904603898798, | |
| "loss": 1.8552, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.18596001859600186, | |
| "grad_norm": 0.3761669993400574, | |
| "learning_rate": 0.00018846951472418084, | |
| "loss": 1.8072, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.19835735316906866, | |
| "grad_norm": 0.2791743576526642, | |
| "learning_rate": 0.00018763998340937373, | |
| "loss": 1.882, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.21075468774213543, | |
| "grad_norm": 0.28887104988098145, | |
| "learning_rate": 0.00018681045209456657, | |
| "loss": 1.945, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.22315202231520223, | |
| "grad_norm": 0.3621903955936432, | |
| "learning_rate": 0.00018598092077975944, | |
| "loss": 1.7506, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.23554935688826903, | |
| "grad_norm": 0.36899787187576294, | |
| "learning_rate": 0.00018515138946495233, | |
| "loss": 1.8959, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.2479466914613358, | |
| "grad_norm": 0.3354776203632355, | |
| "learning_rate": 0.00018432185815014517, | |
| "loss": 1.9282, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.2603440260344026, | |
| "grad_norm": 0.3030059337615967, | |
| "learning_rate": 0.00018349232683533803, | |
| "loss": 1.9876, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.2727413606074694, | |
| "grad_norm": 0.380134254693985, | |
| "learning_rate": 0.00018266279552053093, | |
| "loss": 1.795, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.2851386951805362, | |
| "grad_norm": 0.3484257161617279, | |
| "learning_rate": 0.00018183326420572376, | |
| "loss": 1.9131, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.297536029753603, | |
| "grad_norm": 0.3204387426376343, | |
| "learning_rate": 0.00018100373289091663, | |
| "loss": 1.9829, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.3099333643266698, | |
| "grad_norm": 0.3759450912475586, | |
| "learning_rate": 0.00018017420157610952, | |
| "loss": 1.9002, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.32233069889973653, | |
| "grad_norm": 0.3721698820590973, | |
| "learning_rate": 0.00017934467026130236, | |
| "loss": 1.8657, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.33472803347280333, | |
| "grad_norm": 0.35085615515708923, | |
| "learning_rate": 0.00017851513894649523, | |
| "loss": 1.8623, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.34712536804587013, | |
| "grad_norm": 0.35750696063041687, | |
| "learning_rate": 0.00017768560763168812, | |
| "loss": 1.8743, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.35952270261893693, | |
| "grad_norm": 0.3072109520435333, | |
| "learning_rate": 0.00017685607631688096, | |
| "loss": 1.9626, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.3719200371920037, | |
| "grad_norm": 0.40647512674331665, | |
| "learning_rate": 0.00017602654500207382, | |
| "loss": 1.8034, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.3843173717650705, | |
| "grad_norm": 0.3000311851501465, | |
| "learning_rate": 0.00017519701368726672, | |
| "loss": 1.8995, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.3967147063381373, | |
| "grad_norm": 0.36904624104499817, | |
| "learning_rate": 0.00017436748237245955, | |
| "loss": 1.82, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.40911204091120407, | |
| "grad_norm": 0.337799072265625, | |
| "learning_rate": 0.00017353795105765242, | |
| "loss": 1.8706, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.42150937548427087, | |
| "grad_norm": 0.4223800003528595, | |
| "learning_rate": 0.0001727084197428453, | |
| "loss": 1.8584, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.43390671005733766, | |
| "grad_norm": 0.34497585892677307, | |
| "learning_rate": 0.00017187888842803818, | |
| "loss": 1.8417, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.44630404463040446, | |
| "grad_norm": 0.34032031893730164, | |
| "learning_rate": 0.00017104935711323102, | |
| "loss": 1.7917, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.45870137920347126, | |
| "grad_norm": 0.4158859848976135, | |
| "learning_rate": 0.0001702198257984239, | |
| "loss": 1.8749, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.47109871377653806, | |
| "grad_norm": 0.36545196175575256, | |
| "learning_rate": 0.00016939029448361678, | |
| "loss": 1.8337, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.48349604834960486, | |
| "grad_norm": 0.32123473286628723, | |
| "learning_rate": 0.00016856076316880961, | |
| "loss": 1.9321, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.4958933829226716, | |
| "grad_norm": 0.45476439595222473, | |
| "learning_rate": 0.0001677312318540025, | |
| "loss": 1.8185, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.5082907174957384, | |
| "grad_norm": 0.3410905599594116, | |
| "learning_rate": 0.00016690170053919537, | |
| "loss": 1.9082, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.5206880520688052, | |
| "grad_norm": 0.3436656892299652, | |
| "learning_rate": 0.0001660721692243882, | |
| "loss": 1.7821, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.533085386641872, | |
| "grad_norm": 0.34343594312667847, | |
| "learning_rate": 0.0001652426379095811, | |
| "loss": 1.8184, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.5454827212149388, | |
| "grad_norm": 0.4309318959712982, | |
| "learning_rate": 0.00016441310659477397, | |
| "loss": 1.9124, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.5578800557880056, | |
| "grad_norm": 0.4032953679561615, | |
| "learning_rate": 0.0001635835752799668, | |
| "loss": 1.7962, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.5702773903610724, | |
| "grad_norm": 0.3726664185523987, | |
| "learning_rate": 0.0001627540439651597, | |
| "loss": 1.8158, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.5826747249341392, | |
| "grad_norm": 0.36948224902153015, | |
| "learning_rate": 0.00016192451265035257, | |
| "loss": 1.7539, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.595072059507206, | |
| "grad_norm": 0.33594000339508057, | |
| "learning_rate": 0.0001610949813355454, | |
| "loss": 1.9287, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.6074693940802728, | |
| "grad_norm": 0.3209264576435089, | |
| "learning_rate": 0.0001602654500207383, | |
| "loss": 1.8375, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.6198667286533396, | |
| "grad_norm": 0.38256484270095825, | |
| "learning_rate": 0.00015943591870593116, | |
| "loss": 1.817, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.6322640632264063, | |
| "grad_norm": 0.38966864347457886, | |
| "learning_rate": 0.000158606387391124, | |
| "loss": 1.7397, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.6446613977994731, | |
| "grad_norm": 0.45601052045822144, | |
| "learning_rate": 0.0001577768560763169, | |
| "loss": 1.7448, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.6570587323725399, | |
| "grad_norm": 0.39306697249412537, | |
| "learning_rate": 0.00015694732476150976, | |
| "loss": 1.8505, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.6694560669456067, | |
| "grad_norm": 0.4036141633987427, | |
| "learning_rate": 0.00015611779344670262, | |
| "loss": 1.9946, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.6818534015186735, | |
| "grad_norm": 0.34463125467300415, | |
| "learning_rate": 0.0001552882621318955, | |
| "loss": 2.0461, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.6942507360917403, | |
| "grad_norm": 0.3309987485408783, | |
| "learning_rate": 0.00015445873081708835, | |
| "loss": 1.8714, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.7066480706648071, | |
| "grad_norm": 0.40711745619773865, | |
| "learning_rate": 0.00015362919950228122, | |
| "loss": 1.9515, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.7190454052378739, | |
| "grad_norm": 0.4855351150035858, | |
| "learning_rate": 0.00015279966818747409, | |
| "loss": 1.928, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.7314427398109407, | |
| "grad_norm": 0.3159841299057007, | |
| "learning_rate": 0.00015197013687266695, | |
| "loss": 1.8648, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.7438400743840075, | |
| "grad_norm": 0.34454017877578735, | |
| "learning_rate": 0.00015114060555785982, | |
| "loss": 1.7614, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.7562374089570743, | |
| "grad_norm": 0.42112237215042114, | |
| "learning_rate": 0.00015031107424305268, | |
| "loss": 1.7943, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.768634743530141, | |
| "grad_norm": 0.4868924617767334, | |
| "learning_rate": 0.00014948154292824555, | |
| "loss": 1.8236, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.7810320781032078, | |
| "grad_norm": 0.27235159277915955, | |
| "learning_rate": 0.00014865201161343841, | |
| "loss": 1.8792, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.7934294126762746, | |
| "grad_norm": 0.36492735147476196, | |
| "learning_rate": 0.00014782248029863128, | |
| "loss": 1.8139, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.8058267472493414, | |
| "grad_norm": 0.3278910517692566, | |
| "learning_rate": 0.00014699294898382414, | |
| "loss": 1.9303, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.8182240818224081, | |
| "grad_norm": 0.4410141110420227, | |
| "learning_rate": 0.000146163417669017, | |
| "loss": 1.7458, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.8306214163954749, | |
| "grad_norm": 0.44660821557044983, | |
| "learning_rate": 0.00014533388635420988, | |
| "loss": 1.6484, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.8430187509685417, | |
| "grad_norm": 0.36396560072898865, | |
| "learning_rate": 0.00014450435503940274, | |
| "loss": 1.7457, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.8554160855416085, | |
| "grad_norm": 0.4536712169647217, | |
| "learning_rate": 0.0001436748237245956, | |
| "loss": 1.7763, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.8678134201146753, | |
| "grad_norm": 0.45438772439956665, | |
| "learning_rate": 0.00014284529240978847, | |
| "loss": 1.8736, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.8802107546877421, | |
| "grad_norm": 0.331462562084198, | |
| "learning_rate": 0.00014201576109498134, | |
| "loss": 1.9909, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.8926080892608089, | |
| "grad_norm": 0.29686763882637024, | |
| "learning_rate": 0.00014118622978017423, | |
| "loss": 1.9242, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.9050054238338757, | |
| "grad_norm": 0.4546560049057007, | |
| "learning_rate": 0.00014035669846536707, | |
| "loss": 1.8032, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.9174027584069425, | |
| "grad_norm": 0.3135245442390442, | |
| "learning_rate": 0.00013952716715055993, | |
| "loss": 1.835, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.9298000929800093, | |
| "grad_norm": 0.6448049545288086, | |
| "learning_rate": 0.00013869763583575283, | |
| "loss": 1.9127, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.9421974275530761, | |
| "grad_norm": 0.39725756645202637, | |
| "learning_rate": 0.00013786810452094567, | |
| "loss": 1.8041, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.9545947621261429, | |
| "grad_norm": 0.3762451708316803, | |
| "learning_rate": 0.00013703857320613853, | |
| "loss": 1.8003, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.9669920966992097, | |
| "grad_norm": 0.35813263058662415, | |
| "learning_rate": 0.00013620904189133142, | |
| "loss": 1.8474, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.9793894312722765, | |
| "grad_norm": 0.29999616742134094, | |
| "learning_rate": 0.00013537951057652426, | |
| "loss": 1.8705, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.9917867658453432, | |
| "grad_norm": 0.3202720880508423, | |
| "learning_rate": 0.00013454997926171713, | |
| "loss": 1.7752, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_loss": 1.8193774223327637, | |
| "eval_runtime": 78.8064, | |
| "eval_samples_per_second": 9.098, | |
| "eval_steps_per_second": 1.142, | |
| "step": 807 | |
| }, | |
| { | |
| "epoch": 1.00371920037192, | |
| "grad_norm": 0.30241405963897705, | |
| "learning_rate": 0.00013372044794691002, | |
| "loss": 1.8318, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 1.0161165349449868, | |
| "grad_norm": 0.3700416386127472, | |
| "learning_rate": 0.00013289091663210286, | |
| "loss": 1.8938, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 1.0285138695180536, | |
| "grad_norm": 0.4154430329799652, | |
| "learning_rate": 0.00013206138531729572, | |
| "loss": 1.8733, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 1.0409112040911204, | |
| "grad_norm": 0.38313189148902893, | |
| "learning_rate": 0.00013123185400248862, | |
| "loss": 1.8571, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 1.0533085386641872, | |
| "grad_norm": 0.23230139911174774, | |
| "learning_rate": 0.00013040232268768146, | |
| "loss": 1.798, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 1.065705873237254, | |
| "grad_norm": 0.3701108992099762, | |
| "learning_rate": 0.00012957279137287432, | |
| "loss": 1.8533, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 1.0781032078103208, | |
| "grad_norm": 0.29064834117889404, | |
| "learning_rate": 0.00012874326005806721, | |
| "loss": 1.7453, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 1.0905005423833876, | |
| "grad_norm": 0.3150763213634491, | |
| "learning_rate": 0.00012791372874326005, | |
| "loss": 1.8977, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 1.1028978769564544, | |
| "grad_norm": 0.428843230009079, | |
| "learning_rate": 0.00012708419742845292, | |
| "loss": 1.8688, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 1.1152952115295212, | |
| "grad_norm": 0.2608051896095276, | |
| "learning_rate": 0.0001262546661136458, | |
| "loss": 1.7242, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 1.127692546102588, | |
| "grad_norm": 0.3821583688259125, | |
| "learning_rate": 0.00012542513479883865, | |
| "loss": 1.9037, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 1.1400898806756548, | |
| "grad_norm": 0.28013911843299866, | |
| "learning_rate": 0.00012459560348403151, | |
| "loss": 1.8215, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 1.1524872152487216, | |
| "grad_norm": 0.30506208539009094, | |
| "learning_rate": 0.0001237660721692244, | |
| "loss": 1.8197, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 1.1648845498217884, | |
| "grad_norm": 0.29327717423439026, | |
| "learning_rate": 0.00012293654085441727, | |
| "loss": 1.7784, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 1.1772818843948551, | |
| "grad_norm": 0.23550163209438324, | |
| "learning_rate": 0.0001221070095396101, | |
| "loss": 1.9318, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 1.189679218967922, | |
| "grad_norm": 0.21349768340587616, | |
| "learning_rate": 0.000121277478224803, | |
| "loss": 1.8084, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 1.2020765535409887, | |
| "grad_norm": 0.34790855646133423, | |
| "learning_rate": 0.00012044794690999586, | |
| "loss": 1.7587, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 1.2144738881140555, | |
| "grad_norm": 0.2519979774951935, | |
| "learning_rate": 0.00011961841559518872, | |
| "loss": 1.8055, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 1.2268712226871223, | |
| "grad_norm": 0.3781174123287201, | |
| "learning_rate": 0.0001187888842803816, | |
| "loss": 1.8436, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 1.2392685572601891, | |
| "grad_norm": 0.26533016562461853, | |
| "learning_rate": 0.00011795935296557445, | |
| "loss": 1.6512, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 1.2516658918332557, | |
| "grad_norm": 0.2862655818462372, | |
| "learning_rate": 0.00011712982165076732, | |
| "loss": 2.0418, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 1.2640632264063227, | |
| "grad_norm": 0.27094656229019165, | |
| "learning_rate": 0.0001163002903359602, | |
| "loss": 1.7278, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 1.2764605609793893, | |
| "grad_norm": 0.29184144735336304, | |
| "learning_rate": 0.00011547075902115305, | |
| "loss": 1.6938, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 1.2888578955524563, | |
| "grad_norm": 0.30606576800346375, | |
| "learning_rate": 0.00011464122770634591, | |
| "loss": 1.6627, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 1.301255230125523, | |
| "grad_norm": 0.23827411234378815, | |
| "learning_rate": 0.0001138116963915388, | |
| "loss": 1.7671, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 1.31365256469859, | |
| "grad_norm": 0.3038440942764282, | |
| "learning_rate": 0.00011298216507673165, | |
| "loss": 1.9029, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 1.3260498992716565, | |
| "grad_norm": 0.22325201332569122, | |
| "learning_rate": 0.00011215263376192451, | |
| "loss": 1.7413, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 1.3384472338447233, | |
| "grad_norm": 0.30867546796798706, | |
| "learning_rate": 0.00011132310244711739, | |
| "loss": 1.8252, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 1.35084456841779, | |
| "grad_norm": 0.3236595392227173, | |
| "learning_rate": 0.00011049357113231024, | |
| "loss": 1.8856, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 1.363241902990857, | |
| "grad_norm": 0.32089242339134216, | |
| "learning_rate": 0.00010966403981750311, | |
| "loss": 1.8437, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 1.3756392375639237, | |
| "grad_norm": 0.28864434361457825, | |
| "learning_rate": 0.00010883450850269599, | |
| "loss": 1.7875, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 1.3880365721369905, | |
| "grad_norm": 0.3817216455936432, | |
| "learning_rate": 0.00010800497718788884, | |
| "loss": 1.8455, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 1.4004339067100573, | |
| "grad_norm": 0.22240202128887177, | |
| "learning_rate": 0.0001071754458730817, | |
| "loss": 1.7744, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 1.412831241283124, | |
| "grad_norm": 0.28633448481559753, | |
| "learning_rate": 0.00010634591455827458, | |
| "loss": 1.765, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 1.425228575856191, | |
| "grad_norm": 0.19945985078811646, | |
| "learning_rate": 0.00010551638324346745, | |
| "loss": 1.9739, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 1.4376259104292577, | |
| "grad_norm": 0.2815242111682892, | |
| "learning_rate": 0.0001046868519286603, | |
| "loss": 1.8369, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 1.4500232450023245, | |
| "grad_norm": 0.20174123346805573, | |
| "learning_rate": 0.00010385732061385318, | |
| "loss": 1.9351, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 1.4624205795753913, | |
| "grad_norm": 0.27899351716041565, | |
| "learning_rate": 0.00010302778929904605, | |
| "loss": 1.9544, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 1.474817914148458, | |
| "grad_norm": 0.22028954327106476, | |
| "learning_rate": 0.00010219825798423892, | |
| "loss": 1.852, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 1.4872152487215249, | |
| "grad_norm": 0.2520454525947571, | |
| "learning_rate": 0.00010136872666943178, | |
| "loss": 1.6232, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 1.4996125832945917, | |
| "grad_norm": 0.34104567766189575, | |
| "learning_rate": 0.00010053919535462464, | |
| "loss": 1.9285, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 1.5120099178676585, | |
| "grad_norm": 0.3216676414012909, | |
| "learning_rate": 9.970966403981751e-05, | |
| "loss": 1.6644, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 1.5244072524407253, | |
| "grad_norm": 0.22132286429405212, | |
| "learning_rate": 9.888013272501037e-05, | |
| "loss": 1.796, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 1.536804587013792, | |
| "grad_norm": 0.35185569524765015, | |
| "learning_rate": 9.805060141020324e-05, | |
| "loss": 1.7821, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 1.5492019215868589, | |
| "grad_norm": 0.16420242190361023, | |
| "learning_rate": 9.72210700953961e-05, | |
| "loss": 1.8054, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 1.5615992561599255, | |
| "grad_norm": 0.32100507616996765, | |
| "learning_rate": 9.639153878058897e-05, | |
| "loss": 1.8707, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 1.5739965907329925, | |
| "grad_norm": 0.2014254331588745, | |
| "learning_rate": 9.556200746578184e-05, | |
| "loss": 1.8087, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 1.586393925306059, | |
| "grad_norm": 0.23392640054225922, | |
| "learning_rate": 9.47324761509747e-05, | |
| "loss": 1.7327, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 1.598791259879126, | |
| "grad_norm": 0.19987891614437103, | |
| "learning_rate": 9.390294483616757e-05, | |
| "loss": 1.8399, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 1.6111885944521926, | |
| "grad_norm": 0.13928192853927612, | |
| "learning_rate": 9.307341352136043e-05, | |
| "loss": 1.9271, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 1.6235859290252597, | |
| "grad_norm": 0.3423463702201843, | |
| "learning_rate": 9.22438822065533e-05, | |
| "loss": 1.81, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 1.6359832635983262, | |
| "grad_norm": 0.31117212772369385, | |
| "learning_rate": 9.141435089174618e-05, | |
| "loss": 1.8919, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 1.6483805981713933, | |
| "grad_norm": 0.1769014447927475, | |
| "learning_rate": 9.058481957693903e-05, | |
| "loss": 1.692, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 1.6607779327444598, | |
| "grad_norm": 0.1725306212902069, | |
| "learning_rate": 8.97552882621319e-05, | |
| "loss": 1.7833, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 1.6731752673175269, | |
| "grad_norm": 0.15333665907382965, | |
| "learning_rate": 8.892575694732477e-05, | |
| "loss": 1.7129, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 1.6855726018905934, | |
| "grad_norm": 0.225737065076828, | |
| "learning_rate": 8.809622563251764e-05, | |
| "loss": 2.0034, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 1.6979699364636605, | |
| "grad_norm": 0.27135154604911804, | |
| "learning_rate": 8.726669431771049e-05, | |
| "loss": 1.7641, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 1.710367271036727, | |
| "grad_norm": 0.19318152964115143, | |
| "learning_rate": 8.643716300290337e-05, | |
| "loss": 1.9714, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 1.722764605609794, | |
| "grad_norm": 0.09128980338573456, | |
| "learning_rate": 8.560763168809624e-05, | |
| "loss": 1.6477, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 1.7351619401828606, | |
| "grad_norm": 0.20176462829113007, | |
| "learning_rate": 8.477810037328909e-05, | |
| "loss": 1.8707, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 1.7475592747559274, | |
| "grad_norm": 0.1746564656496048, | |
| "learning_rate": 8.394856905848197e-05, | |
| "loss": 1.8979, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 1.7599566093289942, | |
| "grad_norm": 0.8505027294158936, | |
| "learning_rate": 8.311903774367483e-05, | |
| "loss": 1.71, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 1.772353943902061, | |
| "grad_norm": 0.18071773648262024, | |
| "learning_rate": 8.22895064288677e-05, | |
| "loss": 1.9505, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 1.7847512784751278, | |
| "grad_norm": 0.29732322692871094, | |
| "learning_rate": 8.145997511406056e-05, | |
| "loss": 1.8041, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 1.7971486130481946, | |
| "grad_norm": 0.21795178949832916, | |
| "learning_rate": 8.063044379925343e-05, | |
| "loss": 1.7381, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 1.8095459476212614, | |
| "grad_norm": 0.23459061980247498, | |
| "learning_rate": 7.98009124844463e-05, | |
| "loss": 1.6636, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 1.8219432821943282, | |
| "grad_norm": 0.20895689725875854, | |
| "learning_rate": 7.897138116963916e-05, | |
| "loss": 1.8257, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 1.834340616767395, | |
| "grad_norm": 0.18129809200763702, | |
| "learning_rate": 7.814184985483203e-05, | |
| "loss": 1.9791, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 1.8467379513404618, | |
| "grad_norm": 0.1758158951997757, | |
| "learning_rate": 7.731231854002489e-05, | |
| "loss": 1.7657, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 1.8591352859135286, | |
| "grad_norm": 0.14756101369857788, | |
| "learning_rate": 7.648278722521776e-05, | |
| "loss": 1.7483, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 1.8715326204865954, | |
| "grad_norm": 0.14168867468833923, | |
| "learning_rate": 7.565325591041062e-05, | |
| "loss": 1.8844, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 1.8839299550596622, | |
| "grad_norm": 0.1295221447944641, | |
| "learning_rate": 7.482372459560349e-05, | |
| "loss": 1.8121, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 1.896327289632729, | |
| "grad_norm": 0.19356060028076172, | |
| "learning_rate": 7.399419328079635e-05, | |
| "loss": 1.5851, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 1.9087246242057958, | |
| "grad_norm": 0.20035363733768463, | |
| "learning_rate": 7.316466196598922e-05, | |
| "loss": 1.7063, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 1.9211219587788624, | |
| "grad_norm": 0.12139635533094406, | |
| "learning_rate": 7.233513065118208e-05, | |
| "loss": 1.7962, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 1.9335192933519294, | |
| "grad_norm": 0.10587511211633682, | |
| "learning_rate": 7.150559933637495e-05, | |
| "loss": 1.7143, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 1.945916627924996, | |
| "grad_norm": 0.18997065722942352, | |
| "learning_rate": 7.067606802156782e-05, | |
| "loss": 1.7636, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 1.958313962498063, | |
| "grad_norm": 0.15760765969753265, | |
| "learning_rate": 6.984653670676068e-05, | |
| "loss": 1.7672, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 1.9707112970711296, | |
| "grad_norm": 0.1490117609500885, | |
| "learning_rate": 6.901700539195355e-05, | |
| "loss": 1.6405, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 1.9831086316441966, | |
| "grad_norm": 0.09482846409082413, | |
| "learning_rate": 6.818747407714641e-05, | |
| "loss": 1.8141, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 1.9955059662172632, | |
| "grad_norm": 0.182535782456398, | |
| "learning_rate": 6.735794276233928e-05, | |
| "loss": 1.7745, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_loss": 1.8065738677978516, | |
| "eval_runtime": 78.829, | |
| "eval_samples_per_second": 9.096, | |
| "eval_steps_per_second": 1.142, | |
| "step": 1614 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 2421, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 5.6462587058139955e+17, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |