| { |
| "best_global_step": 1000, |
| "best_metric": 0.24901165068149567, |
| "best_model_checkpoint": "/root/workspace/finetune/checkpoints/train_run_06_qwen3_4b_formal/checkpoint-1000", |
| "epoch": 4.0, |
| "eval_steps": 500, |
| "global_step": 6572, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0121765601217656, |
| "grad_norm": 2.8322572708129883, |
| "learning_rate": 3.166666666666667e-07, |
| "loss": 1.119423007965088, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.0243531202435312, |
| "grad_norm": 2.00534987449646, |
| "learning_rate": 6.5e-07, |
| "loss": 0.8098940849304199, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.0365296803652968, |
| "grad_norm": 1.4319002628326416, |
| "learning_rate": 9.833333333333334e-07, |
| "loss": 0.6310151576995849, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.0487062404870624, |
| "grad_norm": 1.3265702724456787, |
| "learning_rate": 1.3166666666666666e-06, |
| "loss": 0.5526751518249512, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.060882800608828, |
| "grad_norm": 0.8091968297958374, |
| "learning_rate": 1.6500000000000003e-06, |
| "loss": 0.4834024906158447, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.0730593607305936, |
| "grad_norm": 0.8584772348403931, |
| "learning_rate": 1.9833333333333335e-06, |
| "loss": 0.44873433113098143, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.0852359208523592, |
| "grad_norm": 0.9426456093788147, |
| "learning_rate": 2.316666666666667e-06, |
| "loss": 0.38286705017089845, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.0974124809741248, |
| "grad_norm": 0.939085066318512, |
| "learning_rate": 2.6500000000000005e-06, |
| "loss": 0.38465352058410646, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.1095890410958904, |
| "grad_norm": 0.8642272353172302, |
| "learning_rate": 2.9833333333333337e-06, |
| "loss": 0.3305965900421143, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.121765601217656, |
| "grad_norm": 0.7353930473327637, |
| "learning_rate": 3.316666666666667e-06, |
| "loss": 0.2875316619873047, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.1339421613394216, |
| "grad_norm": 0.7686158418655396, |
| "learning_rate": 3.65e-06, |
| "loss": 0.3140716075897217, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.1461187214611872, |
| "grad_norm": 1.153424859046936, |
| "learning_rate": 3.983333333333334e-06, |
| "loss": 0.2985520839691162, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.1582952815829528, |
| "grad_norm": 0.7382195591926575, |
| "learning_rate": 4.316666666666667e-06, |
| "loss": 0.2449601411819458, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.1704718417047184, |
| "grad_norm": 0.7017386555671692, |
| "learning_rate": 4.65e-06, |
| "loss": 0.24566495418548584, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.182648401826484, |
| "grad_norm": 1.0982152223587036, |
| "learning_rate": 4.983333333333334e-06, |
| "loss": 0.20863604545593262, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.1948249619482496, |
| "grad_norm": 0.6582516431808472, |
| "learning_rate": 4.9998867856224845e-06, |
| "loss": 0.19693111181259154, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.2070015220700152, |
| "grad_norm": 0.6201784610748291, |
| "learning_rate": 4.999523005839606e-06, |
| "loss": 0.21833207607269287, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.2191780821917808, |
| "grad_norm": 0.7551426887512207, |
| "learning_rate": 4.998908383543311e-06, |
| "loss": 0.2382877826690674, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.2313546423135464, |
| "grad_norm": 0.6009605526924133, |
| "learning_rate": 4.9980429804147276e-06, |
| "loss": 0.21118538379669188, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.243531202435312, |
| "grad_norm": 0.7086856365203857, |
| "learning_rate": 4.996926883302385e-06, |
| "loss": 0.18367968797683715, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.2557077625570776, |
| "grad_norm": 0.964198648929596, |
| "learning_rate": 4.995560204213496e-06, |
| "loss": 0.22483460903167723, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.2678843226788432, |
| "grad_norm": 0.6596850156784058, |
| "learning_rate": 4.993943080302715e-06, |
| "loss": 0.22673561573028564, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.2800608828006088, |
| "grad_norm": 0.5922938585281372, |
| "learning_rate": 4.992075673858379e-06, |
| "loss": 0.185296630859375, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.2922374429223744, |
| "grad_norm": 0.6671269536018372, |
| "learning_rate": 4.989958172286214e-06, |
| "loss": 0.18437937498092652, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.30441400304414, |
| "grad_norm": 0.7064187526702881, |
| "learning_rate": 4.987590788090533e-06, |
| "loss": 0.1850834846496582, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.30441400304414, |
| "eval_loss": 0.2562323212623596, |
| "eval_runtime": 47.3626, |
| "eval_samples_per_second": 18.58, |
| "eval_steps_per_second": 18.58, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.3165905631659056, |
| "grad_norm": 0.7698928117752075, |
| "learning_rate": 4.984973758852904e-06, |
| "loss": 0.16346561908721924, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.3287671232876712, |
| "grad_norm": 0.4859939515590668, |
| "learning_rate": 4.982107347208317e-06, |
| "loss": 0.18838067054748536, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.3409436834094368, |
| "grad_norm": 0.5321416854858398, |
| "learning_rate": 4.978991840818816e-06, |
| "loss": 0.177593994140625, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.3531202435312024, |
| "grad_norm": 0.593889594078064, |
| "learning_rate": 4.975627552344638e-06, |
| "loss": 0.20775914192199707, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.365296803652968, |
| "grad_norm": 0.7926055192947388, |
| "learning_rate": 4.97201481941283e-06, |
| "loss": 0.16498700380325318, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.3774733637747336, |
| "grad_norm": 0.2909524440765381, |
| "learning_rate": 4.968154004583374e-06, |
| "loss": 0.17565951347351075, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.3896499238964992, |
| "grad_norm": 0.4575704038143158, |
| "learning_rate": 4.964045495312794e-06, |
| "loss": 0.16204673051834106, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.4018264840182648, |
| "grad_norm": 0.5366008281707764, |
| "learning_rate": 4.959689703915272e-06, |
| "loss": 0.17068564891815186, |
| "step": 660 |
| }, |
| { |
| "epoch": 0.4140030441400304, |
| "grad_norm": 0.5129569172859192, |
| "learning_rate": 4.95508706752128e-06, |
| "loss": 0.1589680790901184, |
| "step": 680 |
| }, |
| { |
| "epoch": 0.426179604261796, |
| "grad_norm": 0.4709528684616089, |
| "learning_rate": 4.9502380480337e-06, |
| "loss": 0.17568455934524535, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.4383561643835616, |
| "grad_norm": 0.6092886328697205, |
| "learning_rate": 4.9451431320814715e-06, |
| "loss": 0.16204804182052612, |
| "step": 720 |
| }, |
| { |
| "epoch": 0.4505327245053272, |
| "grad_norm": 0.5957323908805847, |
| "learning_rate": 4.939802830970762e-06, |
| "loss": 0.16562143564224244, |
| "step": 740 |
| }, |
| { |
| "epoch": 0.4627092846270928, |
| "grad_norm": 0.4758240580558777, |
| "learning_rate": 4.934217680633646e-06, |
| "loss": 0.17697544097900392, |
| "step": 760 |
| }, |
| { |
| "epoch": 0.4748858447488584, |
| "grad_norm": 0.865627646446228, |
| "learning_rate": 4.928388241574327e-06, |
| "loss": 0.1649466037750244, |
| "step": 780 |
| }, |
| { |
| "epoch": 0.487062404870624, |
| "grad_norm": 0.466294527053833, |
| "learning_rate": 4.922315098812883e-06, |
| "loss": 0.1602837324142456, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.4992389649923896, |
| "grad_norm": 0.6357060670852661, |
| "learning_rate": 4.9159988618265585e-06, |
| "loss": 0.142719042301178, |
| "step": 820 |
| }, |
| { |
| "epoch": 0.5114155251141552, |
| "grad_norm": 0.6055647730827332, |
| "learning_rate": 4.9094401644886e-06, |
| "loss": 0.14508233070373536, |
| "step": 840 |
| }, |
| { |
| "epoch": 0.5235920852359208, |
| "grad_norm": 0.45214834809303284, |
| "learning_rate": 4.902639665004641e-06, |
| "loss": 0.1821539044380188, |
| "step": 860 |
| }, |
| { |
| "epoch": 0.5357686453576864, |
| "grad_norm": 0.5735688805580139, |
| "learning_rate": 4.89559804584665e-06, |
| "loss": 0.16131887435913086, |
| "step": 880 |
| }, |
| { |
| "epoch": 0.547945205479452, |
| "grad_norm": 0.6279348731040955, |
| "learning_rate": 4.888316013684435e-06, |
| "loss": 0.17404688596725465, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.5601217656012176, |
| "grad_norm": 0.6474089026451111, |
| "learning_rate": 4.880794299314732e-06, |
| "loss": 0.14134640693664552, |
| "step": 920 |
| }, |
| { |
| "epoch": 0.5722983257229832, |
| "grad_norm": 0.5808464884757996, |
| "learning_rate": 4.87303365758786e-06, |
| "loss": 0.14891813993453978, |
| "step": 940 |
| }, |
| { |
| "epoch": 0.5844748858447488, |
| "grad_norm": 0.5440990328788757, |
| "learning_rate": 4.865034867331967e-06, |
| "loss": 0.1696299910545349, |
| "step": 960 |
| }, |
| { |
| "epoch": 0.5966514459665144, |
| "grad_norm": 0.6859214901924133, |
| "learning_rate": 4.856798731274874e-06, |
| "loss": 0.14085158109664916, |
| "step": 980 |
| }, |
| { |
| "epoch": 0.60882800608828, |
| "grad_norm": 0.3178713619709015, |
| "learning_rate": 4.84832607596351e-06, |
| "loss": 0.133053982257843, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.60882800608828, |
| "eval_loss": 0.24901165068149567, |
| "eval_runtime": 46.6976, |
| "eval_samples_per_second": 18.845, |
| "eval_steps_per_second": 18.845, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.6210045662100456, |
| "grad_norm": 0.3812738358974457, |
| "learning_rate": 4.8396177516809695e-06, |
| "loss": 0.12680984735488893, |
| "step": 1020 |
| }, |
| { |
| "epoch": 0.6331811263318112, |
| "grad_norm": 0.5174199342727661, |
| "learning_rate": 4.830674632361178e-06, |
| "loss": 0.14880582094192504, |
| "step": 1040 |
| }, |
| { |
| "epoch": 0.6453576864535768, |
| "grad_norm": 0.4705193042755127, |
| "learning_rate": 4.821497615501186e-06, |
| "loss": 0.1447562575340271, |
| "step": 1060 |
| }, |
| { |
| "epoch": 0.6575342465753424, |
| "grad_norm": 0.42298850417137146, |
| "learning_rate": 4.812087622071104e-06, |
| "loss": 0.15530819892883302, |
| "step": 1080 |
| }, |
| { |
| "epoch": 0.669710806697108, |
| "grad_norm": 0.30658382177352905, |
| "learning_rate": 4.80244559642167e-06, |
| "loss": 0.14426586627960206, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.6818873668188736, |
| "grad_norm": 0.4838867783546448, |
| "learning_rate": 4.792572506189489e-06, |
| "loss": 0.15436025857925414, |
| "step": 1120 |
| }, |
| { |
| "epoch": 0.6940639269406392, |
| "grad_norm": 0.716833770275116, |
| "learning_rate": 4.782469342199915e-06, |
| "loss": 0.14860854148864747, |
| "step": 1140 |
| }, |
| { |
| "epoch": 0.7062404870624048, |
| "grad_norm": 0.36538004875183105, |
| "learning_rate": 4.7721371183676205e-06, |
| "loss": 0.1313084125518799, |
| "step": 1160 |
| }, |
| { |
| "epoch": 0.7184170471841704, |
| "grad_norm": 0.5409316420555115, |
| "learning_rate": 4.761576871594841e-06, |
| "loss": 0.150812029838562, |
| "step": 1180 |
| }, |
| { |
| "epoch": 0.730593607305936, |
| "grad_norm": 0.5275493264198303, |
| "learning_rate": 4.750789661667318e-06, |
| "loss": 0.13278884887695314, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.7427701674277016, |
| "grad_norm": 0.5485584735870361, |
| "learning_rate": 4.739776571147943e-06, |
| "loss": 0.1612934350967407, |
| "step": 1220 |
| }, |
| { |
| "epoch": 0.7549467275494672, |
| "grad_norm": 0.5949460864067078, |
| "learning_rate": 4.728538705268116e-06, |
| "loss": 0.16211290359497071, |
| "step": 1240 |
| }, |
| { |
| "epoch": 0.7671232876712328, |
| "grad_norm": 0.43323376774787903, |
| "learning_rate": 4.717077191816824e-06, |
| "loss": 0.14386119842529296, |
| "step": 1260 |
| }, |
| { |
| "epoch": 0.7792998477929984, |
| "grad_norm": 0.6409174799919128, |
| "learning_rate": 4.705393181027463e-06, |
| "loss": 0.12942540645599365, |
| "step": 1280 |
| }, |
| { |
| "epoch": 0.791476407914764, |
| "grad_norm": 0.4871342182159424, |
| "learning_rate": 4.693487845462413e-06, |
| "loss": 0.14771063327789308, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.8036529680365296, |
| "grad_norm": 0.5108008980751038, |
| "learning_rate": 4.681362379895349e-06, |
| "loss": 0.1276724100112915, |
| "step": 1320 |
| }, |
| { |
| "epoch": 0.8158295281582952, |
| "grad_norm": 0.915285587310791, |
| "learning_rate": 4.6690180011913524e-06, |
| "loss": 0.1319241166114807, |
| "step": 1340 |
| }, |
| { |
| "epoch": 0.8280060882800608, |
| "grad_norm": 0.5282526612281799, |
| "learning_rate": 4.6564559481847795e-06, |
| "loss": 0.1557891011238098, |
| "step": 1360 |
| }, |
| { |
| "epoch": 0.8401826484018264, |
| "grad_norm": 0.46745216846466064, |
| "learning_rate": 4.643677481554947e-06, |
| "loss": 0.11075855493545532, |
| "step": 1380 |
| }, |
| { |
| "epoch": 0.852359208523592, |
| "grad_norm": 0.40246087312698364, |
| "learning_rate": 4.630683883699607e-06, |
| "loss": 0.1580789566040039, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.8645357686453576, |
| "grad_norm": 0.3718211352825165, |
| "learning_rate": 4.6174764586062556e-06, |
| "loss": 0.16006500720977784, |
| "step": 1420 |
| }, |
| { |
| "epoch": 0.8767123287671232, |
| "grad_norm": 0.4359384775161743, |
| "learning_rate": 4.6040565317212685e-06, |
| "loss": 0.1462727189064026, |
| "step": 1440 |
| }, |
| { |
| "epoch": 0.8888888888888888, |
| "grad_norm": 0.3503302037715912, |
| "learning_rate": 4.59042544981688e-06, |
| "loss": 0.14725338220596312, |
| "step": 1460 |
| }, |
| { |
| "epoch": 0.9010654490106544, |
| "grad_norm": 0.3662220537662506, |
| "learning_rate": 4.5765845808560334e-06, |
| "loss": 0.1304166793823242, |
| "step": 1480 |
| }, |
| { |
| "epoch": 0.91324200913242, |
| "grad_norm": 0.45357516407966614, |
| "learning_rate": 4.562535313855094e-06, |
| "loss": 0.1293134570121765, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.91324200913242, |
| "eval_loss": 0.25538697838783264, |
| "eval_runtime": 47.0026, |
| "eval_samples_per_second": 18.722, |
| "eval_steps_per_second": 18.722, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.9254185692541856, |
| "grad_norm": 0.31984779238700867, |
| "learning_rate": 4.548279058744451e-06, |
| "loss": 0.11359381675720215, |
| "step": 1520 |
| }, |
| { |
| "epoch": 0.9375951293759512, |
| "grad_norm": 0.6947388052940369, |
| "learning_rate": 4.533817246227024e-06, |
| "loss": 0.15145074129104613, |
| "step": 1540 |
| }, |
| { |
| "epoch": 0.9497716894977168, |
| "grad_norm": 0.5685822367668152, |
| "learning_rate": 4.519151327634685e-06, |
| "loss": 0.11953675746917725, |
| "step": 1560 |
| }, |
| { |
| "epoch": 0.9619482496194824, |
| "grad_norm": 0.2805669903755188, |
| "learning_rate": 4.504282774782605e-06, |
| "loss": 0.13977375030517578, |
| "step": 1580 |
| }, |
| { |
| "epoch": 0.974124809741248, |
| "grad_norm": 0.333103746175766, |
| "learning_rate": 4.489213079821551e-06, |
| "loss": 0.1338045358657837, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.9863013698630136, |
| "grad_norm": 0.5493115782737732, |
| "learning_rate": 4.4739437550881355e-06, |
| "loss": 0.11776142120361328, |
| "step": 1620 |
| }, |
| { |
| "epoch": 0.9984779299847792, |
| "grad_norm": 0.4903205931186676, |
| "learning_rate": 4.458476332953051e-06, |
| "loss": 0.12504475116729735, |
| "step": 1640 |
| }, |
| { |
| "epoch": 1.0103500761035007, |
| "grad_norm": 0.2682284712791443, |
| "learning_rate": 4.442812365667281e-06, |
| "loss": 0.08379222154617309, |
| "step": 1660 |
| }, |
| { |
| "epoch": 1.0225266362252663, |
| "grad_norm": 0.34869125485420227, |
| "learning_rate": 4.426953425206322e-06, |
| "loss": 0.08407147526741028, |
| "step": 1680 |
| }, |
| { |
| "epoch": 1.034703196347032, |
| "grad_norm": 0.38522422313690186, |
| "learning_rate": 4.410901103112434e-06, |
| "loss": 0.08041079640388489, |
| "step": 1700 |
| }, |
| { |
| "epoch": 1.0468797564687975, |
| "grad_norm": 0.4700411856174469, |
| "learning_rate": 4.394657010334908e-06, |
| "loss": 0.07876392006874085, |
| "step": 1720 |
| }, |
| { |
| "epoch": 1.059056316590563, |
| "grad_norm": 0.3719494938850403, |
| "learning_rate": 4.378222777068406e-06, |
| "loss": 0.10302903652191162, |
| "step": 1740 |
| }, |
| { |
| "epoch": 1.0712328767123287, |
| "grad_norm": 0.406753808259964, |
| "learning_rate": 4.361600052589358e-06, |
| "loss": 0.08733606934547425, |
| "step": 1760 |
| }, |
| { |
| "epoch": 1.0834094368340943, |
| "grad_norm": 0.19822958111763, |
| "learning_rate": 4.344790505090447e-06, |
| "loss": 0.08532609939575195, |
| "step": 1780 |
| }, |
| { |
| "epoch": 1.09558599695586, |
| "grad_norm": 0.3569728434085846, |
| "learning_rate": 4.327795821513195e-06, |
| "loss": 0.08734336495399475, |
| "step": 1800 |
| }, |
| { |
| "epoch": 1.1077625570776255, |
| "grad_norm": 0.4549162685871124, |
| "learning_rate": 4.3106177073786684e-06, |
| "loss": 0.0913870632648468, |
| "step": 1820 |
| }, |
| { |
| "epoch": 1.119939117199391, |
| "grad_norm": 0.5802178978919983, |
| "learning_rate": 4.293257886616318e-06, |
| "loss": 0.08115516304969787, |
| "step": 1840 |
| }, |
| { |
| "epoch": 1.1321156773211567, |
| "grad_norm": 0.30524641275405884, |
| "learning_rate": 4.275718101390975e-06, |
| "loss": 0.08891176581382751, |
| "step": 1860 |
| }, |
| { |
| "epoch": 1.1442922374429223, |
| "grad_norm": 0.33804091811180115, |
| "learning_rate": 4.25800011192801e-06, |
| "loss": 0.07950961589813232, |
| "step": 1880 |
| }, |
| { |
| "epoch": 1.156468797564688, |
| "grad_norm": 0.33472684025764465, |
| "learning_rate": 4.240105696336687e-06, |
| "loss": 0.08310645222663879, |
| "step": 1900 |
| }, |
| { |
| "epoch": 1.1686453576864535, |
| "grad_norm": 0.3032575845718384, |
| "learning_rate": 4.222036650431715e-06, |
| "loss": 0.07682961225509644, |
| "step": 1920 |
| }, |
| { |
| "epoch": 1.180821917808219, |
| "grad_norm": 0.47542238235473633, |
| "learning_rate": 4.203794787553032e-06, |
| "loss": 0.07520227432250977, |
| "step": 1940 |
| }, |
| { |
| "epoch": 1.1929984779299847, |
| "grad_norm": 0.39914897084236145, |
| "learning_rate": 4.185381938383821e-06, |
| "loss": 0.0754019558429718, |
| "step": 1960 |
| }, |
| { |
| "epoch": 1.2051750380517503, |
| "grad_norm": 0.4697635769844055, |
| "learning_rate": 4.166799950766793e-06, |
| "loss": 0.08085885643959045, |
| "step": 1980 |
| }, |
| { |
| "epoch": 1.217351598173516, |
| "grad_norm": 0.28078529238700867, |
| "learning_rate": 4.14805068951874e-06, |
| "loss": 0.0884653627872467, |
| "step": 2000 |
| }, |
| { |
| "epoch": 1.217351598173516, |
| "eval_loss": 0.2661186456680298, |
| "eval_runtime": 46.548, |
| "eval_samples_per_second": 18.905, |
| "eval_steps_per_second": 18.905, |
| "step": 2000 |
| }, |
| { |
| "epoch": 1.2295281582952815, |
| "grad_norm": 0.2885560989379883, |
| "learning_rate": 4.1291360362433965e-06, |
| "loss": 0.06684748530387878, |
| "step": 2020 |
| }, |
| { |
| "epoch": 1.241704718417047, |
| "grad_norm": 0.24163688719272614, |
| "learning_rate": 4.110057889142601e-06, |
| "loss": 0.0720324158668518, |
| "step": 2040 |
| }, |
| { |
| "epoch": 1.2538812785388127, |
| "grad_norm": 0.52589350938797, |
| "learning_rate": 4.090818162825804e-06, |
| "loss": 0.08799988031387329, |
| "step": 2060 |
| }, |
| { |
| "epoch": 1.2660578386605783, |
| "grad_norm": 0.35164448618888855, |
| "learning_rate": 4.071418788117926e-06, |
| "loss": 0.09275985956192016, |
| "step": 2080 |
| }, |
| { |
| "epoch": 1.278234398782344, |
| "grad_norm": 0.4981421232223511, |
| "learning_rate": 4.0518617118655845e-06, |
| "loss": 0.08431113958358764, |
| "step": 2100 |
| }, |
| { |
| "epoch": 1.2904109589041095, |
| "grad_norm": 0.2867731750011444, |
| "learning_rate": 4.032148896741717e-06, |
| "loss": 0.09995608925819396, |
| "step": 2120 |
| }, |
| { |
| "epoch": 1.302587519025875, |
| "grad_norm": 0.5612165331840515, |
| "learning_rate": 4.012282321048618e-06, |
| "loss": 0.07387629747390748, |
| "step": 2140 |
| }, |
| { |
| "epoch": 1.3147640791476407, |
| "grad_norm": 0.4880141317844391, |
| "learning_rate": 3.992263978519398e-06, |
| "loss": 0.07667248249053955, |
| "step": 2160 |
| }, |
| { |
| "epoch": 1.3269406392694063, |
| "grad_norm": 0.3143049478530884, |
| "learning_rate": 3.972095878117904e-06, |
| "loss": 0.09203824400901794, |
| "step": 2180 |
| }, |
| { |
| "epoch": 1.339117199391172, |
| "grad_norm": 0.47859013080596924, |
| "learning_rate": 3.951780043837107e-06, |
| "loss": 0.07835246920585633, |
| "step": 2200 |
| }, |
| { |
| "epoch": 1.3512937595129375, |
| "grad_norm": 0.28448912501335144, |
| "learning_rate": 3.9313185144959835e-06, |
| "loss": 0.08577624559402466, |
| "step": 2220 |
| }, |
| { |
| "epoch": 1.363470319634703, |
| "grad_norm": 0.32139304280281067, |
| "learning_rate": 3.9107133435349025e-06, |
| "loss": 0.0789969801902771, |
| "step": 2240 |
| }, |
| { |
| "epoch": 1.3756468797564687, |
| "grad_norm": 0.4797567129135132, |
| "learning_rate": 3.889966598809557e-06, |
| "loss": 0.07151145935058593, |
| "step": 2260 |
| }, |
| { |
| "epoch": 1.3878234398782343, |
| "grad_norm": 0.2404891699552536, |
| "learning_rate": 3.869080362383437e-06, |
| "loss": 0.09204544425010681, |
| "step": 2280 |
| }, |
| { |
| "epoch": 1.4, |
| "grad_norm": 0.328392893075943, |
| "learning_rate": 3.848056730318881e-06, |
| "loss": 0.11079612970352173, |
| "step": 2300 |
| }, |
| { |
| "epoch": 1.4121765601217655, |
| "grad_norm": 0.2993980646133423, |
| "learning_rate": 3.826897812466728e-06, |
| "loss": 0.06770140528678895, |
| "step": 2320 |
| }, |
| { |
| "epoch": 1.4243531202435311, |
| "grad_norm": 0.47816380858421326, |
| "learning_rate": 3.8056057322545763e-06, |
| "loss": 0.08210510611534119, |
| "step": 2340 |
| }, |
| { |
| "epoch": 1.4365296803652967, |
| "grad_norm": 0.38082119822502136, |
| "learning_rate": 3.7841826264736888e-06, |
| "loss": 0.09583572745323181, |
| "step": 2360 |
| }, |
| { |
| "epoch": 1.4487062404870623, |
| "grad_norm": 0.3811774253845215, |
| "learning_rate": 3.762630645064547e-06, |
| "loss": 0.09235450625419617, |
| "step": 2380 |
| }, |
| { |
| "epoch": 1.460882800608828, |
| "grad_norm": 0.3869916498661041, |
| "learning_rate": 3.7409519509010985e-06, |
| "loss": 0.08658097982406616, |
| "step": 2400 |
| }, |
| { |
| "epoch": 1.4730593607305935, |
| "grad_norm": 0.5042543411254883, |
| "learning_rate": 3.7191487195736915e-06, |
| "loss": 0.08892765045166015, |
| "step": 2420 |
| }, |
| { |
| "epoch": 1.4852359208523591, |
| "grad_norm": 0.4064173996448517, |
| "learning_rate": 3.697223139170748e-06, |
| "loss": 0.07849371433258057, |
| "step": 2440 |
| }, |
| { |
| "epoch": 1.4974124809741247, |
| "grad_norm": 0.3958194851875305, |
| "learning_rate": 3.6751774100591716e-06, |
| "loss": 0.07469035387039184, |
| "step": 2460 |
| }, |
| { |
| "epoch": 1.5095890410958903, |
| "grad_norm": 0.39424121379852295, |
| "learning_rate": 3.6530137446635265e-06, |
| "loss": 0.0782626211643219, |
| "step": 2480 |
| }, |
| { |
| "epoch": 1.521765601217656, |
| "grad_norm": 0.46179333329200745, |
| "learning_rate": 3.630734367244012e-06, |
| "loss": 0.08304058909416198, |
| "step": 2500 |
| }, |
| { |
| "epoch": 1.521765601217656, |
| "eval_loss": 0.2771773338317871, |
| "eval_runtime": 46.9989, |
| "eval_samples_per_second": 18.724, |
| "eval_steps_per_second": 18.724, |
| "step": 2500 |
| }, |
| { |
| "epoch": 1.5339421613394215, |
| "grad_norm": 0.3447970151901245, |
| "learning_rate": 3.6083415136732374e-06, |
| "loss": 0.08037537932395936, |
| "step": 2520 |
| }, |
| { |
| "epoch": 1.5461187214611871, |
| "grad_norm": 0.544224202632904, |
| "learning_rate": 3.585837431211845e-06, |
| "loss": 0.08990358114242554, |
| "step": 2540 |
| }, |
| { |
| "epoch": 1.5582952815829527, |
| "grad_norm": 0.2808915376663208, |
| "learning_rate": 3.563224378282978e-06, |
| "loss": 0.0773526132106781, |
| "step": 2560 |
| }, |
| { |
| "epoch": 1.5704718417047183, |
| "grad_norm": 0.27800413966178894, |
| "learning_rate": 3.5405046242456396e-06, |
| "loss": 0.0777865469455719, |
| "step": 2580 |
| }, |
| { |
| "epoch": 1.582648401826484, |
| "grad_norm": 0.4247933328151703, |
| "learning_rate": 3.517680449166943e-06, |
| "loss": 0.08037815093994141, |
| "step": 2600 |
| }, |
| { |
| "epoch": 1.5948249619482495, |
| "grad_norm": 0.748912513256073, |
| "learning_rate": 3.4947541435932976e-06, |
| "loss": 0.08837634325027466, |
| "step": 2620 |
| }, |
| { |
| "epoch": 1.6070015220700151, |
| "grad_norm": 0.29709526896476746, |
| "learning_rate": 3.471728008320532e-06, |
| "loss": 0.08201563358306885, |
| "step": 2640 |
| }, |
| { |
| "epoch": 1.6191780821917807, |
| "grad_norm": 0.3287765681743622, |
| "learning_rate": 3.4486043541630066e-06, |
| "loss": 0.07379403114318847, |
| "step": 2660 |
| }, |
| { |
| "epoch": 1.6313546423135463, |
| "grad_norm": 0.4224306643009186, |
| "learning_rate": 3.425385501721696e-06, |
| "loss": 0.08871785402297974, |
| "step": 2680 |
| }, |
| { |
| "epoch": 1.643531202435312, |
| "grad_norm": 0.5213542580604553, |
| "learning_rate": 3.4020737811513107e-06, |
| "loss": 0.07757498621940613, |
| "step": 2700 |
| }, |
| { |
| "epoch": 1.6557077625570775, |
| "grad_norm": 0.21830426156520844, |
| "learning_rate": 3.3786715319264483e-06, |
| "loss": 0.08565697669982911, |
| "step": 2720 |
| }, |
| { |
| "epoch": 1.6678843226788431, |
| "grad_norm": 0.30557703971862793, |
| "learning_rate": 3.355181102606816e-06, |
| "loss": 0.08129348754882812, |
| "step": 2740 |
| }, |
| { |
| "epoch": 1.6800608828006087, |
| "grad_norm": 0.453703910112381, |
| "learning_rate": 3.331604850601533e-06, |
| "loss": 0.07875375747680664, |
| "step": 2760 |
| }, |
| { |
| "epoch": 1.6922374429223743, |
| "grad_norm": 0.33889397978782654, |
| "learning_rate": 3.307945141932556e-06, |
| "loss": 0.08989614248275757, |
| "step": 2780 |
| }, |
| { |
| "epoch": 1.70441400304414, |
| "grad_norm": 0.21036894619464874, |
| "learning_rate": 3.2842043509972294e-06, |
| "loss": 0.08251298069953919, |
| "step": 2800 |
| }, |
| { |
| "epoch": 1.7165905631659055, |
| "grad_norm": 0.2764931321144104, |
| "learning_rate": 3.2603848603300026e-06, |
| "loss": 0.07760271430015564, |
| "step": 2820 |
| }, |
| { |
| "epoch": 1.7287671232876711, |
| "grad_norm": 0.3566988706588745, |
| "learning_rate": 3.236489060363329e-06, |
| "loss": 0.07395396232604981, |
| "step": 2840 |
| }, |
| { |
| "epoch": 1.7409436834094367, |
| "grad_norm": 0.3616081476211548, |
| "learning_rate": 3.212519349187766e-06, |
| "loss": 0.07028600573539734, |
| "step": 2860 |
| }, |
| { |
| "epoch": 1.7531202435312023, |
| "grad_norm": 0.41925984621047974, |
| "learning_rate": 3.188478132311319e-06, |
| "loss": 0.08469281196594239, |
| "step": 2880 |
| }, |
| { |
| "epoch": 1.765296803652968, |
| "grad_norm": 0.4866960346698761, |
| "learning_rate": 3.164367822418029e-06, |
| "loss": 0.09567424058914184, |
| "step": 2900 |
| }, |
| { |
| "epoch": 1.7774733637747335, |
| "grad_norm": 0.2977049648761749, |
| "learning_rate": 3.1401908391258474e-06, |
| "loss": 0.07239987254142762, |
| "step": 2920 |
| }, |
| { |
| "epoch": 1.7896499238964991, |
| "grad_norm": 0.3723915219306946, |
| "learning_rate": 3.1159496087438098e-06, |
| "loss": 0.0891954243183136, |
| "step": 2940 |
| }, |
| { |
| "epoch": 1.8018264840182647, |
| "grad_norm": 0.22685140371322632, |
| "learning_rate": 3.0916465640285426e-06, |
| "loss": 0.07796849608421326, |
| "step": 2960 |
| }, |
| { |
| "epoch": 1.8140030441400303, |
| "grad_norm": 0.23534800112247467, |
| "learning_rate": 3.0672841439401223e-06, |
| "loss": 0.08645985722541809, |
| "step": 2980 |
| }, |
| { |
| "epoch": 1.826179604261796, |
| "grad_norm": 0.40188542008399963, |
| "learning_rate": 3.0428647933973103e-06, |
| "loss": 0.08427774310111999, |
| "step": 3000 |
| }, |
| { |
| "epoch": 1.826179604261796, |
| "eval_loss": 0.2774485647678375, |
| "eval_runtime": 46.6449, |
| "eval_samples_per_second": 18.866, |
| "eval_steps_per_second": 18.866, |
| "step": 3000 |
| }, |
| { |
| "epoch": 1.8383561643835615, |
| "grad_norm": 0.35732510685920715, |
| "learning_rate": 3.0183909630321865e-06, |
| "loss": 0.07381275296211243, |
| "step": 3020 |
| }, |
| { |
| "epoch": 1.8505327245053271, |
| "grad_norm": 0.3167949914932251, |
| "learning_rate": 2.9938651089442184e-06, |
| "loss": 0.07289664745330811, |
| "step": 3040 |
| }, |
| { |
| "epoch": 1.8627092846270927, |
| "grad_norm": 0.4893674850463867, |
| "learning_rate": 2.969289692453773e-06, |
| "loss": 0.07124295830726624, |
| "step": 3060 |
| }, |
| { |
| "epoch": 1.8748858447488583, |
| "grad_norm": 0.3017306327819824, |
| "learning_rate": 2.944667179855109e-06, |
| "loss": 0.08125877976417542, |
| "step": 3080 |
| }, |
| { |
| "epoch": 1.887062404870624, |
| "grad_norm": 0.3442900776863098, |
| "learning_rate": 2.920000042168871e-06, |
| "loss": 0.0724608838558197, |
| "step": 3100 |
| }, |
| { |
| "epoch": 1.8992389649923895, |
| "grad_norm": 0.27901750802993774, |
| "learning_rate": 2.8952907548941057e-06, |
| "loss": 0.07104775309562683, |
| "step": 3120 |
| }, |
| { |
| "epoch": 1.9114155251141551, |
| "grad_norm": 0.35838621854782104, |
| "learning_rate": 2.8705417977598277e-06, |
| "loss": 0.0677955150604248, |
| "step": 3140 |
| }, |
| { |
| "epoch": 1.9235920852359207, |
| "grad_norm": 0.3615752160549164, |
| "learning_rate": 2.8457556544761687e-06, |
| "loss": 0.07164496779441834, |
| "step": 3160 |
| }, |
| { |
| "epoch": 1.9357686453576863, |
| "grad_norm": 0.5117827653884888, |
| "learning_rate": 2.8209348124851187e-06, |
| "loss": 0.071807599067688, |
| "step": 3180 |
| }, |
| { |
| "epoch": 1.947945205479452, |
| "grad_norm": 0.36082401871681213, |
| "learning_rate": 2.7960817627108965e-06, |
| "loss": 0.095755535364151, |
| "step": 3200 |
| }, |
| { |
| "epoch": 1.9601217656012175, |
| "grad_norm": 0.27905145287513733, |
| "learning_rate": 2.77119899930997e-06, |
| "loss": 0.07055851817131042, |
| "step": 3220 |
| }, |
| { |
| "epoch": 1.9722983257229831, |
| "grad_norm": 0.5642575621604919, |
| "learning_rate": 2.7462890194207513e-06, |
| "loss": 0.07278798818588257, |
| "step": 3240 |
| }, |
| { |
| "epoch": 1.9844748858447487, |
| "grad_norm": 0.2286670207977295, |
| "learning_rate": 2.7213543229129956e-06, |
| "loss": 0.07153088450431824, |
| "step": 3260 |
| }, |
| { |
| "epoch": 1.9966514459665143, |
| "grad_norm": 0.3228706121444702, |
| "learning_rate": 2.6963974121369242e-06, |
| "loss": 0.07440360188484192, |
| "step": 3280 |
| }, |
| { |
| "epoch": 2.008523592085236, |
| "grad_norm": 0.21613839268684387, |
| "learning_rate": 2.671420791672093e-06, |
| "loss": 0.0517767608165741, |
| "step": 3300 |
| }, |
| { |
| "epoch": 2.0207001522070014, |
| "grad_norm": 0.19271881878376007, |
| "learning_rate": 2.646426968076052e-06, |
| "loss": 0.03812239170074463, |
| "step": 3320 |
| }, |
| { |
| "epoch": 2.032876712328767, |
| "grad_norm": 0.12863056361675262, |
| "learning_rate": 2.6214184496327865e-06, |
| "loss": 0.04107579588890076, |
| "step": 3340 |
| }, |
| { |
| "epoch": 2.0450532724505326, |
| "grad_norm": 0.26679477095603943, |
| "learning_rate": 2.5963977461010022e-06, |
| "loss": 0.04673115909099579, |
| "step": 3360 |
| }, |
| { |
| "epoch": 2.057229832572298, |
| "grad_norm": 0.2743483781814575, |
| "learning_rate": 2.5713673684622524e-06, |
| "loss": 0.03674449622631073, |
| "step": 3380 |
| }, |
| { |
| "epoch": 2.069406392694064, |
| "grad_norm": 0.16999909281730652, |
| "learning_rate": 2.546329828668949e-06, |
| "loss": 0.03422380387783051, |
| "step": 3400 |
| }, |
| { |
| "epoch": 2.0815829528158294, |
| "grad_norm": 0.2931291460990906, |
| "learning_rate": 2.5212876393922657e-06, |
| "loss": 0.035878732800483704, |
| "step": 3420 |
| }, |
| { |
| "epoch": 2.093759512937595, |
| "grad_norm": 0.34781521558761597, |
| "learning_rate": 2.496243313769986e-06, |
| "loss": 0.03577531576156616, |
| "step": 3440 |
| }, |
| { |
| "epoch": 2.1059360730593606, |
| "grad_norm": 0.472351998090744, |
| "learning_rate": 2.471199365154283e-06, |
| "loss": 0.04281675517559051, |
| "step": 3460 |
| }, |
| { |
| "epoch": 2.118112633181126, |
| "grad_norm": 0.38320988416671753, |
| "learning_rate": 2.4461583068595014e-06, |
| "loss": 0.042955422401428224, |
| "step": 3480 |
| }, |
| { |
| "epoch": 2.130289193302892, |
| "grad_norm": 0.12197626382112503, |
| "learning_rate": 2.421122651909918e-06, |
| "loss": 0.04432957172393799, |
| "step": 3500 |
| }, |
| { |
| "epoch": 2.130289193302892, |
| "eval_loss": 0.3123805522918701, |
| "eval_runtime": 46.5258, |
| "eval_samples_per_second": 18.914, |
| "eval_steps_per_second": 18.914, |
| "step": 3500 |
| }, |
| { |
| "epoch": 2.1424657534246574, |
| "grad_norm": 0.21291697025299072, |
| "learning_rate": 2.3960949127875556e-06, |
| "loss": 0.03356837034225464, |
| "step": 3520 |
| }, |
| { |
| "epoch": 2.154642313546423, |
| "grad_norm": 0.44130077958106995, |
| "learning_rate": 2.371077601180031e-06, |
| "loss": 0.036935809254646304, |
| "step": 3540 |
| }, |
| { |
| "epoch": 2.1668188736681886, |
| "grad_norm": 0.49069875478744507, |
| "learning_rate": 2.3460732277284994e-06, |
| "loss": 0.0395690768957138, |
| "step": 3560 |
| }, |
| { |
| "epoch": 2.178995433789954, |
| "grad_norm": 0.2824937701225281, |
| "learning_rate": 2.321084301775689e-06, |
| "loss": 0.044693085551261905, |
| "step": 3580 |
| }, |
| { |
| "epoch": 2.19117199391172, |
| "grad_norm": 0.35114097595214844, |
| "learning_rate": 2.29611333111408e-06, |
| "loss": 0.03243565857410431, |
| "step": 3600 |
| }, |
| { |
| "epoch": 2.2033485540334854, |
| "grad_norm": 0.47931790351867676, |
| "learning_rate": 2.271162821734225e-06, |
| "loss": 0.04325798749923706, |
| "step": 3620 |
| }, |
| { |
| "epoch": 2.215525114155251, |
| "grad_norm": 0.12716218829154968, |
| "learning_rate": 2.2462352775732653e-06, |
| "loss": 0.03856868743896484, |
| "step": 3640 |
| }, |
| { |
| "epoch": 2.2277016742770166, |
| "grad_norm": 0.3522437512874603, |
| "learning_rate": 2.221333200263637e-06, |
| "loss": 0.041602414846420285, |
| "step": 3660 |
| }, |
| { |
| "epoch": 2.239878234398782, |
| "grad_norm": 0.234590083360672, |
| "learning_rate": 2.1964590888820233e-06, |
| "loss": 0.04286134541034699, |
| "step": 3680 |
| }, |
| { |
| "epoch": 2.252054794520548, |
| "grad_norm": 0.2972453534603119, |
| "learning_rate": 2.1716154396985526e-06, |
| "loss": 0.041756758093833925, |
| "step": 3700 |
| }, |
| { |
| "epoch": 2.2642313546423134, |
| "grad_norm": 0.3236760199069977, |
| "learning_rate": 2.1468047459262882e-06, |
| "loss": 0.0359495222568512, |
| "step": 3720 |
| }, |
| { |
| "epoch": 2.276407914764079, |
| "grad_norm": 0.32255831360816956, |
| "learning_rate": 2.12202949747101e-06, |
| "loss": 0.04322676360607147, |
| "step": 3740 |
| }, |
| { |
| "epoch": 2.2885844748858446, |
| "grad_norm": 0.2753404378890991, |
| "learning_rate": 2.0972921806813468e-06, |
| "loss": 0.04191597998142242, |
| "step": 3760 |
| }, |
| { |
| "epoch": 2.30076103500761, |
| "grad_norm": 0.23182159662246704, |
| "learning_rate": 2.072595278099247e-06, |
| "loss": 0.041278204321861266, |
| "step": 3780 |
| }, |
| { |
| "epoch": 2.312937595129376, |
| "grad_norm": 0.25987961888313293, |
| "learning_rate": 2.047941268210849e-06, |
| "loss": 0.04312986135482788, |
| "step": 3800 |
| }, |
| { |
| "epoch": 2.3251141552511414, |
| "grad_norm": 0.3683331310749054, |
| "learning_rate": 2.0233326251977426e-06, |
| "loss": 0.04236046075820923, |
| "step": 3820 |
| }, |
| { |
| "epoch": 2.337290715372907, |
| "grad_norm": 0.2520082890987396, |
| "learning_rate": 1.9987718186886724e-06, |
| "loss": 0.04433901011943817, |
| "step": 3840 |
| }, |
| { |
| "epoch": 2.3494672754946726, |
| "grad_norm": 0.16200299561023712, |
| "learning_rate": 1.9742613135116986e-06, |
| "loss": 0.04127628207206726, |
| "step": 3860 |
| }, |
| { |
| "epoch": 2.361643835616438, |
| "grad_norm": 0.37064701318740845, |
| "learning_rate": 1.949803569446828e-06, |
| "loss": 0.04586326479911804, |
| "step": 3880 |
| }, |
| { |
| "epoch": 2.373820395738204, |
| "grad_norm": 0.416020005941391, |
| "learning_rate": 1.925401040979171e-06, |
| "loss": 0.03624185025691986, |
| "step": 3900 |
| }, |
| { |
| "epoch": 2.3859969558599694, |
| "grad_norm": 0.17131179571151733, |
| "learning_rate": 1.9010561770526076e-06, |
| "loss": 0.035064518451690674, |
| "step": 3920 |
| }, |
| { |
| "epoch": 2.398173515981735, |
| "grad_norm": 0.5188226103782654, |
| "learning_rate": 1.8767714208240312e-06, |
| "loss": 0.042050021886825564, |
| "step": 3940 |
| }, |
| { |
| "epoch": 2.4103500761035006, |
| "grad_norm": 0.3398009240627289, |
| "learning_rate": 1.852549209418154e-06, |
| "loss": 0.038166466355323794, |
| "step": 3960 |
| }, |
| { |
| "epoch": 2.422526636225266, |
| "grad_norm": 0.2541758418083191, |
| "learning_rate": 1.8283919736829332e-06, |
| "loss": 0.040885674953460696, |
| "step": 3980 |
| }, |
| { |
| "epoch": 2.434703196347032, |
| "grad_norm": 0.4074256122112274, |
| "learning_rate": 1.804302137945614e-06, |
| "loss": 0.040162667632102966, |
| "step": 4000 |
| }, |
| { |
| "epoch": 2.434703196347032, |
| "eval_loss": 0.31583163142204285, |
| "eval_runtime": 47.319, |
| "eval_samples_per_second": 18.597, |
| "eval_steps_per_second": 18.597, |
| "step": 4000 |
| }, |
| { |
| "epoch": 2.4468797564687974, |
| "grad_norm": 0.2724802494049072, |
| "learning_rate": 1.7802821197694426e-06, |
| "loss": 0.04170995056629181, |
| "step": 4020 |
| }, |
| { |
| "epoch": 2.459056316590563, |
| "grad_norm": 0.21376508474349976, |
| "learning_rate": 1.7563343297110375e-06, |
| "loss": 0.03834344446659088, |
| "step": 4040 |
| }, |
| { |
| "epoch": 2.4712328767123286, |
| "grad_norm": 0.2933412492275238, |
| "learning_rate": 1.732461171078486e-06, |
| "loss": 0.03928310573101044, |
| "step": 4060 |
| }, |
| { |
| "epoch": 2.483409436834094, |
| "grad_norm": 0.46805083751678467, |
| "learning_rate": 1.7086650396901489e-06, |
| "loss": 0.03358933925628662, |
| "step": 4080 |
| }, |
| { |
| "epoch": 2.49558599695586, |
| "grad_norm": 0.45667552947998047, |
| "learning_rate": 1.6849483236342322e-06, |
| "loss": 0.03547535240650177, |
| "step": 4100 |
| }, |
| { |
| "epoch": 2.5077625570776254, |
| "grad_norm": 0.24512450397014618, |
| "learning_rate": 1.6613134030291217e-06, |
| "loss": 0.03600102663040161, |
| "step": 4120 |
| }, |
| { |
| "epoch": 2.519939117199391, |
| "grad_norm": 0.20636337995529175, |
| "learning_rate": 1.6377626497845278e-06, |
| "loss": 0.04347077012062073, |
| "step": 4140 |
| }, |
| { |
| "epoch": 2.5321156773211566, |
| "grad_norm": 0.4584953486919403, |
| "learning_rate": 1.6142984273634505e-06, |
| "loss": 0.02908192276954651, |
| "step": 4160 |
| }, |
| { |
| "epoch": 2.544292237442922, |
| "grad_norm": 0.26193487644195557, |
| "learning_rate": 1.5909230905449846e-06, |
| "loss": 0.03611198365688324, |
| "step": 4180 |
| }, |
| { |
| "epoch": 2.556468797564688, |
| "grad_norm": 0.20813828706741333, |
| "learning_rate": 1.567638985188012e-06, |
| "loss": 0.03758668601512909, |
| "step": 4200 |
| }, |
| { |
| "epoch": 2.5686453576864534, |
| "grad_norm": 0.3395022749900818, |
| "learning_rate": 1.544448447995773e-06, |
| "loss": 0.033633843064308167, |
| "step": 4220 |
| }, |
| { |
| "epoch": 2.580821917808219, |
| "grad_norm": 0.1472434103488922, |
| "learning_rate": 1.52135380628137e-06, |
| "loss": 0.036797890067100526, |
| "step": 4240 |
| }, |
| { |
| "epoch": 2.5929984779299846, |
| "grad_norm": 0.5788060426712036, |
| "learning_rate": 1.498357377734201e-06, |
| "loss": 0.039166563749313356, |
| "step": 4260 |
| }, |
| { |
| "epoch": 2.60517503805175, |
| "grad_norm": 0.7623679637908936, |
| "learning_rate": 1.4754614701873703e-06, |
| "loss": 0.03717599511146545, |
| "step": 4280 |
| }, |
| { |
| "epoch": 2.6173515981735163, |
| "grad_norm": 0.16205403208732605, |
| "learning_rate": 1.4526683813860792e-06, |
| "loss": 0.03962793946266174, |
| "step": 4300 |
| }, |
| { |
| "epoch": 2.6295281582952814, |
| "grad_norm": 0.11986076086759567, |
| "learning_rate": 1.4299803987570396e-06, |
| "loss": 0.035475924611091614, |
| "step": 4320 |
| }, |
| { |
| "epoch": 2.6417047184170475, |
| "grad_norm": 0.15573006868362427, |
| "learning_rate": 1.4073997991789078e-06, |
| "loss": 0.03256964683532715, |
| "step": 4340 |
| }, |
| { |
| "epoch": 2.6538812785388126, |
| "grad_norm": 0.25151512026786804, |
| "learning_rate": 1.384928848753792e-06, |
| "loss": 0.03712306022644043, |
| "step": 4360 |
| }, |
| { |
| "epoch": 2.6660578386605787, |
| "grad_norm": 0.20408153533935547, |
| "learning_rate": 1.3625698025798322e-06, |
| "loss": 0.041410398483276364, |
| "step": 4380 |
| }, |
| { |
| "epoch": 2.678234398782344, |
| "grad_norm": 0.3156696856021881, |
| "learning_rate": 1.3403249045248907e-06, |
| "loss": 0.03158504366874695, |
| "step": 4400 |
| }, |
| { |
| "epoch": 2.69041095890411, |
| "grad_norm": 0.3835665285587311, |
| "learning_rate": 1.3181963870013604e-06, |
| "loss": 0.03525224924087524, |
| "step": 4420 |
| }, |
| { |
| "epoch": 2.702587519025875, |
| "grad_norm": 0.45423486828804016, |
| "learning_rate": 1.2961864707421345e-06, |
| "loss": 0.03239959478378296, |
| "step": 4440 |
| }, |
| { |
| "epoch": 2.714764079147641, |
| "grad_norm": 0.15982766449451447, |
| "learning_rate": 1.2742973645777394e-06, |
| "loss": 0.031032082438468934, |
| "step": 4460 |
| }, |
| { |
| "epoch": 2.726940639269406, |
| "grad_norm": 0.2770426869392395, |
| "learning_rate": 1.252531265214662e-06, |
| "loss": 0.030566230416297913, |
| "step": 4480 |
| }, |
| { |
| "epoch": 2.7391171993911723, |
| "grad_norm": 0.3693839907646179, |
| "learning_rate": 1.2308903570149048e-06, |
| "loss": 0.041362547874450685, |
| "step": 4500 |
| }, |
| { |
| "epoch": 2.7391171993911723, |
| "eval_loss": 0.3391737937927246, |
| "eval_runtime": 46.9306, |
| "eval_samples_per_second": 18.751, |
| "eval_steps_per_second": 18.751, |
| "step": 4500 |
| }, |
| { |
| "epoch": 2.7512937595129374, |
| "grad_norm": 0.3229275643825531, |
| "learning_rate": 1.2093768117767613e-06, |
| "loss": 0.0388390064239502, |
| "step": 4520 |
| }, |
| { |
| "epoch": 2.7634703196347035, |
| "grad_norm": 0.6786078214645386, |
| "learning_rate": 1.1879927885168733e-06, |
| "loss": 0.032555675506591795, |
| "step": 4540 |
| }, |
| { |
| "epoch": 2.7756468797564686, |
| "grad_norm": 0.32371029257774353, |
| "learning_rate": 1.1667404332535504e-06, |
| "loss": 0.03606459796428681, |
| "step": 4560 |
| }, |
| { |
| "epoch": 2.7878234398782347, |
| "grad_norm": 0.44066882133483887, |
| "learning_rate": 1.1456218787914128e-06, |
| "loss": 0.032086309790611264, |
| "step": 4580 |
| }, |
| { |
| "epoch": 2.8, |
| "grad_norm": 0.5005165338516235, |
| "learning_rate": 1.1246392445073438e-06, |
| "loss": 0.033362787961959836, |
| "step": 4600 |
| }, |
| { |
| "epoch": 2.812176560121766, |
| "grad_norm": 0.22586822509765625, |
| "learning_rate": 1.1037946361378027e-06, |
| "loss": 0.03638745844364166, |
| "step": 4620 |
| }, |
| { |
| "epoch": 2.824353120243531, |
| "grad_norm": 0.2905796766281128, |
| "learning_rate": 1.0830901455674977e-06, |
| "loss": 0.030933958292007447, |
| "step": 4640 |
| }, |
| { |
| "epoch": 2.836529680365297, |
| "grad_norm": 0.10796497762203217, |
| "learning_rate": 1.0625278506194538e-06, |
| "loss": 0.02879139482975006, |
| "step": 4660 |
| }, |
| { |
| "epoch": 2.8487062404870622, |
| "grad_norm": 0.2545916438102722, |
| "learning_rate": 1.04210981484649e-06, |
| "loss": 0.03345020413398743, |
| "step": 4680 |
| }, |
| { |
| "epoch": 2.8608828006088283, |
| "grad_norm": 0.2986568808555603, |
| "learning_rate": 1.0218380873241314e-06, |
| "loss": 0.02593054175376892, |
| "step": 4700 |
| }, |
| { |
| "epoch": 2.8730593607305934, |
| "grad_norm": 0.1755756139755249, |
| "learning_rate": 1.0017147024449674e-06, |
| "loss": 0.03906567096710205, |
| "step": 4720 |
| }, |
| { |
| "epoch": 2.8852359208523595, |
| "grad_norm": 0.18288888037204742, |
| "learning_rate": 9.81741679714493e-07, |
| "loss": 0.03371626436710358, |
| "step": 4740 |
| }, |
| { |
| "epoch": 2.8974124809741246, |
| "grad_norm": 0.368429958820343, |
| "learning_rate": 9.619210235484333e-07, |
| "loss": 0.03090968132019043, |
| "step": 4760 |
| }, |
| { |
| "epoch": 2.9095890410958907, |
| "grad_norm": 0.17118144035339355, |
| "learning_rate": 9.422547230715931e-07, |
| "loss": 0.0322105199098587, |
| "step": 4780 |
| }, |
| { |
| "epoch": 2.921765601217656, |
| "grad_norm": 0.41911277174949646, |
| "learning_rate": 9.227447519182353e-07, |
| "loss": 0.035210177302360535, |
| "step": 4800 |
| }, |
| { |
| "epoch": 2.933942161339422, |
| "grad_norm": 0.3521968722343445, |
| "learning_rate": 9.033930680340097e-07, |
| "loss": 0.026842504739761353, |
| "step": 4820 |
| }, |
| { |
| "epoch": 2.946118721461187, |
| "grad_norm": 0.20812013745307922, |
| "learning_rate": 8.842016134794682e-07, |
| "loss": 0.03439584076404571, |
| "step": 4840 |
| }, |
| { |
| "epoch": 2.958295281582953, |
| "grad_norm": 0.2796875834465027, |
| "learning_rate": 8.651723142351603e-07, |
| "loss": 0.04011322855949402, |
| "step": 4860 |
| }, |
| { |
| "epoch": 2.9704718417047182, |
| "grad_norm": 0.20960687100887299, |
| "learning_rate": 8.463070800083562e-07, |
| "loss": 0.03800423145294189, |
| "step": 4880 |
| }, |
| { |
| "epoch": 2.9826484018264843, |
| "grad_norm": 0.2586495876312256, |
| "learning_rate": 8.276078040413879e-07, |
| "loss": 0.03839131891727447, |
| "step": 4900 |
| }, |
| { |
| "epoch": 2.9948249619482494, |
| "grad_norm": 0.37137141823768616, |
| "learning_rate": 8.090763629216589e-07, |
| "loss": 0.02721840739250183, |
| "step": 4920 |
| }, |
| { |
| "epoch": 3.006697108066971, |
| "grad_norm": 0.3677407503128052, |
| "learning_rate": 7.907146163933102e-07, |
| "loss": 0.023991990089416503, |
| "step": 4940 |
| }, |
| { |
| "epoch": 3.0188736681887365, |
| "grad_norm": 0.11811063438653946, |
| "learning_rate": 7.725244071705871e-07, |
| "loss": 0.01451514959335327, |
| "step": 4960 |
| }, |
| { |
| "epoch": 3.031050228310502, |
| "grad_norm": 0.3449067771434784, |
| "learning_rate": 7.545075607529104e-07, |
| "loss": 0.014327619969844819, |
| "step": 4980 |
| }, |
| { |
| "epoch": 3.0432267884322677, |
| "grad_norm": 0.3237353265285492, |
| "learning_rate": 7.366658852416788e-07, |
| "loss": 0.017832010984420776, |
| "step": 5000 |
| }, |
| { |
| "epoch": 3.0432267884322677, |
| "eval_loss": 0.39115142822265625, |
| "eval_runtime": 47.6431, |
| "eval_samples_per_second": 18.471, |
| "eval_steps_per_second": 18.471, |
| "step": 5000 |
| }, |
| { |
| "epoch": 3.0554033485540333, |
| "grad_norm": 0.2226281613111496, |
| "learning_rate": 7.190011711588101e-07, |
| "loss": 0.011674411594867706, |
| "step": 5020 |
| }, |
| { |
| "epoch": 3.067579908675799, |
| "grad_norm": 0.08729376643896103, |
| "learning_rate": 7.015151912670562e-07, |
| "loss": 0.013690856099128724, |
| "step": 5040 |
| }, |
| { |
| "epoch": 3.0797564687975645, |
| "grad_norm": 0.2745465636253357, |
| "learning_rate": 6.842097003920903e-07, |
| "loss": 0.011978642642498016, |
| "step": 5060 |
| }, |
| { |
| "epoch": 3.09193302891933, |
| "grad_norm": 0.06905842572450638, |
| "learning_rate": 6.67086435246406e-07, |
| "loss": 0.013893941044807434, |
| "step": 5080 |
| }, |
| { |
| "epoch": 3.1041095890410957, |
| "grad_norm": 0.07840315997600555, |
| "learning_rate": 6.501471142550194e-07, |
| "loss": 0.009910025447607041, |
| "step": 5100 |
| }, |
| { |
| "epoch": 3.1162861491628613, |
| "grad_norm": 0.19672124087810516, |
| "learning_rate": 6.333934373830222e-07, |
| "loss": 0.008863755315542222, |
| "step": 5120 |
| }, |
| { |
| "epoch": 3.128462709284627, |
| "grad_norm": 0.37645605206489563, |
| "learning_rate": 6.168270859649761e-07, |
| "loss": 0.010502541810274124, |
| "step": 5140 |
| }, |
| { |
| "epoch": 3.1406392694063925, |
| "grad_norm": 0.2069159746170044, |
| "learning_rate": 6.004497225361786e-07, |
| "loss": 0.012096930295228958, |
| "step": 5160 |
| }, |
| { |
| "epoch": 3.1528158295281585, |
| "grad_norm": 0.2584017217159271, |
| "learning_rate": 5.842629906658226e-07, |
| "loss": 0.013278065621852875, |
| "step": 5180 |
| }, |
| { |
| "epoch": 3.1649923896499237, |
| "grad_norm": 0.2050527036190033, |
| "learning_rate": 5.682685147920481e-07, |
| "loss": 0.013548998534679413, |
| "step": 5200 |
| }, |
| { |
| "epoch": 3.1771689497716897, |
| "grad_norm": 0.13838107883930206, |
| "learning_rate": 5.524679000589256e-07, |
| "loss": 0.013736458122730255, |
| "step": 5220 |
| }, |
| { |
| "epoch": 3.189345509893455, |
| "grad_norm": 0.06378225982189178, |
| "learning_rate": 5.36862732155366e-07, |
| "loss": 0.013177134096622467, |
| "step": 5240 |
| }, |
| { |
| "epoch": 3.201522070015221, |
| "grad_norm": 0.27431613206863403, |
| "learning_rate": 5.214545771559879e-07, |
| "loss": 0.011971819400787353, |
| "step": 5260 |
| }, |
| { |
| "epoch": 3.213698630136986, |
| "grad_norm": 0.529901921749115, |
| "learning_rate": 5.062449813639528e-07, |
| "loss": 0.014422819018363953, |
| "step": 5280 |
| }, |
| { |
| "epoch": 3.225875190258752, |
| "grad_norm": 0.19417761266231537, |
| "learning_rate": 4.912354711557856e-07, |
| "loss": 0.010663678497076034, |
| "step": 5300 |
| }, |
| { |
| "epoch": 3.2380517503805173, |
| "grad_norm": 0.044735077768564224, |
| "learning_rate": 4.764275528281892e-07, |
| "loss": 0.011400717496871948, |
| "step": 5320 |
| }, |
| { |
| "epoch": 3.2502283105022833, |
| "grad_norm": 0.057179443538188934, |
| "learning_rate": 4.6182271244688355e-07, |
| "loss": 0.008456526696681977, |
| "step": 5340 |
| }, |
| { |
| "epoch": 3.2624048706240485, |
| "grad_norm": 0.10396906733512878, |
| "learning_rate": 4.4742241569746407e-07, |
| "loss": 0.014539115130901337, |
| "step": 5360 |
| }, |
| { |
| "epoch": 3.2745814307458145, |
| "grad_norm": 0.32904428243637085, |
| "learning_rate": 4.332281077383177e-07, |
| "loss": 0.017625690996646882, |
| "step": 5380 |
| }, |
| { |
| "epoch": 3.2867579908675797, |
| "grad_norm": 0.20823979377746582, |
| "learning_rate": 4.1924121305558563e-07, |
| "loss": 0.007641100138425827, |
| "step": 5400 |
| }, |
| { |
| "epoch": 3.2989345509893457, |
| "grad_norm": 0.25470009446144104, |
| "learning_rate": 4.054631353202121e-07, |
| "loss": 0.011799700558185577, |
| "step": 5420 |
| }, |
| { |
| "epoch": 3.311111111111111, |
| "grad_norm": 0.3968588709831238, |
| "learning_rate": 3.9189525724707634e-07, |
| "loss": 0.011455408483743667, |
| "step": 5440 |
| }, |
| { |
| "epoch": 3.323287671232877, |
| "grad_norm": 0.10818332433700562, |
| "learning_rate": 3.785389404562259e-07, |
| "loss": 0.012499115616083144, |
| "step": 5460 |
| }, |
| { |
| "epoch": 3.335464231354642, |
| "grad_norm": 0.1818460375070572, |
| "learning_rate": 3.653955253362351e-07, |
| "loss": 0.01148865669965744, |
| "step": 5480 |
| }, |
| { |
| "epoch": 3.347640791476408, |
| "grad_norm": 0.3504088521003723, |
| "learning_rate": 3.5246633090968205e-07, |
| "loss": 0.012819178402423859, |
| "step": 5500 |
| }, |
| { |
| "epoch": 3.347640791476408, |
| "eval_loss": 0.43404534459114075, |
| "eval_runtime": 46.4882, |
| "eval_samples_per_second": 18.93, |
| "eval_steps_per_second": 18.93, |
| "step": 5500 |
| }, |
| { |
| "epoch": 3.3598173515981733, |
| "grad_norm": 0.4551874101161957, |
| "learning_rate": 3.397526547007832e-07, |
| "loss": 0.013325585424900055, |
| "step": 5520 |
| }, |
| { |
| "epoch": 3.3719939117199393, |
| "grad_norm": 0.35187825560569763, |
| "learning_rate": 3.2725577260517396e-07, |
| "loss": 0.011712662875652313, |
| "step": 5540 |
| }, |
| { |
| "epoch": 3.3841704718417045, |
| "grad_norm": 0.6071529984474182, |
| "learning_rate": 3.14976938761867e-07, |
| "loss": 0.01580573171377182, |
| "step": 5560 |
| }, |
| { |
| "epoch": 3.3963470319634705, |
| "grad_norm": 0.18844422698020935, |
| "learning_rate": 3.029173854273909e-07, |
| "loss": 0.012312603741884231, |
| "step": 5580 |
| }, |
| { |
| "epoch": 3.4085235920852357, |
| "grad_norm": 0.13131535053253174, |
| "learning_rate": 2.910783228521269e-07, |
| "loss": 0.011797953397035599, |
| "step": 5600 |
| }, |
| { |
| "epoch": 3.4207001522070017, |
| "grad_norm": 0.4402364492416382, |
| "learning_rate": 2.794609391588504e-07, |
| "loss": 0.012182456254959107, |
| "step": 5620 |
| }, |
| { |
| "epoch": 3.432876712328767, |
| "grad_norm": 0.3497592508792877, |
| "learning_rate": 2.6806640022349897e-07, |
| "loss": 0.013599888980388641, |
| "step": 5640 |
| }, |
| { |
| "epoch": 3.445053272450533, |
| "grad_norm": 0.2316354215145111, |
| "learning_rate": 2.5689584955816497e-07, |
| "loss": 0.009272868931293487, |
| "step": 5660 |
| }, |
| { |
| "epoch": 3.457229832572298, |
| "grad_norm": 0.3858301341533661, |
| "learning_rate": 2.459504081963421e-07, |
| "loss": 0.008165979385375976, |
| "step": 5680 |
| }, |
| { |
| "epoch": 3.469406392694064, |
| "grad_norm": 0.14734333753585815, |
| "learning_rate": 2.3523117458041865e-07, |
| "loss": 0.009182130545377731, |
| "step": 5700 |
| }, |
| { |
| "epoch": 3.4815829528158293, |
| "grad_norm": 0.03280401974916458, |
| "learning_rate": 2.2473922445144485e-07, |
| "loss": 0.0107998326420784, |
| "step": 5720 |
| }, |
| { |
| "epoch": 3.4937595129375953, |
| "grad_norm": 0.1505511999130249, |
| "learning_rate": 2.144756107411733e-07, |
| "loss": 0.014469687640666962, |
| "step": 5740 |
| }, |
| { |
| "epoch": 3.5059360730593605, |
| "grad_norm": 0.2366904318332672, |
| "learning_rate": 2.0444136346639333e-07, |
| "loss": 0.0121701680123806, |
| "step": 5760 |
| }, |
| { |
| "epoch": 3.5181126331811265, |
| "grad_norm": 0.1468425989151001, |
| "learning_rate": 1.9463748962556096e-07, |
| "loss": 0.014668506383895875, |
| "step": 5780 |
| }, |
| { |
| "epoch": 3.5302891933028917, |
| "grad_norm": 0.14534050226211548, |
| "learning_rate": 1.8506497309773885e-07, |
| "loss": 0.010488402843475342, |
| "step": 5800 |
| }, |
| { |
| "epoch": 3.5424657534246577, |
| "grad_norm": 0.15501493215560913, |
| "learning_rate": 1.7572477454386257e-07, |
| "loss": 0.010667071491479874, |
| "step": 5820 |
| }, |
| { |
| "epoch": 3.554642313546423, |
| "grad_norm": 0.26535800099372864, |
| "learning_rate": 1.6661783131032726e-07, |
| "loss": 0.011079683899879456, |
| "step": 5840 |
| }, |
| { |
| "epoch": 3.566818873668189, |
| "grad_norm": 0.24390950798988342, |
| "learning_rate": 1.5774505733492263e-07, |
| "loss": 0.009308797866106033, |
| "step": 5860 |
| }, |
| { |
| "epoch": 3.578995433789954, |
| "grad_norm": 0.3409421443939209, |
| "learning_rate": 1.49107343055111e-07, |
| "loss": 0.012319787591695785, |
| "step": 5880 |
| }, |
| { |
| "epoch": 3.59117199391172, |
| "grad_norm": 0.4800300896167755, |
| "learning_rate": 1.407055553186701e-07, |
| "loss": 0.00843576118350029, |
| "step": 5900 |
| }, |
| { |
| "epoch": 3.6033485540334853, |
| "grad_norm": 0.11663182079792023, |
| "learning_rate": 1.3254053729669564e-07, |
| "loss": 0.00938587412238121, |
| "step": 5920 |
| }, |
| { |
| "epoch": 3.6155251141552514, |
| "grad_norm": 0.29512378573417664, |
| "learning_rate": 1.2461310839898656e-07, |
| "loss": 0.011934128403663636, |
| "step": 5940 |
| }, |
| { |
| "epoch": 3.6277016742770165, |
| "grad_norm": 0.2641650140285492, |
| "learning_rate": 1.169240641918104e-07, |
| "loss": 0.013170333206653595, |
| "step": 5960 |
| }, |
| { |
| "epoch": 3.6398782343987826, |
| "grad_norm": 0.47704726457595825, |
| "learning_rate": 1.0947417631806539e-07, |
| "loss": 0.014534834027290344, |
| "step": 5980 |
| }, |
| { |
| "epoch": 3.6520547945205477, |
| "grad_norm": 0.10114685446023941, |
| "learning_rate": 1.0226419241983865e-07, |
| "loss": 0.011021688580513, |
| "step": 6000 |
| }, |
| { |
| "epoch": 3.6520547945205477, |
| "eval_loss": 0.44063475728034973, |
| "eval_runtime": 46.1734, |
| "eval_samples_per_second": 19.059, |
| "eval_steps_per_second": 19.059, |
| "step": 6000 |
| }, |
| { |
| "epoch": 3.6642313546423138, |
| "grad_norm": 0.2619183659553528, |
| "learning_rate": 9.529483606337902e-08, |
| "loss": 0.010764393210411071, |
| "step": 6020 |
| }, |
| { |
| "epoch": 3.676407914764079, |
| "grad_norm": 0.05733129009604454, |
| "learning_rate": 8.856680666647882e-08, |
| "loss": 0.012128306180238723, |
| "step": 6040 |
| }, |
| { |
| "epoch": 3.688584474885845, |
| "grad_norm": 0.19483673572540283, |
| "learning_rate": 8.208077942828713e-08, |
| "loss": 0.011729901283979416, |
| "step": 6060 |
| }, |
| { |
| "epoch": 3.70076103500761, |
| "grad_norm": 0.2111903578042984, |
| "learning_rate": 7.58374052615457e-08, |
| "loss": 0.009119105339050294, |
| "step": 6080 |
| }, |
| { |
| "epoch": 3.712937595129376, |
| "grad_norm": 0.04995311424136162, |
| "learning_rate": 6.983731072726818e-08, |
| "loss": 0.017101363837718965, |
| "step": 6100 |
| }, |
| { |
| "epoch": 3.7251141552511413, |
| "grad_norm": 0.5839787125587463, |
| "learning_rate": 6.408109797186118e-08, |
| "loss": 0.012368235737085342, |
| "step": 6120 |
| }, |
| { |
| "epoch": 3.7372907153729074, |
| "grad_norm": 0.4685717523097992, |
| "learning_rate": 5.856934466669212e-08, |
| "loss": 0.008782628178596496, |
| "step": 6140 |
| }, |
| { |
| "epoch": 3.7494672754946725, |
| "grad_norm": 0.17204681038856506, |
| "learning_rate": 5.3302603950119994e-08, |
| "loss": 0.008994438499212266, |
| "step": 6160 |
| }, |
| { |
| "epoch": 3.7616438356164386, |
| "grad_norm": 0.07392167299985886, |
| "learning_rate": 4.8281404371981755e-08, |
| "loss": 0.011286454647779465, |
| "step": 6180 |
| }, |
| { |
| "epoch": 3.7738203957382037, |
| "grad_norm": 0.3728208541870117, |
| "learning_rate": 4.350624984055196e-08, |
| "loss": 0.011785905063152313, |
| "step": 6200 |
| }, |
| { |
| "epoch": 3.7859969558599698, |
| "grad_norm": 0.25468680262565613, |
| "learning_rate": 3.897761957196877e-08, |
| "loss": 0.013624191284179688, |
| "step": 6220 |
| }, |
| { |
| "epoch": 3.798173515981735, |
| "grad_norm": 0.09725204110145569, |
| "learning_rate": 3.469596804214548e-08, |
| "loss": 0.011700452119112015, |
| "step": 6240 |
| }, |
| { |
| "epoch": 3.810350076103501, |
| "grad_norm": 0.07126162946224213, |
| "learning_rate": 3.06617249411581e-08, |
| "loss": 0.011029987037181855, |
| "step": 6260 |
| }, |
| { |
| "epoch": 3.822526636225266, |
| "grad_norm": 0.08542267978191376, |
| "learning_rate": 2.687529513012488e-08, |
| "loss": 0.010965974628925323, |
| "step": 6280 |
| }, |
| { |
| "epoch": 3.834703196347032, |
| "grad_norm": 0.2627331018447876, |
| "learning_rate": 2.3337058600575722e-08, |
| "loss": 0.012378603965044022, |
| "step": 6300 |
| }, |
| { |
| "epoch": 3.8468797564687973, |
| "grad_norm": 0.19707690179347992, |
| "learning_rate": 2.0047370436317437e-08, |
| "loss": 0.011792077124118805, |
| "step": 6320 |
| }, |
| { |
| "epoch": 3.8590563165905634, |
| "grad_norm": 0.47547003626823425, |
| "learning_rate": 1.7006560777798608e-08, |
| "loss": 0.01145942509174347, |
| "step": 6340 |
| }, |
| { |
| "epoch": 3.8712328767123285, |
| "grad_norm": 0.3591565489768982, |
| "learning_rate": 1.421493478897945e-08, |
| "loss": 0.011088228970766067, |
| "step": 6360 |
| }, |
| { |
| "epoch": 3.8834094368340946, |
| "grad_norm": 0.20619548857212067, |
| "learning_rate": 1.1672772626704909e-08, |
| "loss": 0.010828402638435364, |
| "step": 6380 |
| }, |
| { |
| "epoch": 3.8955859969558597, |
| "grad_norm": 0.2822403311729431, |
| "learning_rate": 9.38032941258965e-09, |
| "loss": 0.01165580153465271, |
| "step": 6400 |
| }, |
| { |
| "epoch": 3.9077625570776258, |
| "grad_norm": 0.15682674944400787, |
| "learning_rate": 7.3378352074163215e-09, |
| "loss": 0.010783226788043975, |
| "step": 6420 |
| }, |
| { |
| "epoch": 3.919939117199391, |
| "grad_norm": 0.34253379702568054, |
| "learning_rate": 5.545494988045963e-09, |
| "loss": 0.011295531690120698, |
| "step": 6440 |
| }, |
| { |
| "epoch": 3.932115677321157, |
| "grad_norm": 0.27684664726257324, |
| "learning_rate": 4.003488626848073e-09, |
| "loss": 0.013613662123680115, |
| "step": 6460 |
| }, |
| { |
| "epoch": 3.944292237442922, |
| "grad_norm": 0.69688880443573, |
| "learning_rate": 2.7119708736486615e-09, |
| "loss": 0.011696261167526246, |
| "step": 6480 |
| }, |
| { |
| "epoch": 3.956468797564688, |
| "grad_norm": 0.2617769241333008, |
| "learning_rate": 1.6710713402015577e-09, |
| "loss": 0.010873865336179733, |
| "step": 6500 |
| }, |
| { |
| "epoch": 3.956468797564688, |
| "eval_loss": 0.441041499376297, |
| "eval_runtime": 46.1946, |
| "eval_samples_per_second": 19.05, |
| "eval_steps_per_second": 19.05, |
| "step": 6500 |
| }, |
| { |
| "epoch": 3.9686453576864533, |
| "grad_norm": 0.23670868575572968, |
| "learning_rate": 8.80894487179651e-10, |
| "loss": 0.012961818277835846, |
| "step": 6520 |
| }, |
| { |
| "epoch": 3.9808219178082194, |
| "grad_norm": 1.036125659942627, |
| "learning_rate": 3.4151961369188745e-10, |
| "loss": 0.01224210560321808, |
| "step": 6540 |
| }, |
| { |
| "epoch": 3.9929984779299845, |
| "grad_norm": 0.1431870311498642, |
| "learning_rate": 5.300084932574612e-11, |
| "loss": 0.010245455056428909, |
| "step": 6560 |
| }, |
| { |
| "epoch": 4.0, |
| "step": 6572, |
| "total_flos": 3.545203061907456e+17, |
| "train_loss": 0.08566030774240586, |
| "train_runtime": 13933.4985, |
| "train_samples_per_second": 3.772, |
| "train_steps_per_second": 0.472 |
| } |
| ], |
| "logging_steps": 20, |
| "max_steps": 6572, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 4, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 3.545203061907456e+17, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|