| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.0, | |
| "eval_steps": 500, | |
| "global_step": 1073, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.004659832246039142, | |
| "grad_norm": 1.8795702761938184, | |
| "learning_rate": 9.259259259259259e-07, | |
| "loss": 0.3874, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.009319664492078284, | |
| "grad_norm": 1.810960933535657, | |
| "learning_rate": 1.8518518518518519e-06, | |
| "loss": 0.3745, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.013979496738117428, | |
| "grad_norm": 0.689820489630518, | |
| "learning_rate": 2.7777777777777783e-06, | |
| "loss": 0.3574, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.01863932898415657, | |
| "grad_norm": 0.23671360073050024, | |
| "learning_rate": 3.7037037037037037e-06, | |
| "loss": 0.3465, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.023299161230195712, | |
| "grad_norm": 0.5035003762355204, | |
| "learning_rate": 4.62962962962963e-06, | |
| "loss": 0.3273, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.027958993476234855, | |
| "grad_norm": 0.2861970991097442, | |
| "learning_rate": 5.555555555555557e-06, | |
| "loss": 0.3177, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.032618825722273995, | |
| "grad_norm": 0.20083968480929654, | |
| "learning_rate": 6.481481481481482e-06, | |
| "loss": 0.3161, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.03727865796831314, | |
| "grad_norm": 0.2135927219205515, | |
| "learning_rate": 7.4074074074074075e-06, | |
| "loss": 0.3186, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.04193849021435228, | |
| "grad_norm": 0.20149881773912096, | |
| "learning_rate": 8.333333333333334e-06, | |
| "loss": 0.3185, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.046598322460391424, | |
| "grad_norm": 0.17751016241192105, | |
| "learning_rate": 9.25925925925926e-06, | |
| "loss": 0.3248, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.05125815470643057, | |
| "grad_norm": 0.18396942525421686, | |
| "learning_rate": 1.0185185185185186e-05, | |
| "loss": 0.3175, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.05591798695246971, | |
| "grad_norm": 0.17229279499140202, | |
| "learning_rate": 1.1111111111111113e-05, | |
| "loss": 0.3147, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.06057781919850885, | |
| "grad_norm": 0.1737383934121573, | |
| "learning_rate": 1.2037037037037039e-05, | |
| "loss": 0.309, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.06523765144454799, | |
| "grad_norm": 0.1928149352536188, | |
| "learning_rate": 1.2962962962962964e-05, | |
| "loss": 0.3157, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.06989748369058714, | |
| "grad_norm": 0.17826829224749713, | |
| "learning_rate": 1.388888888888889e-05, | |
| "loss": 0.3099, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.07455731593662628, | |
| "grad_norm": 0.18477204996403923, | |
| "learning_rate": 1.4814814814814815e-05, | |
| "loss": 0.3067, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.07921714818266543, | |
| "grad_norm": 0.1897409236877168, | |
| "learning_rate": 1.5740740740740744e-05, | |
| "loss": 0.31, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.08387698042870456, | |
| "grad_norm": 0.17327980240361615, | |
| "learning_rate": 1.6666666666666667e-05, | |
| "loss": 0.3163, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.08853681267474371, | |
| "grad_norm": 0.1841696908552682, | |
| "learning_rate": 1.7592592592592595e-05, | |
| "loss": 0.3083, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.09319664492078285, | |
| "grad_norm": 0.18204159918734777, | |
| "learning_rate": 1.851851851851852e-05, | |
| "loss": 0.3053, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.097856477166822, | |
| "grad_norm": 0.17179833079133283, | |
| "learning_rate": 1.9444444444444445e-05, | |
| "loss": 0.3058, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.10251630941286113, | |
| "grad_norm": 0.19459644226361492, | |
| "learning_rate": 1.995854922279793e-05, | |
| "loss": 0.3065, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.10717614165890028, | |
| "grad_norm": 0.1901341829663568, | |
| "learning_rate": 1.985492227979275e-05, | |
| "loss": 0.3078, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.11183597390493942, | |
| "grad_norm": 0.18214653715281157, | |
| "learning_rate": 1.9751295336787565e-05, | |
| "loss": 0.3059, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.11649580615097857, | |
| "grad_norm": 0.17815775257629182, | |
| "learning_rate": 1.9647668393782386e-05, | |
| "loss": 0.3017, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.1211556383970177, | |
| "grad_norm": 0.1844645558957454, | |
| "learning_rate": 1.9544041450777206e-05, | |
| "loss": 0.3095, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.12581547064305684, | |
| "grad_norm": 0.2030421375294079, | |
| "learning_rate": 1.9440414507772023e-05, | |
| "loss": 0.305, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.13047530288909598, | |
| "grad_norm": 0.2070381771570163, | |
| "learning_rate": 1.9336787564766843e-05, | |
| "loss": 0.3032, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.13513513513513514, | |
| "grad_norm": 0.19750341679314554, | |
| "learning_rate": 1.923316062176166e-05, | |
| "loss": 0.3183, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.13979496738117428, | |
| "grad_norm": 0.19731313864346534, | |
| "learning_rate": 1.9129533678756477e-05, | |
| "loss": 0.3067, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.14445479962721341, | |
| "grad_norm": 0.17260148498158165, | |
| "learning_rate": 1.9025906735751297e-05, | |
| "loss": 0.3186, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.14911463187325255, | |
| "grad_norm": 0.2008627001637739, | |
| "learning_rate": 1.8922279792746117e-05, | |
| "loss": 0.3085, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.15377446411929171, | |
| "grad_norm": 0.18712990541821012, | |
| "learning_rate": 1.8818652849740934e-05, | |
| "loss": 0.3085, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.15843429636533085, | |
| "grad_norm": 0.17292039090794106, | |
| "learning_rate": 1.8715025906735754e-05, | |
| "loss": 0.2908, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.16309412861137, | |
| "grad_norm": 0.19172387862691817, | |
| "learning_rate": 1.861139896373057e-05, | |
| "loss": 0.3081, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.16775396085740912, | |
| "grad_norm": 0.18688008270471848, | |
| "learning_rate": 1.850777202072539e-05, | |
| "loss": 0.306, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.1724137931034483, | |
| "grad_norm": 0.1745486683593482, | |
| "learning_rate": 1.8404145077720208e-05, | |
| "loss": 0.3032, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.17707362534948742, | |
| "grad_norm": 0.19257694931627597, | |
| "learning_rate": 1.8300518134715028e-05, | |
| "loss": 0.3101, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.18173345759552656, | |
| "grad_norm": 0.2104231700336429, | |
| "learning_rate": 1.8196891191709845e-05, | |
| "loss": 0.3073, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.1863932898415657, | |
| "grad_norm": 0.2198068709198768, | |
| "learning_rate": 1.8093264248704665e-05, | |
| "loss": 0.3097, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.19105312208760486, | |
| "grad_norm": 0.19860541284894057, | |
| "learning_rate": 1.7989637305699482e-05, | |
| "loss": 0.3057, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.195712954333644, | |
| "grad_norm": 0.19761516589964884, | |
| "learning_rate": 1.7886010362694302e-05, | |
| "loss": 0.31, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.20037278657968313, | |
| "grad_norm": 0.1857003842960702, | |
| "learning_rate": 1.778238341968912e-05, | |
| "loss": 0.3035, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.20503261882572227, | |
| "grad_norm": 0.19655970430783293, | |
| "learning_rate": 1.767875647668394e-05, | |
| "loss": 0.309, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.2096924510717614, | |
| "grad_norm": 0.18373076556409915, | |
| "learning_rate": 1.757512953367876e-05, | |
| "loss": 0.2992, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.21435228331780057, | |
| "grad_norm": 0.1855241535339271, | |
| "learning_rate": 1.7471502590673576e-05, | |
| "loss": 0.307, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.2190121155638397, | |
| "grad_norm": 0.18789691001904157, | |
| "learning_rate": 1.7367875647668397e-05, | |
| "loss": 0.3034, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.22367194780987884, | |
| "grad_norm": 0.225904963129947, | |
| "learning_rate": 1.7264248704663214e-05, | |
| "loss": 0.3073, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.22833178005591798, | |
| "grad_norm": 0.2176659893463581, | |
| "learning_rate": 1.716062176165803e-05, | |
| "loss": 0.3082, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.23299161230195714, | |
| "grad_norm": 0.19279163806598115, | |
| "learning_rate": 1.705699481865285e-05, | |
| "loss": 0.3049, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.23765144454799628, | |
| "grad_norm": 0.19980935074977862, | |
| "learning_rate": 1.695336787564767e-05, | |
| "loss": 0.3059, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.2423112767940354, | |
| "grad_norm": 0.20286292649885426, | |
| "learning_rate": 1.6849740932642488e-05, | |
| "loss": 0.3104, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.24697110904007455, | |
| "grad_norm": 0.20806622057526192, | |
| "learning_rate": 1.6746113989637308e-05, | |
| "loss": 0.309, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.2516309412861137, | |
| "grad_norm": 0.1837006492305824, | |
| "learning_rate": 1.6642487046632125e-05, | |
| "loss": 0.3054, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.25629077353215285, | |
| "grad_norm": 0.19343204990996146, | |
| "learning_rate": 1.6538860103626945e-05, | |
| "loss": 0.3065, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.26095060577819196, | |
| "grad_norm": 0.17569927907529842, | |
| "learning_rate": 1.6435233160621765e-05, | |
| "loss": 0.3038, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.2656104380242311, | |
| "grad_norm": 0.1964720824239362, | |
| "learning_rate": 1.6331606217616582e-05, | |
| "loss": 0.3057, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.2702702702702703, | |
| "grad_norm": 0.17919773464670466, | |
| "learning_rate": 1.6227979274611402e-05, | |
| "loss": 0.3119, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.2749301025163094, | |
| "grad_norm": 0.20289962735519462, | |
| "learning_rate": 1.612435233160622e-05, | |
| "loss": 0.322, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.27958993476234856, | |
| "grad_norm": 0.20635974949126276, | |
| "learning_rate": 1.6020725388601036e-05, | |
| "loss": 0.3176, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.2842497670083877, | |
| "grad_norm": 0.2128465330229402, | |
| "learning_rate": 1.5917098445595856e-05, | |
| "loss": 0.3076, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.28890959925442683, | |
| "grad_norm": 0.21095377132964316, | |
| "learning_rate": 1.5813471502590673e-05, | |
| "loss": 0.305, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.293569431500466, | |
| "grad_norm": 0.20117897001414128, | |
| "learning_rate": 1.5709844559585493e-05, | |
| "loss": 0.3107, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.2982292637465051, | |
| "grad_norm": 0.210307217026029, | |
| "learning_rate": 1.5606217616580313e-05, | |
| "loss": 0.2969, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.30288909599254427, | |
| "grad_norm": 0.1953188277902909, | |
| "learning_rate": 1.550259067357513e-05, | |
| "loss": 0.2988, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.30754892823858343, | |
| "grad_norm": 0.20624395571895365, | |
| "learning_rate": 1.539896373056995e-05, | |
| "loss": 0.3077, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.31220876048462254, | |
| "grad_norm": 0.19328155978476294, | |
| "learning_rate": 1.5295336787564767e-05, | |
| "loss": 0.3059, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.3168685927306617, | |
| "grad_norm": 0.18260179855898995, | |
| "learning_rate": 1.5191709844559586e-05, | |
| "loss": 0.3044, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.32152842497670087, | |
| "grad_norm": 0.1840438952163473, | |
| "learning_rate": 1.5088082901554406e-05, | |
| "loss": 0.3059, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.32618825722274, | |
| "grad_norm": 0.19461016289059585, | |
| "learning_rate": 1.4984455958549225e-05, | |
| "loss": 0.316, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.33084808946877914, | |
| "grad_norm": 0.19356848595491336, | |
| "learning_rate": 1.4880829015544043e-05, | |
| "loss": 0.3125, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.33550792171481825, | |
| "grad_norm": 0.1903893603560245, | |
| "learning_rate": 1.4777202072538862e-05, | |
| "loss": 0.3046, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.3401677539608574, | |
| "grad_norm": 0.1888576016162768, | |
| "learning_rate": 1.4673575129533678e-05, | |
| "loss": 0.3027, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.3448275862068966, | |
| "grad_norm": 0.18928501902596923, | |
| "learning_rate": 1.4569948186528497e-05, | |
| "loss": 0.31, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.3494874184529357, | |
| "grad_norm": 0.1966355160370978, | |
| "learning_rate": 1.4466321243523317e-05, | |
| "loss": 0.3041, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.35414725069897485, | |
| "grad_norm": 0.1901672871632349, | |
| "learning_rate": 1.4362694300518136e-05, | |
| "loss": 0.3136, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.35880708294501396, | |
| "grad_norm": 0.19972593288293097, | |
| "learning_rate": 1.4259067357512954e-05, | |
| "loss": 0.3097, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.3634669151910531, | |
| "grad_norm": 0.19468214914369028, | |
| "learning_rate": 1.4155440414507773e-05, | |
| "loss": 0.3068, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.3681267474370923, | |
| "grad_norm": 0.24360814017146365, | |
| "learning_rate": 1.4051813471502591e-05, | |
| "loss": 0.3083, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.3727865796831314, | |
| "grad_norm": 0.20121403150161737, | |
| "learning_rate": 1.394818652849741e-05, | |
| "loss": 0.2992, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.37744641192917056, | |
| "grad_norm": 0.18275980204979905, | |
| "learning_rate": 1.384455958549223e-05, | |
| "loss": 0.3191, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.3821062441752097, | |
| "grad_norm": 0.19597361835647514, | |
| "learning_rate": 1.3740932642487049e-05, | |
| "loss": 0.3102, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.38676607642124883, | |
| "grad_norm": 0.19121995822742002, | |
| "learning_rate": 1.3637305699481867e-05, | |
| "loss": 0.3036, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.391425908667288, | |
| "grad_norm": 0.21411409371703122, | |
| "learning_rate": 1.3533678756476684e-05, | |
| "loss": 0.3138, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.3960857409133271, | |
| "grad_norm": 0.20555940672664647, | |
| "learning_rate": 1.3430051813471503e-05, | |
| "loss": 0.3169, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.40074557315936626, | |
| "grad_norm": 0.19221801982606176, | |
| "learning_rate": 1.3326424870466321e-05, | |
| "loss": 0.3121, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.40540540540540543, | |
| "grad_norm": 0.2014890349514093, | |
| "learning_rate": 1.3222797927461141e-05, | |
| "loss": 0.3175, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.41006523765144454, | |
| "grad_norm": 0.24792004508641477, | |
| "learning_rate": 1.311917098445596e-05, | |
| "loss": 0.3084, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.4147250698974837, | |
| "grad_norm": 0.1954769185865929, | |
| "learning_rate": 1.3015544041450778e-05, | |
| "loss": 0.3076, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.4193849021435228, | |
| "grad_norm": 0.20194311212258356, | |
| "learning_rate": 1.2911917098445597e-05, | |
| "loss": 0.3005, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.424044734389562, | |
| "grad_norm": 0.19502464315598184, | |
| "learning_rate": 1.2808290155440415e-05, | |
| "loss": 0.308, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.42870456663560114, | |
| "grad_norm": 0.1985626146517041, | |
| "learning_rate": 1.2704663212435234e-05, | |
| "loss": 0.3057, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.43336439888164024, | |
| "grad_norm": 0.21421287702611994, | |
| "learning_rate": 1.2601036269430054e-05, | |
| "loss": 0.3033, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.4380242311276794, | |
| "grad_norm": 0.1987802920868664, | |
| "learning_rate": 1.2497409326424873e-05, | |
| "loss": 0.3126, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.4426840633737186, | |
| "grad_norm": 0.1895491437169421, | |
| "learning_rate": 1.239378238341969e-05, | |
| "loss": 0.304, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.4473438956197577, | |
| "grad_norm": 0.1985509321068275, | |
| "learning_rate": 1.2290155440414508e-05, | |
| "loss": 0.3189, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.45200372786579684, | |
| "grad_norm": 0.2173910116765384, | |
| "learning_rate": 1.2186528497409327e-05, | |
| "loss": 0.3092, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.45666356011183595, | |
| "grad_norm": 0.17261612546150368, | |
| "learning_rate": 1.2082901554404145e-05, | |
| "loss": 0.3039, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.4613233923578751, | |
| "grad_norm": 0.20048439029565604, | |
| "learning_rate": 1.1979274611398965e-05, | |
| "loss": 0.3077, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.4659832246039143, | |
| "grad_norm": 0.20844885780214473, | |
| "learning_rate": 1.1875647668393784e-05, | |
| "loss": 0.3179, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.4706430568499534, | |
| "grad_norm": 0.19509047480375188, | |
| "learning_rate": 1.1772020725388602e-05, | |
| "loss": 0.307, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 0.47530288909599255, | |
| "grad_norm": 0.18379044121686752, | |
| "learning_rate": 1.1668393782383421e-05, | |
| "loss": 0.3076, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.47996272134203166, | |
| "grad_norm": 0.20046163502948242, | |
| "learning_rate": 1.1564766839378238e-05, | |
| "loss": 0.3073, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 0.4846225535880708, | |
| "grad_norm": 0.1891503524111604, | |
| "learning_rate": 1.1461139896373056e-05, | |
| "loss": 0.3119, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.48928238583411, | |
| "grad_norm": 0.20769187983919965, | |
| "learning_rate": 1.1357512953367878e-05, | |
| "loss": 0.312, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.4939422180801491, | |
| "grad_norm": 0.1956521662876443, | |
| "learning_rate": 1.1253886010362695e-05, | |
| "loss": 0.3069, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.49860205032618826, | |
| "grad_norm": 0.1851287031638996, | |
| "learning_rate": 1.1150259067357514e-05, | |
| "loss": 0.3117, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 0.5032618825722274, | |
| "grad_norm": 0.20050732452739828, | |
| "learning_rate": 1.1046632124352332e-05, | |
| "loss": 0.2985, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.5079217148182665, | |
| "grad_norm": 0.2219598360154172, | |
| "learning_rate": 1.094300518134715e-05, | |
| "loss": 0.3058, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 0.5125815470643057, | |
| "grad_norm": 0.22159014441089023, | |
| "learning_rate": 1.083937823834197e-05, | |
| "loss": 0.3114, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.5172413793103449, | |
| "grad_norm": 0.21189175620802284, | |
| "learning_rate": 1.073575129533679e-05, | |
| "loss": 0.3097, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 0.5219012115563839, | |
| "grad_norm": 0.20194991625949968, | |
| "learning_rate": 1.0632124352331608e-05, | |
| "loss": 0.3195, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.5265610438024231, | |
| "grad_norm": 0.17929820608315875, | |
| "learning_rate": 1.0528497409326426e-05, | |
| "loss": 0.2966, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 0.5312208760484622, | |
| "grad_norm": 0.18847514867987192, | |
| "learning_rate": 1.0424870466321243e-05, | |
| "loss": 0.3013, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.5358807082945014, | |
| "grad_norm": 0.1918865802913081, | |
| "learning_rate": 1.0321243523316062e-05, | |
| "loss": 0.3054, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 0.5405405405405406, | |
| "grad_norm": 0.19646625910996846, | |
| "learning_rate": 1.021761658031088e-05, | |
| "loss": 0.3039, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.5452003727865797, | |
| "grad_norm": 0.1957026436292748, | |
| "learning_rate": 1.01139896373057e-05, | |
| "loss": 0.3085, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 0.5498602050326188, | |
| "grad_norm": 0.2215747140393126, | |
| "learning_rate": 1.0010362694300519e-05, | |
| "loss": 0.3004, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.554520037278658, | |
| "grad_norm": 0.2096798005029143, | |
| "learning_rate": 9.906735751295338e-06, | |
| "loss": 0.3045, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 0.5591798695246971, | |
| "grad_norm": 0.21043804051484524, | |
| "learning_rate": 9.803108808290156e-06, | |
| "loss": 0.3061, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.5638397017707363, | |
| "grad_norm": 0.18777718502993346, | |
| "learning_rate": 9.699481865284975e-06, | |
| "loss": 0.3048, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 0.5684995340167754, | |
| "grad_norm": 0.20237085315201214, | |
| "learning_rate": 9.595854922279793e-06, | |
| "loss": 0.3132, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.5731593662628145, | |
| "grad_norm": 0.193385995407008, | |
| "learning_rate": 9.492227979274612e-06, | |
| "loss": 0.3042, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 0.5778191985088537, | |
| "grad_norm": 0.1933950220530074, | |
| "learning_rate": 9.388601036269432e-06, | |
| "loss": 0.2979, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.5824790307548928, | |
| "grad_norm": 0.19132327866819904, | |
| "learning_rate": 9.284974093264249e-06, | |
| "loss": 0.3121, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 0.587138863000932, | |
| "grad_norm": 0.18145967807189134, | |
| "learning_rate": 9.181347150259067e-06, | |
| "loss": 0.3104, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.5917986952469712, | |
| "grad_norm": 0.1938112202212723, | |
| "learning_rate": 9.077720207253888e-06, | |
| "loss": 0.3114, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 0.5964585274930102, | |
| "grad_norm": 0.18005219766713837, | |
| "learning_rate": 8.974093264248706e-06, | |
| "loss": 0.3109, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.6011183597390494, | |
| "grad_norm": 0.19954264682643283, | |
| "learning_rate": 8.870466321243523e-06, | |
| "loss": 0.3073, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 0.6057781919850885, | |
| "grad_norm": 0.2029508363977868, | |
| "learning_rate": 8.766839378238343e-06, | |
| "loss": 0.3099, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.6104380242311277, | |
| "grad_norm": 0.20067540450864405, | |
| "learning_rate": 8.663212435233162e-06, | |
| "loss": 0.3131, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 0.6150978564771669, | |
| "grad_norm": 0.18278420923843008, | |
| "learning_rate": 8.55958549222798e-06, | |
| "loss": 0.2968, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.6197576887232059, | |
| "grad_norm": 0.19255194330618958, | |
| "learning_rate": 8.455958549222799e-06, | |
| "loss": 0.3, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 0.6244175209692451, | |
| "grad_norm": 0.17805047451733982, | |
| "learning_rate": 8.352331606217617e-06, | |
| "loss": 0.3002, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.6290773532152842, | |
| "grad_norm": 0.1809729112938702, | |
| "learning_rate": 8.248704663212436e-06, | |
| "loss": 0.3004, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 0.6337371854613234, | |
| "grad_norm": 0.19843229643919744, | |
| "learning_rate": 8.145077720207254e-06, | |
| "loss": 0.3093, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.6383970177073626, | |
| "grad_norm": 0.17094593426100432, | |
| "learning_rate": 8.041450777202073e-06, | |
| "loss": 0.3118, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 0.6430568499534017, | |
| "grad_norm": 0.17842406465044058, | |
| "learning_rate": 7.937823834196891e-06, | |
| "loss": 0.3026, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.6477166821994408, | |
| "grad_norm": 0.18735190688774842, | |
| "learning_rate": 7.834196891191712e-06, | |
| "loss": 0.3097, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 0.65237651444548, | |
| "grad_norm": 0.18672212273790229, | |
| "learning_rate": 7.730569948186528e-06, | |
| "loss": 0.3149, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.6570363466915191, | |
| "grad_norm": 0.16216816399314543, | |
| "learning_rate": 7.626943005181348e-06, | |
| "loss": 0.3026, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 0.6616961789375583, | |
| "grad_norm": 0.19617575721215516, | |
| "learning_rate": 7.523316062176167e-06, | |
| "loss": 0.3036, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.6663560111835974, | |
| "grad_norm": 0.1783695592863534, | |
| "learning_rate": 7.419689119170985e-06, | |
| "loss": 0.2985, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 0.6710158434296365, | |
| "grad_norm": 0.17934516453245036, | |
| "learning_rate": 7.3160621761658035e-06, | |
| "loss": 0.3031, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.6756756756756757, | |
| "grad_norm": 0.19399978320829833, | |
| "learning_rate": 7.212435233160623e-06, | |
| "loss": 0.3077, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 0.6803355079217148, | |
| "grad_norm": 0.19060796532512359, | |
| "learning_rate": 7.108808290155441e-06, | |
| "loss": 0.3039, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.684995340167754, | |
| "grad_norm": 0.17663402079064713, | |
| "learning_rate": 7.005181347150259e-06, | |
| "loss": 0.304, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 0.6896551724137931, | |
| "grad_norm": 0.18728492382652162, | |
| "learning_rate": 6.9015544041450784e-06, | |
| "loss": 0.2996, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.6943150046598322, | |
| "grad_norm": 0.16539414234955993, | |
| "learning_rate": 6.797927461139897e-06, | |
| "loss": 0.3007, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 0.6989748369058714, | |
| "grad_norm": 0.18796932042651304, | |
| "learning_rate": 6.6943005181347155e-06, | |
| "loss": 0.3006, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.7036346691519105, | |
| "grad_norm": 0.1934535934904552, | |
| "learning_rate": 6.590673575129535e-06, | |
| "loss": 0.3051, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 0.7082945013979497, | |
| "grad_norm": 0.17511509631442268, | |
| "learning_rate": 6.487046632124353e-06, | |
| "loss": 0.3047, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.7129543336439889, | |
| "grad_norm": 0.16967569477610708, | |
| "learning_rate": 6.383419689119171e-06, | |
| "loss": 0.2981, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 0.7176141658900279, | |
| "grad_norm": 0.20551530112906796, | |
| "learning_rate": 6.2797927461139905e-06, | |
| "loss": 0.3048, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.7222739981360671, | |
| "grad_norm": 0.1772568831952956, | |
| "learning_rate": 6.176165803108809e-06, | |
| "loss": 0.3106, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 0.7269338303821062, | |
| "grad_norm": 0.17122628778280205, | |
| "learning_rate": 6.0725388601036275e-06, | |
| "loss": 0.2986, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.7315936626281454, | |
| "grad_norm": 0.19005996568436556, | |
| "learning_rate": 5.968911917098445e-06, | |
| "loss": 0.3024, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 0.7362534948741846, | |
| "grad_norm": 0.1896569557324295, | |
| "learning_rate": 5.865284974093265e-06, | |
| "loss": 0.3068, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.7409133271202236, | |
| "grad_norm": 0.17553068397844512, | |
| "learning_rate": 5.761658031088083e-06, | |
| "loss": 0.306, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 0.7455731593662628, | |
| "grad_norm": 0.1902146433481209, | |
| "learning_rate": 5.658031088082902e-06, | |
| "loss": 0.3044, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.750232991612302, | |
| "grad_norm": 0.18379958493058496, | |
| "learning_rate": 5.554404145077721e-06, | |
| "loss": 0.3046, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 0.7548928238583411, | |
| "grad_norm": 0.19238104735204387, | |
| "learning_rate": 5.4507772020725395e-06, | |
| "loss": 0.3133, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.7595526561043803, | |
| "grad_norm": 0.16684993046883195, | |
| "learning_rate": 5.347150259067357e-06, | |
| "loss": 0.3019, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 0.7642124883504194, | |
| "grad_norm": 0.17317168720572065, | |
| "learning_rate": 5.243523316062177e-06, | |
| "loss": 0.3092, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.7688723205964585, | |
| "grad_norm": 0.17257530643463354, | |
| "learning_rate": 5.139896373056995e-06, | |
| "loss": 0.3012, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 0.7735321528424977, | |
| "grad_norm": 0.19022509153976733, | |
| "learning_rate": 5.036269430051814e-06, | |
| "loss": 0.3006, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.7781919850885368, | |
| "grad_norm": 0.18273563180618016, | |
| "learning_rate": 4.932642487046633e-06, | |
| "loss": 0.302, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 0.782851817334576, | |
| "grad_norm": 0.2068935985590348, | |
| "learning_rate": 4.829015544041451e-06, | |
| "loss": 0.3083, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.7875116495806151, | |
| "grad_norm": 0.1787063525187819, | |
| "learning_rate": 4.72538860103627e-06, | |
| "loss": 0.3158, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 0.7921714818266542, | |
| "grad_norm": 0.17589355462106077, | |
| "learning_rate": 4.621761658031089e-06, | |
| "loss": 0.3015, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.7968313140726934, | |
| "grad_norm": 0.19008199962840902, | |
| "learning_rate": 4.518134715025907e-06, | |
| "loss": 0.3087, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 0.8014911463187325, | |
| "grad_norm": 0.1744284628031719, | |
| "learning_rate": 4.414507772020726e-06, | |
| "loss": 0.3146, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.8061509785647717, | |
| "grad_norm": 0.1764675363887709, | |
| "learning_rate": 4.310880829015544e-06, | |
| "loss": 0.3028, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 0.8108108108108109, | |
| "grad_norm": 0.18723433018807362, | |
| "learning_rate": 4.207253886010363e-06, | |
| "loss": 0.3082, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.8154706430568499, | |
| "grad_norm": 0.16928853740679736, | |
| "learning_rate": 4.103626943005182e-06, | |
| "loss": 0.3059, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 0.8201304753028891, | |
| "grad_norm": 0.17841937199548402, | |
| "learning_rate": 4.000000000000001e-06, | |
| "loss": 0.3042, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.8247903075489282, | |
| "grad_norm": 0.1656726857328673, | |
| "learning_rate": 3.896373056994819e-06, | |
| "loss": 0.3079, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 0.8294501397949674, | |
| "grad_norm": 0.17487630016211303, | |
| "learning_rate": 3.7927461139896377e-06, | |
| "loss": 0.3027, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.8341099720410066, | |
| "grad_norm": 0.16843425617538177, | |
| "learning_rate": 3.6891191709844567e-06, | |
| "loss": 0.3172, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 0.8387698042870456, | |
| "grad_norm": 0.17177773487516515, | |
| "learning_rate": 3.5854922279792748e-06, | |
| "loss": 0.3055, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.8434296365330848, | |
| "grad_norm": 0.16684455749445157, | |
| "learning_rate": 3.4818652849740937e-06, | |
| "loss": 0.3031, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 0.848089468779124, | |
| "grad_norm": 0.171250184663666, | |
| "learning_rate": 3.3782383419689123e-06, | |
| "loss": 0.3068, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.8527493010251631, | |
| "grad_norm": 0.1904755654711732, | |
| "learning_rate": 3.274611398963731e-06, | |
| "loss": 0.3091, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 0.8574091332712023, | |
| "grad_norm": 0.17322382387681076, | |
| "learning_rate": 3.1709844559585493e-06, | |
| "loss": 0.3029, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.8620689655172413, | |
| "grad_norm": 0.17587838098911934, | |
| "learning_rate": 3.0673575129533683e-06, | |
| "loss": 0.2978, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 0.8667287977632805, | |
| "grad_norm": 0.17389743789103038, | |
| "learning_rate": 2.963730569948187e-06, | |
| "loss": 0.2979, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.8713886300093197, | |
| "grad_norm": 0.18137733240826867, | |
| "learning_rate": 2.8601036269430053e-06, | |
| "loss": 0.3027, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 0.8760484622553588, | |
| "grad_norm": 0.17290656767359902, | |
| "learning_rate": 2.7564766839378243e-06, | |
| "loss": 0.3054, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.880708294501398, | |
| "grad_norm": 0.17569375365058235, | |
| "learning_rate": 2.6528497409326424e-06, | |
| "loss": 0.3023, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 0.8853681267474371, | |
| "grad_norm": 0.1727950452551694, | |
| "learning_rate": 2.5492227979274614e-06, | |
| "loss": 0.3137, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.8900279589934762, | |
| "grad_norm": 0.1740200974029668, | |
| "learning_rate": 2.44559585492228e-06, | |
| "loss": 0.307, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 0.8946877912395154, | |
| "grad_norm": 0.1732738746462953, | |
| "learning_rate": 2.3419689119170984e-06, | |
| "loss": 0.3111, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.8993476234855545, | |
| "grad_norm": 0.18648726628836773, | |
| "learning_rate": 2.2383419689119174e-06, | |
| "loss": 0.3069, | |
| "step": 965 | |
| }, | |
| { | |
| "epoch": 0.9040074557315937, | |
| "grad_norm": 0.1697331695842795, | |
| "learning_rate": 2.134715025906736e-06, | |
| "loss": 0.297, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.9086672879776329, | |
| "grad_norm": 0.16507665028070173, | |
| "learning_rate": 2.0310880829015544e-06, | |
| "loss": 0.3125, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 0.9133271202236719, | |
| "grad_norm": 0.17494951036076584, | |
| "learning_rate": 1.9274611398963734e-06, | |
| "loss": 0.3041, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.9179869524697111, | |
| "grad_norm": 0.17876267200484872, | |
| "learning_rate": 1.823834196891192e-06, | |
| "loss": 0.3024, | |
| "step": 985 | |
| }, | |
| { | |
| "epoch": 0.9226467847157502, | |
| "grad_norm": 0.17300125358384327, | |
| "learning_rate": 1.7202072538860104e-06, | |
| "loss": 0.3007, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.9273066169617894, | |
| "grad_norm": 0.1739041985560771, | |
| "learning_rate": 1.6165803108808292e-06, | |
| "loss": 0.3039, | |
| "step": 995 | |
| }, | |
| { | |
| "epoch": 0.9319664492078286, | |
| "grad_norm": 0.17113300633163106, | |
| "learning_rate": 1.5129533678756477e-06, | |
| "loss": 0.3035, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.9366262814538676, | |
| "grad_norm": 0.16281913618369226, | |
| "learning_rate": 1.4093264248704663e-06, | |
| "loss": 0.3071, | |
| "step": 1005 | |
| }, | |
| { | |
| "epoch": 0.9412861136999068, | |
| "grad_norm": 0.17924475619300242, | |
| "learning_rate": 1.3056994818652852e-06, | |
| "loss": 0.3058, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.9459459459459459, | |
| "grad_norm": 0.17688052898291365, | |
| "learning_rate": 1.2020725388601037e-06, | |
| "loss": 0.3087, | |
| "step": 1015 | |
| }, | |
| { | |
| "epoch": 0.9506057781919851, | |
| "grad_norm": 0.16179921554930488, | |
| "learning_rate": 1.0984455958549225e-06, | |
| "loss": 0.3044, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.9552656104380243, | |
| "grad_norm": 0.17553736079048324, | |
| "learning_rate": 9.94818652849741e-07, | |
| "loss": 0.3129, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 0.9599254426840633, | |
| "grad_norm": 0.1704446487232818, | |
| "learning_rate": 8.911917098445596e-07, | |
| "loss": 0.3046, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.9645852749301025, | |
| "grad_norm": 0.169056664565789, | |
| "learning_rate": 7.875647668393784e-07, | |
| "loss": 0.3072, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 0.9692451071761417, | |
| "grad_norm": 0.16668251867193293, | |
| "learning_rate": 6.839378238341969e-07, | |
| "loss": 0.3036, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.9739049394221808, | |
| "grad_norm": 0.1639424483827266, | |
| "learning_rate": 5.803108808290156e-07, | |
| "loss": 0.2989, | |
| "step": 1045 | |
| }, | |
| { | |
| "epoch": 0.97856477166822, | |
| "grad_norm": 0.17474417966467756, | |
| "learning_rate": 4.7668393782383424e-07, | |
| "loss": 0.3187, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.983224603914259, | |
| "grad_norm": 0.17412618700034416, | |
| "learning_rate": 3.730569948186528e-07, | |
| "loss": 0.2996, | |
| "step": 1055 | |
| }, | |
| { | |
| "epoch": 0.9878844361602982, | |
| "grad_norm": 0.16453567016761128, | |
| "learning_rate": 2.694300518134715e-07, | |
| "loss": 0.3028, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.9925442684063374, | |
| "grad_norm": 0.1609387610584271, | |
| "learning_rate": 1.6580310880829015e-07, | |
| "loss": 0.3061, | |
| "step": 1065 | |
| }, | |
| { | |
| "epoch": 0.9972041006523765, | |
| "grad_norm": 0.1665161978210062, | |
| "learning_rate": 6.217616580310881e-08, | |
| "loss": 0.303, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "step": 1073, | |
| "total_flos": 9.186429923093381e+17, | |
| "train_loss": 0.30834408108585926, | |
| "train_runtime": 35203.5942, | |
| "train_samples_per_second": 0.488, | |
| "train_steps_per_second": 0.03 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 1073, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 9.186429923093381e+17, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |