| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.9991503823279524, | |
| "eval_steps": 29, | |
| "global_step": 147, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.006796941376380629, | |
| "grad_norm": 4.598693370819092, | |
| "learning_rate": 1.111111111111111e-06, | |
| "loss": 0.5628, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.013593882752761258, | |
| "grad_norm": 3.3473575115203857, | |
| "learning_rate": 2.222222222222222e-06, | |
| "loss": 0.5195, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.020390824129141887, | |
| "grad_norm": 4.778041362762451, | |
| "learning_rate": 3.3333333333333333e-06, | |
| "loss": 0.5557, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.027187765505522515, | |
| "grad_norm": 4.810835838317871, | |
| "learning_rate": 4.444444444444444e-06, | |
| "loss": 0.5973, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.033984706881903144, | |
| "grad_norm": 4.811781406402588, | |
| "learning_rate": 5.555555555555557e-06, | |
| "loss": 0.5647, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.04078164825828377, | |
| "grad_norm": 2.4056437015533447, | |
| "learning_rate": 6.666666666666667e-06, | |
| "loss": 0.5343, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.0475785896346644, | |
| "grad_norm": 1.500022530555725, | |
| "learning_rate": 7.77777777777778e-06, | |
| "loss": 0.4847, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.05437553101104503, | |
| "grad_norm": 1.5126186609268188, | |
| "learning_rate": 8.888888888888888e-06, | |
| "loss": 0.5285, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.06117247238742566, | |
| "grad_norm": 1.3590396642684937, | |
| "learning_rate": 1e-05, | |
| "loss": 0.4215, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.06796941376380629, | |
| "grad_norm": 1.0946522951126099, | |
| "learning_rate": 1.1111111111111113e-05, | |
| "loss": 0.4077, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.07476635514018691, | |
| "grad_norm": 1.035597801208496, | |
| "learning_rate": 1.2222222222222224e-05, | |
| "loss": 0.3961, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.08156329651656755, | |
| "grad_norm": 0.9033994078636169, | |
| "learning_rate": 1.3333333333333333e-05, | |
| "loss": 0.3715, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.08836023789294817, | |
| "grad_norm": 1.1996865272521973, | |
| "learning_rate": 1.4444444444444446e-05, | |
| "loss": 0.3576, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.0951571792693288, | |
| "grad_norm": 0.9262693524360657, | |
| "learning_rate": 1.555555555555556e-05, | |
| "loss": 0.3483, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.10195412064570943, | |
| "grad_norm": 0.8593380451202393, | |
| "learning_rate": 1.6666666666666667e-05, | |
| "loss": 0.3552, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.10875106202209006, | |
| "grad_norm": 0.9698188900947571, | |
| "learning_rate": 1.7777777777777777e-05, | |
| "loss": 0.329, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.11554800339847068, | |
| "grad_norm": 1.0091280937194824, | |
| "learning_rate": 1.888888888888889e-05, | |
| "loss": 0.3793, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.12234494477485132, | |
| "grad_norm": 0.8758748173713684, | |
| "learning_rate": 2e-05, | |
| "loss": 0.3134, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.12914188615123195, | |
| "grad_norm": 0.7287446856498718, | |
| "learning_rate": 1.9999724204599748e-05, | |
| "loss": 0.3261, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.13593882752761258, | |
| "grad_norm": 0.9394727349281311, | |
| "learning_rate": 1.9998896833611603e-05, | |
| "loss": 0.3553, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.1427357689039932, | |
| "grad_norm": 0.8428260684013367, | |
| "learning_rate": 1.9997517932672592e-05, | |
| "loss": 0.3411, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.14953271028037382, | |
| "grad_norm": 0.8483632802963257, | |
| "learning_rate": 1.999558757784162e-05, | |
| "loss": 0.351, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.15632965165675447, | |
| "grad_norm": 0.7444515228271484, | |
| "learning_rate": 1.999310587559529e-05, | |
| "loss": 0.3487, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.1631265930331351, | |
| "grad_norm": 0.8404412865638733, | |
| "learning_rate": 1.999007296282201e-05, | |
| "loss": 0.3439, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.16992353440951571, | |
| "grad_norm": 0.7445095181465149, | |
| "learning_rate": 1.9986489006814454e-05, | |
| "loss": 0.2972, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.17672047578589634, | |
| "grad_norm": 0.7440016269683838, | |
| "learning_rate": 1.9982354205260347e-05, | |
| "loss": 0.3391, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.18351741716227699, | |
| "grad_norm": 0.6863638162612915, | |
| "learning_rate": 1.9977668786231536e-05, | |
| "loss": 0.2993, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.1903143585386576, | |
| "grad_norm": 0.7364312410354614, | |
| "learning_rate": 1.9972433008171417e-05, | |
| "loss": 0.3169, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.19711129991503823, | |
| "grad_norm": 0.7718295454978943, | |
| "learning_rate": 1.9966647159880703e-05, | |
| "loss": 0.3291, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.19711129991503823, | |
| "eval_loss": 0.3171866536140442, | |
| "eval_runtime": 25.1652, | |
| "eval_samples_per_second": 19.869, | |
| "eval_steps_per_second": 0.636, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.20390824129141885, | |
| "grad_norm": 0.7047903537750244, | |
| "learning_rate": 1.9960311560501457e-05, | |
| "loss": 0.3326, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.2107051826677995, | |
| "grad_norm": 0.7129018902778625, | |
| "learning_rate": 1.9953426559499508e-05, | |
| "loss": 0.3404, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.21750212404418012, | |
| "grad_norm": 0.740166187286377, | |
| "learning_rate": 1.9945992536645188e-05, | |
| "loss": 0.3283, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.22429906542056074, | |
| "grad_norm": 0.6873639225959778, | |
| "learning_rate": 1.993800990199235e-05, | |
| "loss": 0.3064, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.23109600679694137, | |
| "grad_norm": 0.7272984981536865, | |
| "learning_rate": 1.9929479095855783e-05, | |
| "loss": 0.3305, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.23789294817332202, | |
| "grad_norm": 0.7347909808158875, | |
| "learning_rate": 1.99204005887869e-05, | |
| "loss": 0.323, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.24468988954970264, | |
| "grad_norm": 0.7388918399810791, | |
| "learning_rate": 1.9910774881547803e-05, | |
| "loss": 0.3013, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.25148683092608326, | |
| "grad_norm": 0.702777087688446, | |
| "learning_rate": 1.990060250508365e-05, | |
| "loss": 0.3274, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.2582837723024639, | |
| "grad_norm": 0.7246643900871277, | |
| "learning_rate": 1.9889884020493363e-05, | |
| "loss": 0.3362, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.2650807136788445, | |
| "grad_norm": 0.6060255765914917, | |
| "learning_rate": 1.9878620018998696e-05, | |
| "loss": 0.287, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.27187765505522515, | |
| "grad_norm": 0.6417918801307678, | |
| "learning_rate": 1.986681112191161e-05, | |
| "loss": 0.2818, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.2786745964316058, | |
| "grad_norm": 0.7274541258811951, | |
| "learning_rate": 1.98544579806e-05, | |
| "loss": 0.3222, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.2854715378079864, | |
| "grad_norm": 0.7122014164924622, | |
| "learning_rate": 1.984156127645178e-05, | |
| "loss": 0.3392, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.29226847918436705, | |
| "grad_norm": 0.6548435688018799, | |
| "learning_rate": 1.9828121720837288e-05, | |
| "loss": 0.3056, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.29906542056074764, | |
| "grad_norm": 0.6896780133247375, | |
| "learning_rate": 1.9814140055070044e-05, | |
| "loss": 0.3107, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.3058623619371283, | |
| "grad_norm": 0.7305508255958557, | |
| "learning_rate": 1.979961705036587e-05, | |
| "loss": 0.3077, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.31265930331350894, | |
| "grad_norm": 0.6713555455207825, | |
| "learning_rate": 1.9784553507800346e-05, | |
| "loss": 0.2747, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.31945624468988953, | |
| "grad_norm": 0.7295777797698975, | |
| "learning_rate": 1.9768950258264625e-05, | |
| "loss": 0.3126, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.3262531860662702, | |
| "grad_norm": 0.6292139291763306, | |
| "learning_rate": 1.975280816241959e-05, | |
| "loss": 0.2966, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.33305012744265083, | |
| "grad_norm": 0.6628125309944153, | |
| "learning_rate": 1.9736128110648407e-05, | |
| "loss": 0.3266, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.33984706881903143, | |
| "grad_norm": 0.6593727469444275, | |
| "learning_rate": 1.9718911023007382e-05, | |
| "loss": 0.3055, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.3466440101954121, | |
| "grad_norm": 0.737305760383606, | |
| "learning_rate": 1.970115784917523e-05, | |
| "loss": 0.3038, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.35344095157179267, | |
| "grad_norm": 0.6167359948158264, | |
| "learning_rate": 1.9682869568400683e-05, | |
| "loss": 0.3112, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.3602378929481733, | |
| "grad_norm": 0.6573403477668762, | |
| "learning_rate": 1.9664047189448496e-05, | |
| "loss": 0.2971, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.36703483432455397, | |
| "grad_norm": 0.7991875410079956, | |
| "learning_rate": 1.964469175054377e-05, | |
| "loss": 0.3504, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.37383177570093457, | |
| "grad_norm": 0.6709067225456238, | |
| "learning_rate": 1.9624804319314704e-05, | |
| "loss": 0.3307, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.3806287170773152, | |
| "grad_norm": 0.5573310852050781, | |
| "learning_rate": 1.9604385992733718e-05, | |
| "loss": 0.2759, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.3874256584536958, | |
| "grad_norm": 0.6823217868804932, | |
| "learning_rate": 1.9583437897056915e-05, | |
| "loss": 0.2965, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.39422259983007646, | |
| "grad_norm": 0.6070125699043274, | |
| "learning_rate": 1.9561961187761987e-05, | |
| "loss": 0.2893, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.39422259983007646, | |
| "eval_loss": 0.29359060525894165, | |
| "eval_runtime": 24.6748, | |
| "eval_samples_per_second": 20.264, | |
| "eval_steps_per_second": 0.648, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.4010195412064571, | |
| "grad_norm": 0.6562219858169556, | |
| "learning_rate": 1.953995704948446e-05, | |
| "loss": 0.2821, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.4078164825828377, | |
| "grad_norm": 0.6055108904838562, | |
| "learning_rate": 1.9517426695952358e-05, | |
| "loss": 0.297, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.41461342395921835, | |
| "grad_norm": 0.7652341723442078, | |
| "learning_rate": 1.9494371369919253e-05, | |
| "loss": 0.3092, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.421410365335599, | |
| "grad_norm": 0.612332284450531, | |
| "learning_rate": 1.9470792343095718e-05, | |
| "loss": 0.291, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.4282073067119796, | |
| "grad_norm": 0.5866986513137817, | |
| "learning_rate": 1.944669091607919e-05, | |
| "loss": 0.2586, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.43500424808836025, | |
| "grad_norm": 0.847610592842102, | |
| "learning_rate": 1.9422068418282204e-05, | |
| "loss": 0.3567, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.44180118946474084, | |
| "grad_norm": 0.6735373139381409, | |
| "learning_rate": 1.9396926207859085e-05, | |
| "loss": 0.3018, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.4485981308411215, | |
| "grad_norm": 0.7430619597434998, | |
| "learning_rate": 1.9371265671631038e-05, | |
| "loss": 0.3245, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.45539507221750214, | |
| "grad_norm": 0.7295202016830444, | |
| "learning_rate": 1.9345088225009626e-05, | |
| "loss": 0.3464, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.46219201359388273, | |
| "grad_norm": 0.6956584453582764, | |
| "learning_rate": 1.931839531191873e-05, | |
| "loss": 0.3116, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.4689889549702634, | |
| "grad_norm": 0.7618936896324158, | |
| "learning_rate": 1.9291188404714876e-05, | |
| "loss": 0.2802, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.47578589634664403, | |
| "grad_norm": 0.6563156843185425, | |
| "learning_rate": 1.926346900410604e-05, | |
| "loss": 0.3162, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.4825828377230246, | |
| "grad_norm": 0.6584069728851318, | |
| "learning_rate": 1.9235238639068855e-05, | |
| "loss": 0.3147, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.4893797790994053, | |
| "grad_norm": 0.7286915183067322, | |
| "learning_rate": 1.920649886676429e-05, | |
| "loss": 0.3232, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.49617672047578587, | |
| "grad_norm": 0.6320315599441528, | |
| "learning_rate": 1.9177251272451742e-05, | |
| "loss": 0.2946, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.5029736618521665, | |
| "grad_norm": 0.7056065797805786, | |
| "learning_rate": 1.914749746940161e-05, | |
| "loss": 0.3148, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.5097706032285472, | |
| "grad_norm": 0.6424931287765503, | |
| "learning_rate": 1.9117239098806296e-05, | |
| "loss": 0.2757, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.5165675446049278, | |
| "grad_norm": 0.6416671872138977, | |
| "learning_rate": 1.9086477829689688e-05, | |
| "loss": 0.2962, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.5233644859813084, | |
| "grad_norm": 0.7531472444534302, | |
| "learning_rate": 1.905521535881509e-05, | |
| "loss": 0.3129, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.530161427357689, | |
| "grad_norm": 0.631662130355835, | |
| "learning_rate": 1.902345341059164e-05, | |
| "loss": 0.3076, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.5369583687340697, | |
| "grad_norm": 0.6587047576904297, | |
| "learning_rate": 1.8991193736979176e-05, | |
| "loss": 0.3031, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.5437553101104503, | |
| "grad_norm": 0.6439694762229919, | |
| "learning_rate": 1.895843811739162e-05, | |
| "loss": 0.2765, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.550552251486831, | |
| "grad_norm": 0.584459662437439, | |
| "learning_rate": 1.8925188358598815e-05, | |
| "loss": 0.2602, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.5573491928632116, | |
| "grad_norm": 0.6689626574516296, | |
| "learning_rate": 1.8891446294626868e-05, | |
| "loss": 0.2842, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.5641461342395921, | |
| "grad_norm": 0.664592981338501, | |
| "learning_rate": 1.8857213786656986e-05, | |
| "loss": 0.2827, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.5709430756159728, | |
| "grad_norm": 0.5949574708938599, | |
| "learning_rate": 1.882249272292282e-05, | |
| "loss": 0.2781, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.5777400169923534, | |
| "grad_norm": 0.6539214253425598, | |
| "learning_rate": 1.87872850186063e-05, | |
| "loss": 0.2772, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.5845369583687341, | |
| "grad_norm": 0.6778468489646912, | |
| "learning_rate": 1.8751592615732007e-05, | |
| "loss": 0.2938, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.5913338997451147, | |
| "grad_norm": 0.7007662057876587, | |
| "learning_rate": 1.871541748306005e-05, | |
| "loss": 0.2962, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.5913338997451147, | |
| "eval_loss": 0.288781076669693, | |
| "eval_runtime": 24.6643, | |
| "eval_samples_per_second": 20.272, | |
| "eval_steps_per_second": 0.649, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.5981308411214953, | |
| "grad_norm": 0.6959559321403503, | |
| "learning_rate": 1.867876161597747e-05, | |
| "loss": 0.2944, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.6049277824978759, | |
| "grad_norm": 0.6510220766067505, | |
| "learning_rate": 1.8641627036388168e-05, | |
| "loss": 0.2974, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.6117247238742566, | |
| "grad_norm": 0.6992867588996887, | |
| "learning_rate": 1.8604015792601395e-05, | |
| "loss": 0.2915, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.6185216652506372, | |
| "grad_norm": 0.6997932195663452, | |
| "learning_rate": 1.856592995921876e-05, | |
| "loss": 0.2795, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.6253186066270179, | |
| "grad_norm": 0.747304379940033, | |
| "learning_rate": 1.852737163701979e-05, | |
| "loss": 0.2991, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.6321155480033984, | |
| "grad_norm": 0.6337850093841553, | |
| "learning_rate": 1.8488342952846074e-05, | |
| "loss": 0.2816, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.6389124893797791, | |
| "grad_norm": 0.6768142580986023, | |
| "learning_rate": 1.844884605948392e-05, | |
| "loss": 0.2882, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.6457094307561597, | |
| "grad_norm": 0.6271469593048096, | |
| "learning_rate": 1.8408883135545634e-05, | |
| "loss": 0.2767, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.6525063721325404, | |
| "grad_norm": 0.751917839050293, | |
| "learning_rate": 1.8368456385349333e-05, | |
| "loss": 0.3126, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.659303313508921, | |
| "grad_norm": 0.6843672394752502, | |
| "learning_rate": 1.832756803879737e-05, | |
| "loss": 0.3092, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.6661002548853017, | |
| "grad_norm": 0.6191291213035583, | |
| "learning_rate": 1.8286220351253324e-05, | |
| "loss": 0.27, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.6728971962616822, | |
| "grad_norm": 0.7665583491325378, | |
| "learning_rate": 1.8244415603417603e-05, | |
| "loss": 0.3247, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.6796941376380629, | |
| "grad_norm": 0.5895417332649231, | |
| "learning_rate": 1.8202156101201646e-05, | |
| "loss": 0.2967, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.6864910790144435, | |
| "grad_norm": 0.697557270526886, | |
| "learning_rate": 1.8159444175600706e-05, | |
| "loss": 0.3319, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.6932880203908242, | |
| "grad_norm": 0.6296725273132324, | |
| "learning_rate": 1.8116282182565313e-05, | |
| "loss": 0.3177, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.7000849617672048, | |
| "grad_norm": 0.5775023102760315, | |
| "learning_rate": 1.8072672502871295e-05, | |
| "loss": 0.2712, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.7068819031435853, | |
| "grad_norm": 0.5521108508110046, | |
| "learning_rate": 1.802861754198847e-05, | |
| "loss": 0.2724, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.713678844519966, | |
| "grad_norm": 0.7251706719398499, | |
| "learning_rate": 1.7984119729947944e-05, | |
| "loss": 0.3194, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.7204757858963466, | |
| "grad_norm": 0.6509170532226562, | |
| "learning_rate": 1.79391815212081e-05, | |
| "loss": 0.2859, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.7272727272727273, | |
| "grad_norm": 0.7021235227584839, | |
| "learning_rate": 1.7893805394519187e-05, | |
| "loss": 0.3059, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.7340696686491079, | |
| "grad_norm": 0.6381179690361023, | |
| "learning_rate": 1.7847993852786612e-05, | |
| "loss": 0.2556, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.7408666100254885, | |
| "grad_norm": 0.6494466662406921, | |
| "learning_rate": 1.780174942293287e-05, | |
| "loss": 0.2933, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.7476635514018691, | |
| "grad_norm": 0.7242010831832886, | |
| "learning_rate": 1.7755074655758174e-05, | |
| "loss": 0.3111, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.7544604927782498, | |
| "grad_norm": 0.6928083896636963, | |
| "learning_rate": 1.7707972125799738e-05, | |
| "loss": 0.2913, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.7612574341546304, | |
| "grad_norm": 0.6862010955810547, | |
| "learning_rate": 1.766044443118978e-05, | |
| "loss": 0.2961, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.7680543755310111, | |
| "grad_norm": 0.622751772403717, | |
| "learning_rate": 1.761249419351222e-05, | |
| "loss": 0.2581, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.7748513169073916, | |
| "grad_norm": 0.6430375576019287, | |
| "learning_rate": 1.7564124057658057e-05, | |
| "loss": 0.2817, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.7816482582837723, | |
| "grad_norm": 0.5642279386520386, | |
| "learning_rate": 1.7515336691679478e-05, | |
| "loss": 0.2759, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.7884451996601529, | |
| "grad_norm": 0.5759322047233582, | |
| "learning_rate": 1.746613478664271e-05, | |
| "loss": 0.2708, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.7884451996601529, | |
| "eval_loss": 0.28095486760139465, | |
| "eval_runtime": 24.6484, | |
| "eval_samples_per_second": 20.285, | |
| "eval_steps_per_second": 0.649, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.7952421410365336, | |
| "grad_norm": 0.6173487305641174, | |
| "learning_rate": 1.7416521056479577e-05, | |
| "loss": 0.3005, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.8020390824129142, | |
| "grad_norm": 0.6516419649124146, | |
| "learning_rate": 1.736649823783779e-05, | |
| "loss": 0.2788, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.8088360237892949, | |
| "grad_norm": 0.6238332986831665, | |
| "learning_rate": 1.7316069089930007e-05, | |
| "loss": 0.2789, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.8156329651656754, | |
| "grad_norm": 0.5984514355659485, | |
| "learning_rate": 1.7265236394381634e-05, | |
| "loss": 0.3, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.822429906542056, | |
| "grad_norm": 0.6230698823928833, | |
| "learning_rate": 1.7214002955077394e-05, | |
| "loss": 0.2579, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.8292268479184367, | |
| "grad_norm": 0.6018764972686768, | |
| "learning_rate": 1.7162371598006668e-05, | |
| "loss": 0.3065, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.8360237892948174, | |
| "grad_norm": 0.6184946298599243, | |
| "learning_rate": 1.711034517110761e-05, | |
| "loss": 0.2699, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.842820730671198, | |
| "grad_norm": 0.5912549495697021, | |
| "learning_rate": 1.705792654411007e-05, | |
| "loss": 0.2656, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.8496176720475785, | |
| "grad_norm": 0.6374722123146057, | |
| "learning_rate": 1.7005118608377288e-05, | |
| "loss": 0.3074, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.8564146134239592, | |
| "grad_norm": 0.6522223353385925, | |
| "learning_rate": 1.6951924276746425e-05, | |
| "loss": 0.2842, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.8632115548003398, | |
| "grad_norm": 0.6337326765060425, | |
| "learning_rate": 1.689834648336787e-05, | |
| "loss": 0.3009, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.8700084961767205, | |
| "grad_norm": 0.6591935753822327, | |
| "learning_rate": 1.6844388183543418e-05, | |
| "loss": 0.2758, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.8768054375531011, | |
| "grad_norm": 0.71869957447052, | |
| "learning_rate": 1.6790052353563254e-05, | |
| "loss": 0.3036, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.8836023789294817, | |
| "grad_norm": 0.6463243961334229, | |
| "learning_rate": 1.6735341990541766e-05, | |
| "loss": 0.2749, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.8903993203058623, | |
| "grad_norm": 0.5988742709159851, | |
| "learning_rate": 1.6680260112252253e-05, | |
| "loss": 0.2857, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.897196261682243, | |
| "grad_norm": 0.5640173554420471, | |
| "learning_rate": 1.6624809756960445e-05, | |
| "loss": 0.265, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.9039932030586236, | |
| "grad_norm": 0.6231476068496704, | |
| "learning_rate": 1.656899398325693e-05, | |
| "loss": 0.2837, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.9107901444350043, | |
| "grad_norm": 0.644237220287323, | |
| "learning_rate": 1.651281586988844e-05, | |
| "loss": 0.2911, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.9175870858113849, | |
| "grad_norm": 0.6296971440315247, | |
| "learning_rate": 1.6456278515588023e-05, | |
| "loss": 0.3025, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.9243840271877655, | |
| "grad_norm": 0.6154587864875793, | |
| "learning_rate": 1.639938503890414e-05, | |
| "loss": 0.2427, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.9311809685641461, | |
| "grad_norm": 0.6521453261375427, | |
| "learning_rate": 1.6342138578028613e-05, | |
| "loss": 0.2742, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.9379779099405268, | |
| "grad_norm": 0.7209259271621704, | |
| "learning_rate": 1.6284542290623568e-05, | |
| "loss": 0.3017, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.9447748513169074, | |
| "grad_norm": 0.6163744330406189, | |
| "learning_rate": 1.622659935364723e-05, | |
| "loss": 0.2779, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.9515717926932881, | |
| "grad_norm": 0.622027575969696, | |
| "learning_rate": 1.61683129631787e-05, | |
| "loss": 0.2767, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.9583687340696686, | |
| "grad_norm": 0.6272425055503845, | |
| "learning_rate": 1.6109686334241655e-05, | |
| "loss": 0.2644, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.9651656754460493, | |
| "grad_norm": 0.5646295547485352, | |
| "learning_rate": 1.605072270062701e-05, | |
| "loss": 0.2589, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.9719626168224299, | |
| "grad_norm": 0.6147757768630981, | |
| "learning_rate": 1.599142531471456e-05, | |
| "loss": 0.29, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.9787595581988106, | |
| "grad_norm": 0.610895574092865, | |
| "learning_rate": 1.5931797447293553e-05, | |
| "loss": 0.262, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.9855564995751912, | |
| "grad_norm": 0.5855327248573303, | |
| "learning_rate": 1.5871842387382307e-05, | |
| "loss": 0.2435, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.9855564995751912, | |
| "eval_loss": 0.27577897906303406, | |
| "eval_runtime": 24.6998, | |
| "eval_samples_per_second": 20.243, | |
| "eval_steps_per_second": 0.648, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.9923534409515717, | |
| "grad_norm": 0.6426906585693359, | |
| "learning_rate": 1.5811563442046768e-05, | |
| "loss": 0.2755, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.9991503823279524, | |
| "grad_norm": 0.5879192352294922, | |
| "learning_rate": 1.5750963936218104e-05, | |
| "loss": 0.2515, | |
| "step": 147 | |
| } | |
| ], | |
| "logging_steps": 1.0, | |
| "max_steps": 441, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 10.0, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.877564828072018e+17, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |