{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 0.32786885245901637, "eval_steps": 500, "global_step": 100, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.003278688524590164, "grad_norm": 58.999969482421875, "learning_rate": 0.0, "loss": 3.8125, "step": 1 }, { "epoch": 0.006557377049180328, "grad_norm": 56.5999870300293, "learning_rate": 3.5714285714285716e-07, "loss": 3.8984, "step": 2 }, { "epoch": 0.009836065573770493, "grad_norm": 58.442317962646484, "learning_rate": 7.142857142857143e-07, "loss": 3.8359, "step": 3 }, { "epoch": 0.013114754098360656, "grad_norm": 59.13441467285156, "learning_rate": 1.0714285714285714e-06, "loss": 3.7422, "step": 4 }, { "epoch": 0.01639344262295082, "grad_norm": 59.55280685424805, "learning_rate": 1.4285714285714286e-06, "loss": 3.7383, "step": 5 }, { "epoch": 0.019672131147540985, "grad_norm": 57.113956451416016, "learning_rate": 1.7857142857142859e-06, "loss": 3.8281, "step": 6 }, { "epoch": 0.022950819672131147, "grad_norm": 44.69753646850586, "learning_rate": 2.1428571428571427e-06, "loss": 3.6562, "step": 7 }, { "epoch": 0.02622950819672131, "grad_norm": 42.74599075317383, "learning_rate": 2.5e-06, "loss": 3.668, "step": 8 }, { "epoch": 0.029508196721311476, "grad_norm": 27.557815551757812, "learning_rate": 2.8571428571428573e-06, "loss": 3.3438, "step": 9 }, { "epoch": 0.03278688524590164, "grad_norm": 25.87570571899414, "learning_rate": 3.2142857142857147e-06, "loss": 3.3633, "step": 10 }, { "epoch": 0.036065573770491806, "grad_norm": 25.204158782958984, "learning_rate": 3.5714285714285718e-06, "loss": 3.2852, "step": 11 }, { "epoch": 0.03934426229508197, "grad_norm": 25.082277297973633, "learning_rate": 3.928571428571429e-06, "loss": 3.2188, "step": 12 }, { "epoch": 0.04262295081967213, "grad_norm": 21.52399253845215, "learning_rate": 4.2857142857142855e-06, "loss": 3.1016, "step": 13 }, { "epoch": 0.04590163934426229, "grad_norm": 21.302213668823242, "learning_rate": 4.642857142857144e-06, "loss": 3.0938, "step": 14 }, { "epoch": 0.04918032786885246, "grad_norm": 17.48920249938965, "learning_rate": 5e-06, "loss": 3.0703, "step": 15 }, { "epoch": 0.05245901639344262, "grad_norm": 14.723689079284668, "learning_rate": 5.357142857142857e-06, "loss": 2.7812, "step": 16 }, { "epoch": 0.05573770491803279, "grad_norm": 13.425204277038574, "learning_rate": 5.7142857142857145e-06, "loss": 2.8398, "step": 17 }, { "epoch": 0.05901639344262295, "grad_norm": 12.121846199035645, "learning_rate": 6.071428571428571e-06, "loss": 2.7539, "step": 18 }, { "epoch": 0.06229508196721312, "grad_norm": 12.97847843170166, "learning_rate": 6.4285714285714295e-06, "loss": 2.6719, "step": 19 }, { "epoch": 0.06557377049180328, "grad_norm": 12.588919639587402, "learning_rate": 6.785714285714287e-06, "loss": 2.6758, "step": 20 }, { "epoch": 0.06885245901639345, "grad_norm": 10.492659568786621, "learning_rate": 7.1428571428571436e-06, "loss": 2.6562, "step": 21 }, { "epoch": 0.07213114754098361, "grad_norm": 9.676924705505371, "learning_rate": 7.500000000000001e-06, "loss": 2.6094, "step": 22 }, { "epoch": 0.07540983606557378, "grad_norm": 10.222829818725586, "learning_rate": 7.857142857142858e-06, "loss": 2.6211, "step": 23 }, { "epoch": 0.07868852459016394, "grad_norm": 10.020594596862793, "learning_rate": 8.214285714285714e-06, "loss": 2.5273, "step": 24 }, { "epoch": 0.08196721311475409, "grad_norm": 9.404995918273926, "learning_rate": 8.571428571428571e-06, "loss": 2.5352, "step": 25 }, { "epoch": 0.08524590163934426, "grad_norm": 10.194731712341309, "learning_rate": 8.92857142857143e-06, "loss": 2.543, "step": 26 }, { "epoch": 0.08852459016393442, "grad_norm": 9.045639991760254, "learning_rate": 9.285714285714288e-06, "loss": 2.4961, "step": 27 }, { "epoch": 0.09180327868852459, "grad_norm": 9.108716011047363, "learning_rate": 9.642857142857144e-06, "loss": 2.4336, "step": 28 }, { "epoch": 0.09508196721311475, "grad_norm": 9.336297988891602, "learning_rate": 1e-05, "loss": 2.457, "step": 29 }, { "epoch": 0.09836065573770492, "grad_norm": 8.170738220214844, "learning_rate": 9.999968638844173e-06, "loss": 2.4688, "step": 30 }, { "epoch": 0.10163934426229508, "grad_norm": 8.507989883422852, "learning_rate": 9.999874555770099e-06, "loss": 2.4062, "step": 31 }, { "epoch": 0.10491803278688525, "grad_norm": 8.38835334777832, "learning_rate": 9.999717751958002e-06, "loss": 2.3711, "step": 32 }, { "epoch": 0.10819672131147541, "grad_norm": 8.36460018157959, "learning_rate": 9.9994982293749e-06, "loss": 2.4805, "step": 33 }, { "epoch": 0.11147540983606558, "grad_norm": 8.547989845275879, "learning_rate": 9.999215990774584e-06, "loss": 2.3906, "step": 34 }, { "epoch": 0.11475409836065574, "grad_norm": 8.938812255859375, "learning_rate": 9.998871039697589e-06, "loss": 2.3906, "step": 35 }, { "epoch": 0.1180327868852459, "grad_norm": 8.230469703674316, "learning_rate": 9.998463380471138e-06, "loss": 2.3164, "step": 36 }, { "epoch": 0.12131147540983607, "grad_norm": 7.883755683898926, "learning_rate": 9.997993018209098e-06, "loss": 2.3633, "step": 37 }, { "epoch": 0.12459016393442623, "grad_norm": 8.220748901367188, "learning_rate": 9.997459958811911e-06, "loss": 2.3516, "step": 38 }, { "epoch": 0.12786885245901639, "grad_norm": 8.738327980041504, "learning_rate": 9.996864208966522e-06, "loss": 2.3398, "step": 39 }, { "epoch": 0.13114754098360656, "grad_norm": 7.944454669952393, "learning_rate": 9.996205776146288e-06, "loss": 2.3867, "step": 40 }, { "epoch": 0.13442622950819672, "grad_norm": 7.631208419799805, "learning_rate": 9.995484668610897e-06, "loss": 2.3242, "step": 41 }, { "epoch": 0.1377049180327869, "grad_norm": 9.301526069641113, "learning_rate": 9.994700895406258e-06, "loss": 2.3203, "step": 42 }, { "epoch": 0.14098360655737704, "grad_norm": 7.625834941864014, "learning_rate": 9.99385446636438e-06, "loss": 2.3398, "step": 43 }, { "epoch": 0.14426229508196722, "grad_norm": 7.908705234527588, "learning_rate": 9.992945392103264e-06, "loss": 2.3477, "step": 44 }, { "epoch": 0.14754098360655737, "grad_norm": 8.002193450927734, "learning_rate": 9.991973684026755e-06, "loss": 2.2852, "step": 45 }, { "epoch": 0.15081967213114755, "grad_norm": 8.058524131774902, "learning_rate": 9.99093935432441e-06, "loss": 2.3711, "step": 46 }, { "epoch": 0.1540983606557377, "grad_norm": 8.803773880004883, "learning_rate": 9.989842415971338e-06, "loss": 2.2344, "step": 47 }, { "epoch": 0.15737704918032788, "grad_norm": 7.851852893829346, "learning_rate": 9.988682882728043e-06, "loss": 2.3047, "step": 48 }, { "epoch": 0.16065573770491803, "grad_norm": 8.29660415649414, "learning_rate": 9.987460769140242e-06, "loss": 2.3047, "step": 49 }, { "epoch": 0.16393442622950818, "grad_norm": 7.779457092285156, "learning_rate": 9.986176090538697e-06, "loss": 2.1992, "step": 50 }, { "epoch": 0.16721311475409836, "grad_norm": 8.602660179138184, "learning_rate": 9.984828863039008e-06, "loss": 2.25, "step": 51 }, { "epoch": 0.17049180327868851, "grad_norm": 8.079516410827637, "learning_rate": 9.983419103541421e-06, "loss": 2.3203, "step": 52 }, { "epoch": 0.1737704918032787, "grad_norm": 8.443635940551758, "learning_rate": 9.981946829730611e-06, "loss": 2.3242, "step": 53 }, { "epoch": 0.17704918032786884, "grad_norm": 7.873276710510254, "learning_rate": 9.980412060075459e-06, "loss": 2.2656, "step": 54 }, { "epoch": 0.18032786885245902, "grad_norm": 8.633429527282715, "learning_rate": 9.978814813828827e-06, "loss": 2.2227, "step": 55 }, { "epoch": 0.18360655737704917, "grad_norm": 9.135279655456543, "learning_rate": 9.97715511102731e-06, "loss": 2.2148, "step": 56 }, { "epoch": 0.18688524590163935, "grad_norm": 7.584745407104492, "learning_rate": 9.975432972490985e-06, "loss": 2.2773, "step": 57 }, { "epoch": 0.1901639344262295, "grad_norm": 9.086216926574707, "learning_rate": 9.973648419823161e-06, "loss": 2.2656, "step": 58 }, { "epoch": 0.19344262295081968, "grad_norm": 8.223559379577637, "learning_rate": 9.971801475410084e-06, "loss": 2.2773, "step": 59 }, { "epoch": 0.19672131147540983, "grad_norm": 8.726387023925781, "learning_rate": 9.969892162420682e-06, "loss": 2.3125, "step": 60 }, { "epoch": 0.2, "grad_norm": 8.304938316345215, "learning_rate": 9.96792050480626e-06, "loss": 2.168, "step": 61 }, { "epoch": 0.20327868852459016, "grad_norm": 8.20266342163086, "learning_rate": 9.965886527300201e-06, "loss": 2.25, "step": 62 }, { "epoch": 0.20655737704918034, "grad_norm": 7.833676815032959, "learning_rate": 9.963790255417663e-06, "loss": 2.2422, "step": 63 }, { "epoch": 0.2098360655737705, "grad_norm": 7.891449928283691, "learning_rate": 9.961631715455245e-06, "loss": 2.1758, "step": 64 }, { "epoch": 0.21311475409836064, "grad_norm": 8.999512672424316, "learning_rate": 9.959410934490673e-06, "loss": 2.1484, "step": 65 }, { "epoch": 0.21639344262295082, "grad_norm": 7.718410015106201, "learning_rate": 9.95712794038245e-06, "loss": 2.1172, "step": 66 }, { "epoch": 0.21967213114754097, "grad_norm": 8.272541999816895, "learning_rate": 9.954782761769509e-06, "loss": 2.25, "step": 67 }, { "epoch": 0.22295081967213115, "grad_norm": 9.607975006103516, "learning_rate": 9.952375428070853e-06, "loss": 2.2812, "step": 68 }, { "epoch": 0.2262295081967213, "grad_norm": 8.297865867614746, "learning_rate": 9.949905969485192e-06, "loss": 2.1406, "step": 69 }, { "epoch": 0.22950819672131148, "grad_norm": 8.192915916442871, "learning_rate": 9.947374416990554e-06, "loss": 2.2109, "step": 70 }, { "epoch": 0.23278688524590163, "grad_norm": 8.489333152770996, "learning_rate": 9.944780802343906e-06, "loss": 2.1914, "step": 71 }, { "epoch": 0.2360655737704918, "grad_norm": 8.244175910949707, "learning_rate": 9.942125158080747e-06, "loss": 2.1797, "step": 72 }, { "epoch": 0.23934426229508196, "grad_norm": 8.009477615356445, "learning_rate": 9.939407517514709e-06, "loss": 2.2578, "step": 73 }, { "epoch": 0.24262295081967214, "grad_norm": 8.284704208374023, "learning_rate": 9.936627914737129e-06, "loss": 2.2305, "step": 74 }, { "epoch": 0.2459016393442623, "grad_norm": 9.479103088378906, "learning_rate": 9.933786384616631e-06, "loss": 2.1406, "step": 75 }, { "epoch": 0.24918032786885247, "grad_norm": 8.558723449707031, "learning_rate": 9.930882962798683e-06, "loss": 2.2812, "step": 76 }, { "epoch": 0.25245901639344265, "grad_norm": 8.608718872070312, "learning_rate": 9.927917685705148e-06, "loss": 2.2461, "step": 77 }, { "epoch": 0.25573770491803277, "grad_norm": 7.863804817199707, "learning_rate": 9.924890590533837e-06, "loss": 2.1914, "step": 78 }, { "epoch": 0.25901639344262295, "grad_norm": 8.842366218566895, "learning_rate": 9.921801715258027e-06, "loss": 2.0996, "step": 79 }, { "epoch": 0.26229508196721313, "grad_norm": 8.273311614990234, "learning_rate": 9.918651098626e-06, "loss": 2.1641, "step": 80 }, { "epoch": 0.26557377049180325, "grad_norm": 8.011392593383789, "learning_rate": 9.915438780160547e-06, "loss": 2.168, "step": 81 }, { "epoch": 0.26885245901639343, "grad_norm": 7.946706771850586, "learning_rate": 9.912164800158474e-06, "loss": 2.2227, "step": 82 }, { "epoch": 0.2721311475409836, "grad_norm": 9.032291412353516, "learning_rate": 9.908829199690106e-06, "loss": 2.2891, "step": 83 }, { "epoch": 0.2754098360655738, "grad_norm": 8.33333969116211, "learning_rate": 9.905432020598751e-06, "loss": 2.2344, "step": 84 }, { "epoch": 0.2786885245901639, "grad_norm": 7.967901229858398, "learning_rate": 9.901973305500197e-06, "loss": 2.1172, "step": 85 }, { "epoch": 0.2819672131147541, "grad_norm": 8.24918270111084, "learning_rate": 9.898453097782164e-06, "loss": 2.1602, "step": 86 }, { "epoch": 0.28524590163934427, "grad_norm": 8.877111434936523, "learning_rate": 9.894871441603766e-06, "loss": 2.2383, "step": 87 }, { "epoch": 0.28852459016393445, "grad_norm": 8.32304573059082, "learning_rate": 9.891228381894954e-06, "loss": 2.1719, "step": 88 }, { "epoch": 0.29180327868852457, "grad_norm": 9.091806411743164, "learning_rate": 9.887523964355953e-06, "loss": 2.168, "step": 89 }, { "epoch": 0.29508196721311475, "grad_norm": 8.54736614227295, "learning_rate": 9.88375823545669e-06, "loss": 2.1211, "step": 90 }, { "epoch": 0.2983606557377049, "grad_norm": 9.3049955368042, "learning_rate": 9.879931242436208e-06, "loss": 2.1719, "step": 91 }, { "epoch": 0.3016393442622951, "grad_norm": 8.523479461669922, "learning_rate": 9.876043033302079e-06, "loss": 2.1719, "step": 92 }, { "epoch": 0.30491803278688523, "grad_norm": 7.741846561431885, "learning_rate": 9.872093656829792e-06, "loss": 2.0391, "step": 93 }, { "epoch": 0.3081967213114754, "grad_norm": 8.36119270324707, "learning_rate": 9.868083162562155e-06, "loss": 2.0938, "step": 94 }, { "epoch": 0.3114754098360656, "grad_norm": 8.163023948669434, "learning_rate": 9.864011600808663e-06, "loss": 2.0938, "step": 95 }, { "epoch": 0.31475409836065577, "grad_norm": 8.827863693237305, "learning_rate": 9.859879022644865e-06, "loss": 2.1406, "step": 96 }, { "epoch": 0.3180327868852459, "grad_norm": 8.434041023254395, "learning_rate": 9.855685479911736e-06, "loss": 2.168, "step": 97 }, { "epoch": 0.32131147540983607, "grad_norm": 8.144036293029785, "learning_rate": 9.851431025215012e-06, "loss": 2.0742, "step": 98 }, { "epoch": 0.32459016393442625, "grad_norm": 8.843683242797852, "learning_rate": 9.847115711924542e-06, "loss": 2.1016, "step": 99 }, { "epoch": 0.32786885245901637, "grad_norm": 8.300407409667969, "learning_rate": 9.842739594173607e-06, "loss": 2.1953, "step": 100 } ], "logging_steps": 1, "max_steps": 915, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.1732503551279104e+16, "train_batch_size": 512, "trial_name": null, "trial_params": null }