| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.9890909090909092, | |
| "eval_steps": 500, | |
| "global_step": 153, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.019393939393939394, | |
| "grad_norm": 87.14871215820312, | |
| "learning_rate": 0.0, | |
| "loss": 2.7491, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.03878787878787879, | |
| "grad_norm": 88.4898452758789, | |
| "learning_rate": 4e-05, | |
| "loss": 2.7719, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.05818181818181818, | |
| "grad_norm": 55.55705642700195, | |
| "learning_rate": 8e-05, | |
| "loss": 1.3445, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.07757575757575758, | |
| "grad_norm": 68.65280151367188, | |
| "learning_rate": 0.00012, | |
| "loss": 2.8264, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.09696969696969697, | |
| "grad_norm": 60.984527587890625, | |
| "learning_rate": 0.00016, | |
| "loss": 1.2853, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.11636363636363636, | |
| "grad_norm": 79.5303955078125, | |
| "learning_rate": 0.0002, | |
| "loss": 1.5542, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.13575757575757577, | |
| "grad_norm": 122.48136138916016, | |
| "learning_rate": 0.00019997747161747695, | |
| "loss": 4.8128, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.15515151515151515, | |
| "grad_norm": 115.57413482666016, | |
| "learning_rate": 0.00019990989662046818, | |
| "loss": 7.7786, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.17454545454545456, | |
| "grad_norm": 106.0721664428711, | |
| "learning_rate": 0.00019979730545608126, | |
| "loss": 14.3271, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.19393939393939394, | |
| "grad_norm": 187.45887756347656, | |
| "learning_rate": 0.00019963974885425266, | |
| "loss": 10.5438, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.21333333333333335, | |
| "grad_norm": 67.8653564453125, | |
| "learning_rate": 0.00019943729780489027, | |
| "loss": 8.8936, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.23272727272727273, | |
| "grad_norm": 53.751426696777344, | |
| "learning_rate": 0.00019919004352588767, | |
| "loss": 6.9044, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.25212121212121213, | |
| "grad_norm": 68.24635314941406, | |
| "learning_rate": 0.00019889809742202455, | |
| "loss": 6.6015, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.27151515151515154, | |
| "grad_norm": 49.91704559326172, | |
| "learning_rate": 0.00019856159103477086, | |
| "loss": 4.7958, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.2909090909090909, | |
| "grad_norm": 84.79557800292969, | |
| "learning_rate": 0.0001981806759830189, | |
| "loss": 4.4889, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.3103030303030303, | |
| "grad_norm": 42.62947463989258, | |
| "learning_rate": 0.00019775552389476864, | |
| "loss": 3.852, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.3296969696969697, | |
| "grad_norm": 56.70987319946289, | |
| "learning_rate": 0.00019728632632979746, | |
| "loss": 3.7326, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.3490909090909091, | |
| "grad_norm": 48.838932037353516, | |
| "learning_rate": 0.0001967732946933499, | |
| "loss": 3.5072, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.36848484848484847, | |
| "grad_norm": 36.361454010009766, | |
| "learning_rate": 0.00019621666014088494, | |
| "loss": 3.3605, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.3878787878787879, | |
| "grad_norm": 42.21967697143555, | |
| "learning_rate": 0.00019561667347392508, | |
| "loss": 3.405, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.4072727272727273, | |
| "grad_norm": 38.328067779541016, | |
| "learning_rate": 0.0001949736050270532, | |
| "loss": 3.2796, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.4266666666666667, | |
| "grad_norm": 35.21055221557617, | |
| "learning_rate": 0.00019428774454610843, | |
| "loss": 3.1108, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.44606060606060605, | |
| "grad_norm": 36.039955139160156, | |
| "learning_rate": 0.0001935594010576362, | |
| "loss": 2.989, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.46545454545454545, | |
| "grad_norm": 46.623226165771484, | |
| "learning_rate": 0.00019278890272965096, | |
| "loss": 2.9046, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.48484848484848486, | |
| "grad_norm": 42.583335876464844, | |
| "learning_rate": 0.0001919765967237739, | |
| "loss": 2.7366, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.5042424242424243, | |
| "grad_norm": 27.625574111938477, | |
| "learning_rate": 0.0001911228490388136, | |
| "loss": 2.538, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.5236363636363637, | |
| "grad_norm": 22.885229110717773, | |
| "learning_rate": 0.00019022804434585852, | |
| "loss": 2.4008, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.5430303030303031, | |
| "grad_norm": 25.352842330932617, | |
| "learning_rate": 0.00018929258581495685, | |
| "loss": 2.2888, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.5624242424242424, | |
| "grad_norm": 25.622007369995117, | |
| "learning_rate": 0.00018831689493346095, | |
| "loss": 2.1831, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.5818181818181818, | |
| "grad_norm": 24.7369327545166, | |
| "learning_rate": 0.00018730141131611882, | |
| "loss": 2.0357, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.6012121212121212, | |
| "grad_norm": 24.530712127685547, | |
| "learning_rate": 0.00018624659250699805, | |
| "loss": 1.91, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.6206060606060606, | |
| "grad_norm": 21.335567474365234, | |
| "learning_rate": 0.00018515291377333112, | |
| "loss": 1.7888, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 21.168838500976562, | |
| "learning_rate": 0.00018402086789137546, | |
| "loss": 1.6856, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.6593939393939394, | |
| "grad_norm": 22.99504280090332, | |
| "learning_rate": 0.00018285096492438424, | |
| "loss": 1.5913, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.6787878787878788, | |
| "grad_norm": 19.430891036987305, | |
| "learning_rate": 0.00018164373199278856, | |
| "loss": 1.5296, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.6981818181818182, | |
| "grad_norm": 17.32403564453125, | |
| "learning_rate": 0.00018039971303669407, | |
| "loss": 1.4921, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.7175757575757575, | |
| "grad_norm": 13.494332313537598, | |
| "learning_rate": 0.00017911946857079888, | |
| "loss": 1.4273, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.7369696969696969, | |
| "grad_norm": 13.716304779052734, | |
| "learning_rate": 0.00017780357543184397, | |
| "loss": 1.3661, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.7563636363636363, | |
| "grad_norm": 18.049741744995117, | |
| "learning_rate": 0.00017645262651870926, | |
| "loss": 1.3225, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.7757575757575758, | |
| "grad_norm": 12.06884765625, | |
| "learning_rate": 0.00017506723052527242, | |
| "loss": 1.2614, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.7951515151515152, | |
| "grad_norm": 10.950291633605957, | |
| "learning_rate": 0.00017364801166615124, | |
| "loss": 1.2235, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.8145454545454546, | |
| "grad_norm": 10.462716102600098, | |
| "learning_rate": 0.00017219560939545246, | |
| "loss": 1.1887, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.833939393939394, | |
| "grad_norm": 14.221378326416016, | |
| "learning_rate": 0.00017071067811865476, | |
| "loss": 1.1758, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.8533333333333334, | |
| "grad_norm": 9.887701034545898, | |
| "learning_rate": 0.00016919388689775464, | |
| "loss": 1.1214, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.8727272727272727, | |
| "grad_norm": 13.127241134643555, | |
| "learning_rate": 0.0001676459191498087, | |
| "loss": 1.0835, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.8921212121212121, | |
| "grad_norm": 16.6353816986084, | |
| "learning_rate": 0.00016606747233900815, | |
| "loss": 1.0416, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.9115151515151515, | |
| "grad_norm": 14.955666542053223, | |
| "learning_rate": 0.00016445925766242391, | |
| "loss": 1.0035, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.9309090909090909, | |
| "grad_norm": 9.893986701965332, | |
| "learning_rate": 0.00016282199972956425, | |
| "loss": 0.9478, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.9503030303030303, | |
| "grad_norm": 7.6303606033325195, | |
| "learning_rate": 0.00016115643623588915, | |
| "loss": 0.9117, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.9696969696969697, | |
| "grad_norm": 6.176935195922852, | |
| "learning_rate": 0.00015946331763042867, | |
| "loss": 0.8864, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.9890909090909091, | |
| "grad_norm": 5.813427448272705, | |
| "learning_rate": 0.0001577434067776548, | |
| "loss": 0.8923, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 1.0193939393939393, | |
| "grad_norm": 5.501270294189453, | |
| "learning_rate": 0.00015599747861375955, | |
| "loss": 1.1926, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 1.0387878787878788, | |
| "grad_norm": 6.7219743728637695, | |
| "learning_rate": 0.00015422631979749354, | |
| "loss": 0.8721, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 1.0581818181818181, | |
| "grad_norm": 5.997743606567383, | |
| "learning_rate": 0.00015243072835572318, | |
| "loss": 0.8581, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 1.0775757575757576, | |
| "grad_norm": 6.01072359085083, | |
| "learning_rate": 0.00015061151332386566, | |
| "loss": 0.8375, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 1.096969696969697, | |
| "grad_norm": 7.074451923370361, | |
| "learning_rate": 0.00014876949438136347, | |
| "loss": 0.8521, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 1.1163636363636364, | |
| "grad_norm": 6.878782272338867, | |
| "learning_rate": 0.0001469055014823637, | |
| "loss": 0.8376, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 1.1357575757575757, | |
| "grad_norm": 7.401325225830078, | |
| "learning_rate": 0.00014502037448176734, | |
| "loss": 0.8258, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 1.1551515151515153, | |
| "grad_norm": 6.986009120941162, | |
| "learning_rate": 0.00014311496275681783, | |
| "loss": 0.8152, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 1.1745454545454546, | |
| "grad_norm": 3.373999834060669, | |
| "learning_rate": 0.0001411901248243993, | |
| "loss": 0.8212, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 1.1939393939393939, | |
| "grad_norm": 4.864870548248291, | |
| "learning_rate": 0.00013924672795421637, | |
| "loss": 0.8272, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 1.2133333333333334, | |
| "grad_norm": 5.978538513183594, | |
| "learning_rate": 0.00013728564777803088, | |
| "loss": 0.8253, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 1.2327272727272727, | |
| "grad_norm": 5.256051540374756, | |
| "learning_rate": 0.0001353077678951301, | |
| "loss": 0.8115, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 1.2521212121212122, | |
| "grad_norm": 4.685987949371338, | |
| "learning_rate": 0.00013331397947420576, | |
| "loss": 0.806, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 1.2715151515151515, | |
| "grad_norm": 5.210947513580322, | |
| "learning_rate": 0.00013130518085182225, | |
| "loss": 0.8114, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 1.290909090909091, | |
| "grad_norm": 7.189337730407715, | |
| "learning_rate": 0.00012928227712765504, | |
| "loss": 0.8136, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 1.3103030303030303, | |
| "grad_norm": 4.788082599639893, | |
| "learning_rate": 0.0001272461797566823, | |
| "loss": 0.798, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 1.3296969696969696, | |
| "grad_norm": 6.347202301025391, | |
| "learning_rate": 0.00012519780613851254, | |
| "loss": 0.8164, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 1.3490909090909091, | |
| "grad_norm": 4.3109307289123535, | |
| "learning_rate": 0.00012313807920403419, | |
| "loss": 0.7968, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 1.3684848484848484, | |
| "grad_norm": 4.202847480773926, | |
| "learning_rate": 0.00012106792699957263, | |
| "loss": 0.8084, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 1.387878787878788, | |
| "grad_norm": 5.670694351196289, | |
| "learning_rate": 0.00011898828226874284, | |
| "loss": 0.8105, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 1.4072727272727272, | |
| "grad_norm": 6.197960376739502, | |
| "learning_rate": 0.00011690008203218493, | |
| "loss": 0.8075, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 1.4266666666666667, | |
| "grad_norm": 5.756638526916504, | |
| "learning_rate": 0.00011480426716537315, | |
| "loss": 0.791, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 1.446060606060606, | |
| "grad_norm": 4.505667209625244, | |
| "learning_rate": 0.00011270178197468789, | |
| "loss": 0.8066, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 1.4654545454545453, | |
| "grad_norm": 4.306074142456055, | |
| "learning_rate": 0.00011059357377194161, | |
| "loss": 0.7946, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 1.4848484848484849, | |
| "grad_norm": 5.666556358337402, | |
| "learning_rate": 0.00010848059244755093, | |
| "loss": 0.8056, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 1.5042424242424244, | |
| "grad_norm": 4.2490315437316895, | |
| "learning_rate": 0.00010636379004254664, | |
| "loss": 0.8098, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 1.5236363636363637, | |
| "grad_norm": 4.633942604064941, | |
| "learning_rate": 0.00010424412031961484, | |
| "loss": 0.8032, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 1.543030303030303, | |
| "grad_norm": 4.449273109436035, | |
| "learning_rate": 0.00010212253833336237, | |
| "loss": 0.7973, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 1.5624242424242425, | |
| "grad_norm": 4.270414352416992, | |
| "learning_rate": 0.0001, | |
| "loss": 0.8055, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 1.5818181818181818, | |
| "grad_norm": 3.798518657684326, | |
| "learning_rate": 9.787746166663764e-05, | |
| "loss": 0.7913, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 1.601212121212121, | |
| "grad_norm": 4.9506425857543945, | |
| "learning_rate": 9.57558796803852e-05, | |
| "loss": 0.7878, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 1.6206060606060606, | |
| "grad_norm": 2.22216534614563, | |
| "learning_rate": 9.363620995745337e-05, | |
| "loss": 0.7962, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 1.6400000000000001, | |
| "grad_norm": 4.047158241271973, | |
| "learning_rate": 9.151940755244912e-05, | |
| "loss": 0.7892, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 1.6593939393939394, | |
| "grad_norm": 3.936147689819336, | |
| "learning_rate": 8.94064262280584e-05, | |
| "loss": 0.7937, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 1.6787878787878787, | |
| "grad_norm": 2.8039066791534424, | |
| "learning_rate": 8.729821802531212e-05, | |
| "loss": 0.7902, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 1.6981818181818182, | |
| "grad_norm": 3.999570608139038, | |
| "learning_rate": 8.519573283462687e-05, | |
| "loss": 0.7902, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 1.7175757575757575, | |
| "grad_norm": 6.398386478424072, | |
| "learning_rate": 8.309991796781511e-05, | |
| "loss": 0.7874, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 1.7369696969696968, | |
| "grad_norm": 3.3710646629333496, | |
| "learning_rate": 8.101171773125716e-05, | |
| "loss": 0.8017, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 1.7563636363636363, | |
| "grad_norm": 4.403554439544678, | |
| "learning_rate": 7.89320730004274e-05, | |
| "loss": 0.7909, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 1.7757575757575759, | |
| "grad_norm": 3.6332623958587646, | |
| "learning_rate": 7.686192079596586e-05, | |
| "loss": 0.7914, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 1.7951515151515152, | |
| "grad_norm": 2.992091655731201, | |
| "learning_rate": 7.48021938614875e-05, | |
| "loss": 0.8005, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 1.8145454545454545, | |
| "grad_norm": 3.749993324279785, | |
| "learning_rate": 7.275382024331772e-05, | |
| "loss": 0.7916, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 1.833939393939394, | |
| "grad_norm": 5.296722888946533, | |
| "learning_rate": 7.071772287234497e-05, | |
| "loss": 0.7968, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 1.8533333333333335, | |
| "grad_norm": 3.761016607284546, | |
| "learning_rate": 6.869481914817779e-05, | |
| "loss": 0.7856, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 1.8727272727272726, | |
| "grad_norm": 2.1738433837890625, | |
| "learning_rate": 6.668602052579424e-05, | |
| "loss": 0.7907, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 1.892121212121212, | |
| "grad_norm": 4.787725448608398, | |
| "learning_rate": 6.469223210486992e-05, | |
| "loss": 0.7969, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 1.9115151515151516, | |
| "grad_norm": 3.5583176612854004, | |
| "learning_rate": 6.271435222196916e-05, | |
| "loss": 0.789, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 1.930909090909091, | |
| "grad_norm": 3.7247867584228516, | |
| "learning_rate": 6.0753272045783625e-05, | |
| "loss": 0.7984, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 1.9503030303030302, | |
| "grad_norm": 3.7892508506774902, | |
| "learning_rate": 5.880987517560075e-05, | |
| "loss": 0.7951, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 1.9696969696969697, | |
| "grad_norm": 3.6910500526428223, | |
| "learning_rate": 5.688503724318217e-05, | |
| "loss": 0.7995, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 1.9890909090909092, | |
| "grad_norm": 3.318086862564087, | |
| "learning_rate": 5.497962551823266e-05, | |
| "loss": 0.7816, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 2.0193939393939395, | |
| "grad_norm": 4.396894454956055, | |
| "learning_rate": 5.309449851763633e-05, | |
| "loss": 1.092, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 2.0387878787878786, | |
| "grad_norm": 2.913602828979492, | |
| "learning_rate": 5.123050561863657e-05, | |
| "loss": 0.7868, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 2.058181818181818, | |
| "grad_norm": 5.820463180541992, | |
| "learning_rate": 4.938848667613436e-05, | |
| "loss": 0.792, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 2.0775757575757576, | |
| "grad_norm": 3.3575682640075684, | |
| "learning_rate": 4.756927164427685e-05, | |
| "loss": 0.7894, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 2.096969696969697, | |
| "grad_norm": 4.6300458908081055, | |
| "learning_rate": 4.57736802025065e-05, | |
| "loss": 0.7882, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 2.1163636363636362, | |
| "grad_norm": 3.036693572998047, | |
| "learning_rate": 4.4002521386240466e-05, | |
| "loss": 0.7934, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 2.1357575757575757, | |
| "grad_norm": 3.7198660373687744, | |
| "learning_rate": 4.2256593222345185e-05, | |
| "loss": 0.7864, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 2.1551515151515153, | |
| "grad_norm": 3.7882423400878906, | |
| "learning_rate": 4.053668236957134e-05, | |
| "loss": 0.785, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 2.174545454545455, | |
| "grad_norm": 2.843419075012207, | |
| "learning_rate": 3.884356376411089e-05, | |
| "loss": 0.7836, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 2.193939393939394, | |
| "grad_norm": 2.6506614685058594, | |
| "learning_rate": 3.717800027043576e-05, | |
| "loss": 0.7894, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 2.2133333333333334, | |
| "grad_norm": 3.587935447692871, | |
| "learning_rate": 3.554074233757608e-05, | |
| "loss": 0.7884, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 2.232727272727273, | |
| "grad_norm": 3.2883381843566895, | |
| "learning_rate": 3.393252766099187e-05, | |
| "loss": 0.7926, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 2.252121212121212, | |
| "grad_norm": 4.512854099273682, | |
| "learning_rate": 3.2354080850191324e-05, | |
| "loss": 0.7839, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 2.2715151515151515, | |
| "grad_norm": 4.273723125457764, | |
| "learning_rate": 3.080611310224539e-05, | |
| "loss": 0.7902, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 2.290909090909091, | |
| "grad_norm": 3.590773820877075, | |
| "learning_rate": 2.9289321881345254e-05, | |
| "loss": 0.7818, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 2.3103030303030305, | |
| "grad_norm": 3.1294124126434326, | |
| "learning_rate": 2.7804390604547557e-05, | |
| "loss": 0.786, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 2.3296969696969696, | |
| "grad_norm": 3.600764751434326, | |
| "learning_rate": 2.6351988333848788e-05, | |
| "loss": 0.7853, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 2.349090909090909, | |
| "grad_norm": 3.205591917037964, | |
| "learning_rate": 2.493276947472756e-05, | |
| "loss": 0.7835, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 2.3684848484848486, | |
| "grad_norm": 2.840078592300415, | |
| "learning_rate": 2.354737348129077e-05, | |
| "loss": 0.785, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 2.3878787878787877, | |
| "grad_norm": 3.47402286529541, | |
| "learning_rate": 2.2196424568156073e-05, | |
| "loss": 0.7822, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 2.4072727272727272, | |
| "grad_norm": 3.246713638305664, | |
| "learning_rate": 2.0880531429201145e-05, | |
| "loss": 0.779, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 2.4266666666666667, | |
| "grad_norm": 3.6749753952026367, | |
| "learning_rate": 1.9600286963305957e-05, | |
| "loss": 0.7974, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 2.4460606060606063, | |
| "grad_norm": 3.423552989959717, | |
| "learning_rate": 1.835626800721144e-05, | |
| "loss": 0.795, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 2.4654545454545453, | |
| "grad_norm": 5.248719215393066, | |
| "learning_rate": 1.7149035075615794e-05, | |
| "loss": 0.7816, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 2.484848484848485, | |
| "grad_norm": 2.561350107192993, | |
| "learning_rate": 1.5979132108624574e-05, | |
| "loss": 0.7886, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 2.5042424242424244, | |
| "grad_norm": 2.6997573375701904, | |
| "learning_rate": 1.4847086226668872e-05, | |
| "loss": 0.788, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 2.5236363636363635, | |
| "grad_norm": 2.400233268737793, | |
| "learning_rate": 1.3753407493001968e-05, | |
| "loss": 0.7785, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 2.543030303030303, | |
| "grad_norm": 4.139961242675781, | |
| "learning_rate": 1.2698588683881186e-05, | |
| "loss": 0.7846, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 2.5624242424242425, | |
| "grad_norm": 3.8779916763305664, | |
| "learning_rate": 1.1683105066539068e-05, | |
| "loss": 0.7877, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 2.581818181818182, | |
| "grad_norm": 3.5718817710876465, | |
| "learning_rate": 1.0707414185043163e-05, | |
| "loss": 0.7755, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 2.601212121212121, | |
| "grad_norm": 2.733354330062866, | |
| "learning_rate": 9.771955654141496e-06, | |
| "loss": 0.7923, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 2.6206060606060606, | |
| "grad_norm": 3.5972788333892822, | |
| "learning_rate": 8.87715096118642e-06, | |
| "loss": 0.7829, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "grad_norm": 2.625112533569336, | |
| "learning_rate": 8.023403276226126e-06, | |
| "loss": 0.7758, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 2.659393939393939, | |
| "grad_norm": 1.7941921949386597, | |
| "learning_rate": 7.211097270349066e-06, | |
| "loss": 0.7896, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 2.6787878787878787, | |
| "grad_norm": 3.4884462356567383, | |
| "learning_rate": 6.440598942363796e-06, | |
| "loss": 0.787, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 2.6981818181818182, | |
| "grad_norm": 3.0694332122802734, | |
| "learning_rate": 5.71225545389158e-06, | |
| "loss": 0.7787, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 2.7175757575757578, | |
| "grad_norm": 3.9330761432647705, | |
| "learning_rate": 5.026394972946813e-06, | |
| "loss": 0.7724, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 2.736969696969697, | |
| "grad_norm": 4.874517917633057, | |
| "learning_rate": 4.383326526074916e-06, | |
| "loss": 0.7737, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 2.7563636363636363, | |
| "grad_norm": 3.3836331367492676, | |
| "learning_rate": 3.783339859115065e-06, | |
| "loss": 0.7892, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 2.775757575757576, | |
| "grad_norm": 3.0882411003112793, | |
| "learning_rate": 3.226705306650113e-06, | |
| "loss": 0.7905, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 2.795151515151515, | |
| "grad_norm": 3.052604913711548, | |
| "learning_rate": 2.7136736702025433e-06, | |
| "loss": 0.7853, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 2.8145454545454545, | |
| "grad_norm": 1.9272351264953613, | |
| "learning_rate": 2.2444761052313856e-06, | |
| "loss": 0.7732, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 2.833939393939394, | |
| "grad_norm": 2.5749495029449463, | |
| "learning_rate": 1.8193240169810943e-06, | |
| "loss": 0.7846, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 2.8533333333333335, | |
| "grad_norm": 1.8724554777145386, | |
| "learning_rate": 1.4384089652291543e-06, | |
| "loss": 0.7803, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 2.8727272727272726, | |
| "grad_norm": 2.784907102584839, | |
| "learning_rate": 1.1019025779754666e-06, | |
| "loss": 0.7838, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 2.892121212121212, | |
| "grad_norm": 4.25577449798584, | |
| "learning_rate": 8.099564741123166e-07, | |
| "loss": 0.7863, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 2.9115151515151516, | |
| "grad_norm": 2.0086898803710938, | |
| "learning_rate": 5.627021951097545e-07, | |
| "loss": 0.7808, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 2.9309090909090907, | |
| "grad_norm": 2.6451563835144043, | |
| "learning_rate": 3.6025114574734785e-07, | |
| "loss": 0.7877, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 2.95030303030303, | |
| "grad_norm": 3.0813286304473877, | |
| "learning_rate": 2.0269454391874666e-07, | |
| "loss": 0.7727, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 2.9696969696969697, | |
| "grad_norm": 2.1434407234191895, | |
| "learning_rate": 9.010337953185843e-08, | |
| "loss": 0.7861, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 2.9890909090909092, | |
| "grad_norm": 2.7659671306610107, | |
| "learning_rate": 2.2528382523057113e-08, | |
| "loss": 0.7782, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 2.9890909090909092, | |
| "step": 153, | |
| "total_flos": 2.4132084696088576e+18, | |
| "train_loss": 1.5338537124247333, | |
| "train_runtime": 26135.9705, | |
| "train_samples_per_second": 0.757, | |
| "train_steps_per_second": 0.006 | |
| } | |
| ], | |
| "logging_steps": 1.0, | |
| "max_steps": 153, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 100, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 2.4132084696088576e+18, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |