| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.1946587537091988, | |
| "eval_steps": 126, | |
| "global_step": 252, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0, | |
| "eval_loss": 3.176351308822632, | |
| "eval_runtime": 160.501, | |
| "eval_samples_per_second": 11.19, | |
| "eval_steps_per_second": 0.704, | |
| "step": 0 | |
| }, | |
| { | |
| "epoch": 0.004747774480712166, | |
| "grad_norm": 14.394155502319336, | |
| "learning_rate": 0.0, | |
| "loss": 3.5173, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.009495548961424332, | |
| "grad_norm": 15.094048500061035, | |
| "learning_rate": 3.174603174603175e-07, | |
| "loss": 3.7637, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.014243323442136498, | |
| "grad_norm": 15.361183166503906, | |
| "learning_rate": 6.34920634920635e-07, | |
| "loss": 3.7578, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.018991097922848664, | |
| "grad_norm": 15.74487018585205, | |
| "learning_rate": 9.523809523809525e-07, | |
| "loss": 3.7525, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.02373887240356083, | |
| "grad_norm": 13.999384880065918, | |
| "learning_rate": 1.26984126984127e-06, | |
| "loss": 3.6421, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.028486646884272996, | |
| "grad_norm": 14.453660011291504, | |
| "learning_rate": 1.5873015873015873e-06, | |
| "loss": 3.656, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.03323442136498516, | |
| "grad_norm": 14.581559181213379, | |
| "learning_rate": 1.904761904761905e-06, | |
| "loss": 3.7368, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.03798219584569733, | |
| "grad_norm": 13.600449562072754, | |
| "learning_rate": 2.222222222222222e-06, | |
| "loss": 3.5658, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.042729970326409496, | |
| "grad_norm": 11.520191192626953, | |
| "learning_rate": 2.53968253968254e-06, | |
| "loss": 3.3966, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.04747774480712166, | |
| "grad_norm": 11.848709106445312, | |
| "learning_rate": 2.8571428571428573e-06, | |
| "loss": 3.4279, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.05222551928783383, | |
| "grad_norm": 11.579791069030762, | |
| "learning_rate": 3.1746031746031746e-06, | |
| "loss": 3.4126, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.05697329376854599, | |
| "grad_norm": 10.861617088317871, | |
| "learning_rate": 3.492063492063492e-06, | |
| "loss": 3.1806, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.06172106824925816, | |
| "grad_norm": 11.614352226257324, | |
| "learning_rate": 3.80952380952381e-06, | |
| "loss": 3.0254, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.06646884272997032, | |
| "grad_norm": 11.77472972869873, | |
| "learning_rate": 4.126984126984127e-06, | |
| "loss": 2.9602, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.0712166172106825, | |
| "grad_norm": 11.360806465148926, | |
| "learning_rate": 4.444444444444444e-06, | |
| "loss": 2.9155, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.07596439169139466, | |
| "grad_norm": 12.338720321655273, | |
| "learning_rate": 4.761904761904762e-06, | |
| "loss": 2.4742, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.08071216617210683, | |
| "grad_norm": 13.61253833770752, | |
| "learning_rate": 5.07936507936508e-06, | |
| "loss": 2.4264, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.08545994065281899, | |
| "grad_norm": 13.436763763427734, | |
| "learning_rate": 5.396825396825397e-06, | |
| "loss": 2.323, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.09020771513353115, | |
| "grad_norm": 11.5920991897583, | |
| "learning_rate": 5.7142857142857145e-06, | |
| "loss": 2.0951, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.09495548961424333, | |
| "grad_norm": 10.541767120361328, | |
| "learning_rate": 6.031746031746032e-06, | |
| "loss": 1.9129, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.09970326409495549, | |
| "grad_norm": 10.552443504333496, | |
| "learning_rate": 6.349206349206349e-06, | |
| "loss": 1.7977, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.10445103857566766, | |
| "grad_norm": 10.990846633911133, | |
| "learning_rate": 6.666666666666667e-06, | |
| "loss": 1.584, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.10919881305637982, | |
| "grad_norm": 9.763855934143066, | |
| "learning_rate": 6.984126984126984e-06, | |
| "loss": 1.401, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.11394658753709198, | |
| "grad_norm": 7.580389499664307, | |
| "learning_rate": 7.301587301587301e-06, | |
| "loss": 1.2337, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.11869436201780416, | |
| "grad_norm": 7.879193305969238, | |
| "learning_rate": 7.61904761904762e-06, | |
| "loss": 1.185, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.12344213649851632, | |
| "grad_norm": 7.337639331817627, | |
| "learning_rate": 7.936507936507936e-06, | |
| "loss": 1.0778, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.1281899109792285, | |
| "grad_norm": 5.66098690032959, | |
| "learning_rate": 8.253968253968254e-06, | |
| "loss": 0.9721, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.13293768545994064, | |
| "grad_norm": 4.907323360443115, | |
| "learning_rate": 8.571428571428571e-06, | |
| "loss": 0.9584, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.13768545994065282, | |
| "grad_norm": 3.510972023010254, | |
| "learning_rate": 8.888888888888888e-06, | |
| "loss": 0.8819, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.142433234421365, | |
| "grad_norm": 3.220461368560791, | |
| "learning_rate": 9.206349206349207e-06, | |
| "loss": 0.8042, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.14718100890207717, | |
| "grad_norm": 2.63944935798645, | |
| "learning_rate": 9.523809523809525e-06, | |
| "loss": 0.7757, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.1519287833827893, | |
| "grad_norm": 11.566150665283203, | |
| "learning_rate": 9.841269841269842e-06, | |
| "loss": 0.6714, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.1566765578635015, | |
| "grad_norm": 2.1101737022399902, | |
| "learning_rate": 1.015873015873016e-05, | |
| "loss": 0.6488, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.16142433234421366, | |
| "grad_norm": 3.368835926055908, | |
| "learning_rate": 1.0476190476190477e-05, | |
| "loss": 0.6551, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.1661721068249258, | |
| "grad_norm": 2.465441942214966, | |
| "learning_rate": 1.0793650793650794e-05, | |
| "loss": 0.6039, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.17091988130563798, | |
| "grad_norm": 1.3464806079864502, | |
| "learning_rate": 1.1111111111111113e-05, | |
| "loss": 0.5671, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.17566765578635016, | |
| "grad_norm": 2.223472833633423, | |
| "learning_rate": 1.1428571428571429e-05, | |
| "loss": 0.5821, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.1804154302670623, | |
| "grad_norm": 1.5176966190338135, | |
| "learning_rate": 1.1746031746031748e-05, | |
| "loss": 0.5397, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.18516320474777448, | |
| "grad_norm": 2.944708824157715, | |
| "learning_rate": 1.2063492063492064e-05, | |
| "loss": 0.5455, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.18991097922848665, | |
| "grad_norm": 1.5701731443405151, | |
| "learning_rate": 1.2380952380952383e-05, | |
| "loss": 0.5256, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.1946587537091988, | |
| "grad_norm": 1.0416580438613892, | |
| "learning_rate": 1.2698412698412699e-05, | |
| "loss": 0.4746, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.19940652818991098, | |
| "grad_norm": 1.0439484119415283, | |
| "learning_rate": 1.3015873015873018e-05, | |
| "loss": 0.4844, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.20415430267062315, | |
| "grad_norm": 2.974339485168457, | |
| "learning_rate": 1.3333333333333333e-05, | |
| "loss": 0.509, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.20890207715133532, | |
| "grad_norm": 1.0552765130996704, | |
| "learning_rate": 1.3650793650793652e-05, | |
| "loss": 0.5047, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.21364985163204747, | |
| "grad_norm": 3.1413824558258057, | |
| "learning_rate": 1.3968253968253968e-05, | |
| "loss": 0.4924, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.21839762611275965, | |
| "grad_norm": 1.9047034978866577, | |
| "learning_rate": 1.4285714285714287e-05, | |
| "loss": 0.456, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.22314540059347182, | |
| "grad_norm": 1.2171484231948853, | |
| "learning_rate": 1.4603174603174603e-05, | |
| "loss": 0.4459, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.22789317507418397, | |
| "grad_norm": 4.339561462402344, | |
| "learning_rate": 1.4920634920634922e-05, | |
| "loss": 0.4565, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.23264094955489614, | |
| "grad_norm": 2.3773863315582275, | |
| "learning_rate": 1.523809523809524e-05, | |
| "loss": 0.4568, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.23738872403560832, | |
| "grad_norm": 1.8882341384887695, | |
| "learning_rate": 1.555555555555556e-05, | |
| "loss": 0.4452, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.24213649851632046, | |
| "grad_norm": 1.1871311664581299, | |
| "learning_rate": 1.5873015873015872e-05, | |
| "loss": 0.4238, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.24688427299703264, | |
| "grad_norm": 2.8959860801696777, | |
| "learning_rate": 1.6190476190476193e-05, | |
| "loss": 0.426, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.2516320474777448, | |
| "grad_norm": 0.9582217335700989, | |
| "learning_rate": 1.6507936507936507e-05, | |
| "loss": 0.4246, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.256379821958457, | |
| "grad_norm": 0.8898813128471375, | |
| "learning_rate": 1.6825396825396828e-05, | |
| "loss": 0.4072, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.26112759643916916, | |
| "grad_norm": 0.9023370742797852, | |
| "learning_rate": 1.7142857142857142e-05, | |
| "loss": 0.4083, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.2658753709198813, | |
| "grad_norm": 1.1952933073043823, | |
| "learning_rate": 1.7460317460317463e-05, | |
| "loss": 0.401, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.27062314540059346, | |
| "grad_norm": 1.1986902952194214, | |
| "learning_rate": 1.7777777777777777e-05, | |
| "loss": 0.4006, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.27537091988130563, | |
| "grad_norm": 0.6974486112594604, | |
| "learning_rate": 1.8095238095238097e-05, | |
| "loss": 0.4173, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.2801186943620178, | |
| "grad_norm": 1.2794380187988281, | |
| "learning_rate": 1.8412698412698415e-05, | |
| "loss": 0.3893, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.28486646884273, | |
| "grad_norm": 0.7234415411949158, | |
| "learning_rate": 1.8730158730158732e-05, | |
| "loss": 0.3904, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.28961424332344216, | |
| "grad_norm": 0.6361059546470642, | |
| "learning_rate": 1.904761904761905e-05, | |
| "loss": 0.4136, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.29436201780415433, | |
| "grad_norm": 1.7151949405670166, | |
| "learning_rate": 1.9365079365079367e-05, | |
| "loss": 0.3943, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.29910979228486645, | |
| "grad_norm": 0.6753223538398743, | |
| "learning_rate": 1.9682539682539684e-05, | |
| "loss": 0.3938, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.3038575667655786, | |
| "grad_norm": 2.0604114532470703, | |
| "learning_rate": 2e-05, | |
| "loss": 0.3909, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.3086053412462908, | |
| "grad_norm": 0.7779602408409119, | |
| "learning_rate": 1.9999846502070808e-05, | |
| "loss": 0.3661, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.313353115727003, | |
| "grad_norm": 0.7002611756324768, | |
| "learning_rate": 1.9999386012995554e-05, | |
| "loss": 0.3522, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.31810089020771515, | |
| "grad_norm": 0.6202103495597839, | |
| "learning_rate": 1.999861854691106e-05, | |
| "loss": 0.3831, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.3228486646884273, | |
| "grad_norm": 0.44983768463134766, | |
| "learning_rate": 1.9997544127378217e-05, | |
| "loss": 0.37, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.32759643916913944, | |
| "grad_norm": 0.6840509176254272, | |
| "learning_rate": 1.999616278738126e-05, | |
| "loss": 0.3751, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.3323442136498516, | |
| "grad_norm": 0.416205495595932, | |
| "learning_rate": 1.999447456932676e-05, | |
| "loss": 0.3535, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.3370919881305638, | |
| "grad_norm": 0.5574557185173035, | |
| "learning_rate": 1.9992479525042305e-05, | |
| "loss": 0.3792, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.34183976261127597, | |
| "grad_norm": 0.46201279759407043, | |
| "learning_rate": 1.9990177715774927e-05, | |
| "loss": 0.3519, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.34658753709198814, | |
| "grad_norm": 0.4977426826953888, | |
| "learning_rate": 1.9987569212189224e-05, | |
| "loss": 0.3488, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.3513353115727003, | |
| "grad_norm": 0.45472055673599243, | |
| "learning_rate": 1.9984654094365175e-05, | |
| "loss": 0.3602, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.3560830860534125, | |
| "grad_norm": 0.4730567932128906, | |
| "learning_rate": 1.9981432451795687e-05, | |
| "loss": 0.3583, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.3608308605341246, | |
| "grad_norm": 0.49239611625671387, | |
| "learning_rate": 1.997790438338385e-05, | |
| "loss": 0.3474, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.3655786350148368, | |
| "grad_norm": 0.3924044370651245, | |
| "learning_rate": 1.997406999743991e-05, | |
| "loss": 0.3538, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.37032640949554896, | |
| "grad_norm": 0.4983470141887665, | |
| "learning_rate": 1.996992941167792e-05, | |
| "loss": 0.3699, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.37507418397626113, | |
| "grad_norm": 0.4259602129459381, | |
| "learning_rate": 1.9965482753212154e-05, | |
| "loss": 0.3529, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.3798219584569733, | |
| "grad_norm": 0.3953029215335846, | |
| "learning_rate": 1.9960730158553186e-05, | |
| "loss": 0.3514, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.3845697329376855, | |
| "grad_norm": 0.7401227951049805, | |
| "learning_rate": 1.99556717736037e-05, | |
| "loss": 0.3795, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.3893175074183976, | |
| "grad_norm": 2.0949933528900146, | |
| "learning_rate": 1.9950307753654016e-05, | |
| "loss": 0.3747, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.3940652818991098, | |
| "grad_norm": 0.48020580410957336, | |
| "learning_rate": 1.9944638263377332e-05, | |
| "loss": 0.3678, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.39881305637982195, | |
| "grad_norm": 0.4131617248058319, | |
| "learning_rate": 1.9938663476824646e-05, | |
| "loss": 0.3787, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.4035608308605341, | |
| "grad_norm": 0.4024046063423157, | |
| "learning_rate": 1.9932383577419432e-05, | |
| "loss": 0.3402, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.4083086053412463, | |
| "grad_norm": 0.35955172777175903, | |
| "learning_rate": 1.9925798757952003e-05, | |
| "loss": 0.3605, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.4130563798219585, | |
| "grad_norm": 0.4338579773902893, | |
| "learning_rate": 1.9918909220573588e-05, | |
| "loss": 0.357, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.41780415430267065, | |
| "grad_norm": 0.3916882872581482, | |
| "learning_rate": 1.991171517679013e-05, | |
| "loss": 0.354, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.42255192878338277, | |
| "grad_norm": 0.39583539962768555, | |
| "learning_rate": 1.9904216847455795e-05, | |
| "loss": 0.3284, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.42729970326409494, | |
| "grad_norm": 0.43609434366226196, | |
| "learning_rate": 1.9896414462766188e-05, | |
| "loss": 0.3424, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.4320474777448071, | |
| "grad_norm": 0.4122212529182434, | |
| "learning_rate": 1.9888308262251286e-05, | |
| "loss": 0.3386, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.4367952522255193, | |
| "grad_norm": 0.3249572515487671, | |
| "learning_rate": 1.9879898494768093e-05, | |
| "loss": 0.3157, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.44154302670623147, | |
| "grad_norm": 0.745229959487915, | |
| "learning_rate": 1.9871185418492978e-05, | |
| "loss": 0.3738, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.44629080118694364, | |
| "grad_norm": 0.39437487721443176, | |
| "learning_rate": 1.9862169300913784e-05, | |
| "loss": 0.3386, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.45103857566765576, | |
| "grad_norm": 0.3622061312198639, | |
| "learning_rate": 1.985285041882158e-05, | |
| "loss": 0.3234, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.45578635014836794, | |
| "grad_norm": 0.4053417146205902, | |
| "learning_rate": 1.9843229058302192e-05, | |
| "loss": 0.3608, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.4605341246290801, | |
| "grad_norm": 0.4641744792461395, | |
| "learning_rate": 1.9833305514727396e-05, | |
| "loss": 0.3431, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.4652818991097923, | |
| "grad_norm": 0.39851275086402893, | |
| "learning_rate": 1.9823080092745878e-05, | |
| "loss": 0.3531, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.47002967359050446, | |
| "grad_norm": 0.3545701205730438, | |
| "learning_rate": 1.9812553106273848e-05, | |
| "loss": 0.3842, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.47477744807121663, | |
| "grad_norm": 0.37058180570602417, | |
| "learning_rate": 1.9801724878485438e-05, | |
| "loss": 0.3407, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.4795252225519288, | |
| "grad_norm": 0.33017316460609436, | |
| "learning_rate": 1.9790595741802757e-05, | |
| "loss": 0.3369, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.48427299703264093, | |
| "grad_norm": 0.3639221787452698, | |
| "learning_rate": 1.9779166037885692e-05, | |
| "loss": 0.3427, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.4890207715133531, | |
| "grad_norm": 0.43298575282096863, | |
| "learning_rate": 1.9767436117621416e-05, | |
| "loss": 0.3693, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.4937685459940653, | |
| "grad_norm": 0.33556169271469116, | |
| "learning_rate": 1.9755406341113622e-05, | |
| "loss": 0.3266, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.49851632047477745, | |
| "grad_norm": 0.33984947204589844, | |
| "learning_rate": 1.974307707767147e-05, | |
| "loss": 0.3309, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.5032640949554896, | |
| "grad_norm": 0.35068896412849426, | |
| "learning_rate": 1.973044870579824e-05, | |
| "loss": 0.3516, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.5080118694362018, | |
| "grad_norm": 0.3809346556663513, | |
| "learning_rate": 1.971752161317972e-05, | |
| "loss": 0.3361, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.512759643916914, | |
| "grad_norm": 0.359453946352005, | |
| "learning_rate": 1.9704296196672298e-05, | |
| "loss": 0.3737, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.5175074183976262, | |
| "grad_norm": 0.3776707947254181, | |
| "learning_rate": 1.969077286229078e-05, | |
| "loss": 0.336, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.5222551928783383, | |
| "grad_norm": 0.37052029371261597, | |
| "learning_rate": 1.9676952025195937e-05, | |
| "loss": 0.3278, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.5270029673590505, | |
| "grad_norm": 0.37363845109939575, | |
| "learning_rate": 1.966283410968174e-05, | |
| "loss": 0.3481, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.5317507418397626, | |
| "grad_norm": 0.3355433940887451, | |
| "learning_rate": 1.964841954916235e-05, | |
| "loss": 0.3379, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.5364985163204747, | |
| "grad_norm": 0.3838210105895996, | |
| "learning_rate": 1.9633708786158803e-05, | |
| "loss": 0.3388, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.5412462908011869, | |
| "grad_norm": 0.3923501670360565, | |
| "learning_rate": 1.9618702272285434e-05, | |
| "loss": 0.3749, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.5459940652818991, | |
| "grad_norm": 0.4121781885623932, | |
| "learning_rate": 1.9603400468236e-05, | |
| "loss": 0.3149, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.5507418397626113, | |
| "grad_norm": 0.31165602803230286, | |
| "learning_rate": 1.9587803843769547e-05, | |
| "loss": 0.3162, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.5554896142433234, | |
| "grad_norm": 0.3426300883293152, | |
| "learning_rate": 1.9571912877695995e-05, | |
| "loss": 0.3408, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.5602373887240356, | |
| "grad_norm": 0.33053115010261536, | |
| "learning_rate": 1.955572805786141e-05, | |
| "loss": 0.3397, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.5649851632047478, | |
| "grad_norm": 0.30746808648109436, | |
| "learning_rate": 1.9539249881133062e-05, | |
| "loss": 0.3287, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.56973293768546, | |
| "grad_norm": 0.3441820442676544, | |
| "learning_rate": 1.9522478853384154e-05, | |
| "loss": 0.3225, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.5744807121661721, | |
| "grad_norm": 0.35982808470726013, | |
| "learning_rate": 1.9505415489478293e-05, | |
| "loss": 0.3429, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.5792284866468843, | |
| "grad_norm": 0.3274133801460266, | |
| "learning_rate": 1.948806031325368e-05, | |
| "loss": 0.3436, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.5839762611275965, | |
| "grad_norm": 0.38329532742500305, | |
| "learning_rate": 1.9470413857507036e-05, | |
| "loss": 0.3211, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.5887240356083087, | |
| "grad_norm": 0.38898247480392456, | |
| "learning_rate": 1.945247666397725e-05, | |
| "loss": 0.3409, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.5934718100890207, | |
| "grad_norm": 0.3356146216392517, | |
| "learning_rate": 1.943424928332873e-05, | |
| "loss": 0.3223, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.5982195845697329, | |
| "grad_norm": 0.33519861102104187, | |
| "learning_rate": 1.9415732275134515e-05, | |
| "loss": 0.3369, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.5982195845697329, | |
| "eval_loss": 0.15260468423366547, | |
| "eval_runtime": 142.1665, | |
| "eval_samples_per_second": 12.633, | |
| "eval_steps_per_second": 0.795, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.6029673590504451, | |
| "grad_norm": 0.31465238332748413, | |
| "learning_rate": 1.9396926207859085e-05, | |
| "loss": 0.3189, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.6077151335311572, | |
| "grad_norm": 0.33517467975616455, | |
| "learning_rate": 1.937783165884092e-05, | |
| "loss": 0.3525, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.6124629080118694, | |
| "grad_norm": 0.35138949751853943, | |
| "learning_rate": 1.9358449214274763e-05, | |
| "loss": 0.3466, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.6172106824925816, | |
| "grad_norm": 0.3277546167373657, | |
| "learning_rate": 1.9338779469193638e-05, | |
| "loss": 0.3304, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.6219584569732938, | |
| "grad_norm": 5.854161262512207, | |
| "learning_rate": 1.931882302745057e-05, | |
| "loss": 0.4189, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.626706231454006, | |
| "grad_norm": 0.3688313961029053, | |
| "learning_rate": 1.9298580501700058e-05, | |
| "loss": 0.3393, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.6314540059347181, | |
| "grad_norm": 0.3257688581943512, | |
| "learning_rate": 1.9278052513379256e-05, | |
| "loss": 0.331, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.6362017804154303, | |
| "grad_norm": 0.33901330828666687, | |
| "learning_rate": 1.9257239692688907e-05, | |
| "loss": 0.342, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.6409495548961425, | |
| "grad_norm": 0.3109529912471771, | |
| "learning_rate": 1.9236142678573983e-05, | |
| "loss": 0.3151, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.6456973293768546, | |
| "grad_norm": 0.34090420603752136, | |
| "learning_rate": 1.921476211870408e-05, | |
| "loss": 0.3293, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.6504451038575668, | |
| "grad_norm": 0.3669479787349701, | |
| "learning_rate": 1.9193098669453532e-05, | |
| "loss": 0.3574, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.6551928783382789, | |
| "grad_norm": 0.3427859842777252, | |
| "learning_rate": 1.9171152995881257e-05, | |
| "loss": 0.3352, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.6599406528189911, | |
| "grad_norm": 0.3781029284000397, | |
| "learning_rate": 1.9148925771710347e-05, | |
| "loss": 0.3384, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.6646884272997032, | |
| "grad_norm": 0.4394756257534027, | |
| "learning_rate": 1.912641767930738e-05, | |
| "loss": 0.3234, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.6694362017804154, | |
| "grad_norm": 0.3640480637550354, | |
| "learning_rate": 1.9103629409661468e-05, | |
| "loss": 0.3266, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.6741839762611276, | |
| "grad_norm": 0.32769304513931274, | |
| "learning_rate": 1.908056166236305e-05, | |
| "loss": 0.3255, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.6789317507418398, | |
| "grad_norm": 0.3409084379673004, | |
| "learning_rate": 1.9057215145582418e-05, | |
| "loss": 0.3285, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.6836795252225519, | |
| "grad_norm": 0.3296876549720764, | |
| "learning_rate": 1.9033590576047967e-05, | |
| "loss": 0.3197, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.6884272997032641, | |
| "grad_norm": 0.3788384199142456, | |
| "learning_rate": 1.900968867902419e-05, | |
| "loss": 0.33, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.6931750741839763, | |
| "grad_norm": 0.31414154171943665, | |
| "learning_rate": 1.898551018828944e-05, | |
| "loss": 0.313, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.6979228486646885, | |
| "grad_norm": 0.4816839396953583, | |
| "learning_rate": 1.8961055846113358e-05, | |
| "loss": 0.3407, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.7026706231454006, | |
| "grad_norm": 0.3418005406856537, | |
| "learning_rate": 1.8936326403234125e-05, | |
| "loss": 0.3311, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.7074183976261128, | |
| "grad_norm": 0.31916749477386475, | |
| "learning_rate": 1.8911322618835393e-05, | |
| "loss": 0.3042, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.712166172106825, | |
| "grad_norm": 1.1523184776306152, | |
| "learning_rate": 1.888604526052299e-05, | |
| "loss": 0.3218, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.716913946587537, | |
| "grad_norm": 0.2973329424858093, | |
| "learning_rate": 1.8860495104301346e-05, | |
| "loss": 0.3061, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.7216617210682492, | |
| "grad_norm": 0.3566143214702606, | |
| "learning_rate": 1.8834672934549677e-05, | |
| "loss": 0.3287, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.7264094955489614, | |
| "grad_norm": 0.3308376967906952, | |
| "learning_rate": 1.8808579543997892e-05, | |
| "loss": 0.3034, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.7311572700296736, | |
| "grad_norm": 0.33520743250846863, | |
| "learning_rate": 1.8782215733702286e-05, | |
| "loss": 0.3296, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.7359050445103857, | |
| "grad_norm": 0.3926364481449127, | |
| "learning_rate": 1.8755582313020912e-05, | |
| "loss": 0.3135, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.7406528189910979, | |
| "grad_norm": 0.37003567814826965, | |
| "learning_rate": 1.8728680099588748e-05, | |
| "loss": 0.3455, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.7454005934718101, | |
| "grad_norm": 0.31512436270713806, | |
| "learning_rate": 1.870150991929261e-05, | |
| "loss": 0.3146, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.7501483679525223, | |
| "grad_norm": 0.35992729663848877, | |
| "learning_rate": 1.867407260624578e-05, | |
| "loss": 0.3138, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.7548961424332344, | |
| "grad_norm": 0.33734387159347534, | |
| "learning_rate": 1.864636900276241e-05, | |
| "loss": 0.3453, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.7596439169139466, | |
| "grad_norm": 0.403726726770401, | |
| "learning_rate": 1.8618399959331642e-05, | |
| "loss": 0.3176, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.7643916913946588, | |
| "grad_norm": 0.33617129921913147, | |
| "learning_rate": 1.8590166334591533e-05, | |
| "loss": 0.3235, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.769139465875371, | |
| "grad_norm": 0.3692370355129242, | |
| "learning_rate": 1.8561668995302668e-05, | |
| "loss": 0.3167, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.7738872403560831, | |
| "grad_norm": 0.31387221813201904, | |
| "learning_rate": 1.8532908816321557e-05, | |
| "loss": 0.2978, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.7786350148367952, | |
| "grad_norm": 0.39496850967407227, | |
| "learning_rate": 1.850388668057379e-05, | |
| "loss": 0.3167, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.7833827893175074, | |
| "grad_norm": 0.29015660285949707, | |
| "learning_rate": 1.8474603479026912e-05, | |
| "loss": 0.3111, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.7881305637982196, | |
| "grad_norm": 0.4949202835559845, | |
| "learning_rate": 1.844506011066308e-05, | |
| "loss": 0.3231, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.7928783382789317, | |
| "grad_norm": 0.3161013126373291, | |
| "learning_rate": 1.841525748245147e-05, | |
| "loss": 0.3235, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.7976261127596439, | |
| "grad_norm": 0.3269205093383789, | |
| "learning_rate": 1.8385196509320424e-05, | |
| "loss": 0.3302, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.8023738872403561, | |
| "grad_norm": 0.2890471816062927, | |
| "learning_rate": 1.8354878114129368e-05, | |
| "loss": 0.297, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.8071216617210683, | |
| "grad_norm": 0.3379496932029724, | |
| "learning_rate": 1.8324303227640472e-05, | |
| "loss": 0.3032, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.8118694362017804, | |
| "grad_norm": 0.3012542128562927, | |
| "learning_rate": 1.8293472788490096e-05, | |
| "loss": 0.3119, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.8166172106824926, | |
| "grad_norm": 0.31910914182662964, | |
| "learning_rate": 1.826238774315995e-05, | |
| "loss": 0.3249, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.8213649851632048, | |
| "grad_norm": 0.2994840741157532, | |
| "learning_rate": 1.8231049045948054e-05, | |
| "loss": 0.3106, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.826112759643917, | |
| "grad_norm": 0.31799837946891785, | |
| "learning_rate": 1.8199457658939425e-05, | |
| "loss": 0.3109, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.8308605341246291, | |
| "grad_norm": 0.3020702302455902, | |
| "learning_rate": 1.816761455197657e-05, | |
| "loss": 0.3016, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.8356083086053413, | |
| "grad_norm": 0.33007511496543884, | |
| "learning_rate": 1.8135520702629677e-05, | |
| "loss": 0.3326, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.8403560830860534, | |
| "grad_norm": 0.34143343567848206, | |
| "learning_rate": 1.8103177096166632e-05, | |
| "loss": 0.3245, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.8451038575667655, | |
| "grad_norm": 0.325300008058548, | |
| "learning_rate": 1.8070584725522763e-05, | |
| "loss": 0.32, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.8498516320474777, | |
| "grad_norm": 0.38607487082481384, | |
| "learning_rate": 1.803774459127034e-05, | |
| "loss": 0.289, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.8545994065281899, | |
| "grad_norm": 0.31634339690208435, | |
| "learning_rate": 1.8004657701587893e-05, | |
| "loss": 0.3185, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.8593471810089021, | |
| "grad_norm": 0.32501840591430664, | |
| "learning_rate": 1.7971325072229227e-05, | |
| "loss": 0.3291, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.8640949554896142, | |
| "grad_norm": 0.3348029553890228, | |
| "learning_rate": 1.7937747726492256e-05, | |
| "loss": 0.324, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.8688427299703264, | |
| "grad_norm": 0.34350311756134033, | |
| "learning_rate": 1.7903926695187595e-05, | |
| "loss": 0.3351, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.8735905044510386, | |
| "grad_norm": 0.34770363569259644, | |
| "learning_rate": 1.7869863016606893e-05, | |
| "loss": 0.3368, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.8783382789317508, | |
| "grad_norm": 0.40538617968559265, | |
| "learning_rate": 1.783555773649097e-05, | |
| "loss": 0.3027, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.8830860534124629, | |
| "grad_norm": 0.37002649903297424, | |
| "learning_rate": 1.7801011907997728e-05, | |
| "loss": 0.3296, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.8878338278931751, | |
| "grad_norm": 0.3094744384288788, | |
| "learning_rate": 1.7766226591669787e-05, | |
| "loss": 0.3211, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.8925816023738873, | |
| "grad_norm": 0.3551942706108093, | |
| "learning_rate": 1.773120285540195e-05, | |
| "loss": 0.315, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.8973293768545995, | |
| "grad_norm": 0.3153332471847534, | |
| "learning_rate": 1.7695941774408424e-05, | |
| "loss": 0.312, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.9020771513353115, | |
| "grad_norm": 0.30479609966278076, | |
| "learning_rate": 1.766044443118978e-05, | |
| "loss": 0.3176, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.9068249258160237, | |
| "grad_norm": 0.3118721842765808, | |
| "learning_rate": 1.7624711915499767e-05, | |
| "loss": 0.3366, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.9115727002967359, | |
| "grad_norm": 0.3304102122783661, | |
| "learning_rate": 1.75887453243118e-05, | |
| "loss": 0.3321, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.916320474777448, | |
| "grad_norm": 0.29333484172821045, | |
| "learning_rate": 1.755254576178535e-05, | |
| "loss": 0.306, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.9210682492581602, | |
| "grad_norm": 0.34330320358276367, | |
| "learning_rate": 1.7516114339231984e-05, | |
| "loss": 0.3173, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.9258160237388724, | |
| "grad_norm": 0.28864824771881104, | |
| "learning_rate": 1.747945217508129e-05, | |
| "loss": 0.2955, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.9305637982195846, | |
| "grad_norm": 0.32548072934150696, | |
| "learning_rate": 1.7442560394846518e-05, | |
| "loss": 0.3284, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.9353115727002967, | |
| "grad_norm": 0.3424006402492523, | |
| "learning_rate": 1.740544013109005e-05, | |
| "loss": 0.315, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.9400593471810089, | |
| "grad_norm": 0.31073182821273804, | |
| "learning_rate": 1.73680925233886e-05, | |
| "loss": 0.3186, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.9448071216617211, | |
| "grad_norm": 0.37291258573532104, | |
| "learning_rate": 1.7330518718298263e-05, | |
| "loss": 0.2982, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.9495548961424333, | |
| "grad_norm": 0.2813059091567993, | |
| "learning_rate": 1.7292719869319295e-05, | |
| "loss": 0.3189, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.9543026706231454, | |
| "grad_norm": 0.297116219997406, | |
| "learning_rate": 1.72546971368607e-05, | |
| "loss": 0.2888, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.9590504451038576, | |
| "grad_norm": 0.3253256380558014, | |
| "learning_rate": 1.7216451688204623e-05, | |
| "loss": 0.327, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.9637982195845697, | |
| "grad_norm": 0.3983827233314514, | |
| "learning_rate": 1.717798469747049e-05, | |
| "loss": 0.309, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.9685459940652819, | |
| "grad_norm": 0.30336683988571167, | |
| "learning_rate": 1.7139297345578992e-05, | |
| "loss": 0.3222, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.973293768545994, | |
| "grad_norm": 0.3853195309638977, | |
| "learning_rate": 1.7100390820215805e-05, | |
| "loss": 0.3182, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.9780415430267062, | |
| "grad_norm": 0.3312664031982422, | |
| "learning_rate": 1.7061266315795146e-05, | |
| "loss": 0.2972, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.9827893175074184, | |
| "grad_norm": 0.3694183826446533, | |
| "learning_rate": 1.7021925033423096e-05, | |
| "loss": 0.3177, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.9875370919881306, | |
| "grad_norm": 0.3218733072280884, | |
| "learning_rate": 1.698236818086073e-05, | |
| "loss": 0.304, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.9922848664688427, | |
| "grad_norm": 0.33860743045806885, | |
| "learning_rate": 1.694259697248704e-05, | |
| "loss": 0.3031, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.9970326409495549, | |
| "grad_norm": 0.42556554079055786, | |
| "learning_rate": 1.690261262926165e-05, | |
| "loss": 0.3196, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 0.38205158710479736, | |
| "learning_rate": 1.686241637868734e-05, | |
| "loss": 0.2881, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 1.004747774480712, | |
| "grad_norm": 0.36370348930358887, | |
| "learning_rate": 1.682200945477235e-05, | |
| "loss": 0.3082, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 1.0094955489614243, | |
| "grad_norm": 0.3192894160747528, | |
| "learning_rate": 1.6781393097992512e-05, | |
| "loss": 0.3233, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 1.0142433234421364, | |
| "grad_norm": 0.34595000743865967, | |
| "learning_rate": 1.6740568555253153e-05, | |
| "loss": 0.2932, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 1.0189910979228487, | |
| "grad_norm": 0.32576438784599304, | |
| "learning_rate": 1.669953707985084e-05, | |
| "loss": 0.2944, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 1.0237388724035608, | |
| "grad_norm": 0.3424221873283386, | |
| "learning_rate": 1.6658299931434857e-05, | |
| "loss": 0.305, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 1.028486646884273, | |
| "grad_norm": 0.32766616344451904, | |
| "learning_rate": 1.6616858375968596e-05, | |
| "loss": 0.3266, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 1.0332344213649851, | |
| "grad_norm": 0.34646332263946533, | |
| "learning_rate": 1.657521368569064e-05, | |
| "loss": 0.316, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 1.0379821958456974, | |
| "grad_norm": 0.30172574520111084, | |
| "learning_rate": 1.6533367139075732e-05, | |
| "loss": 0.2973, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 1.0427299703264095, | |
| "grad_norm": 0.33457738161087036, | |
| "learning_rate": 1.649132002079552e-05, | |
| "loss": 0.3015, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 1.0474777448071217, | |
| "grad_norm": 0.32709312438964844, | |
| "learning_rate": 1.6449073621679128e-05, | |
| "loss": 0.2967, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 1.0522255192878338, | |
| "grad_norm": 0.32702985405921936, | |
| "learning_rate": 1.6406629238673507e-05, | |
| "loss": 0.3375, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 1.056973293768546, | |
| "grad_norm": 0.3030819296836853, | |
| "learning_rate": 1.6363988174803638e-05, | |
| "loss": 0.3171, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 1.0617210682492582, | |
| "grad_norm": 0.3053749203681946, | |
| "learning_rate": 1.632115173913252e-05, | |
| "loss": 0.3188, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 1.0664688427299702, | |
| "grad_norm": 0.33132317662239075, | |
| "learning_rate": 1.627812124672099e-05, | |
| "loss": 0.325, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 1.0712166172106825, | |
| "grad_norm": 0.32221683859825134, | |
| "learning_rate": 1.6234898018587336e-05, | |
| "loss": 0.3249, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 1.0759643916913946, | |
| "grad_norm": 0.32629886269569397, | |
| "learning_rate": 1.619148338166677e-05, | |
| "loss": 0.308, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 1.0807121661721069, | |
| "grad_norm": 0.31524035334587097, | |
| "learning_rate": 1.614787866877066e-05, | |
| "loss": 0.304, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 1.085459940652819, | |
| "grad_norm": 0.33473050594329834, | |
| "learning_rate": 1.6104085218545633e-05, | |
| "loss": 0.3222, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 1.0902077151335312, | |
| "grad_norm": 0.28698474168777466, | |
| "learning_rate": 1.6060104375432476e-05, | |
| "loss": 0.3057, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 1.0949554896142433, | |
| "grad_norm": 0.2982139587402344, | |
| "learning_rate": 1.601593748962485e-05, | |
| "loss": 0.3074, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 1.0997032640949556, | |
| "grad_norm": 0.29852068424224854, | |
| "learning_rate": 1.5971585917027864e-05, | |
| "loss": 0.3152, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 1.1044510385756676, | |
| "grad_norm": 0.28795862197875977, | |
| "learning_rate": 1.5927051019216428e-05, | |
| "loss": 0.3084, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 1.10919881305638, | |
| "grad_norm": 0.27636635303497314, | |
| "learning_rate": 1.588233416339345e-05, | |
| "loss": 0.3079, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 1.113946587537092, | |
| "grad_norm": 0.3473345637321472, | |
| "learning_rate": 1.5837436722347902e-05, | |
| "loss": 0.2836, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 1.1186943620178043, | |
| "grad_norm": 0.3350068926811218, | |
| "learning_rate": 1.5792360074412612e-05, | |
| "loss": 0.3066, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 1.1234421364985163, | |
| "grad_norm": 0.33414971828460693, | |
| "learning_rate": 1.5747105603422013e-05, | |
| "loss": 0.3082, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 1.1281899109792284, | |
| "grad_norm": 0.32059019804000854, | |
| "learning_rate": 1.570167469866962e-05, | |
| "loss": 0.3103, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 1.1329376854599407, | |
| "grad_norm": 0.32460328936576843, | |
| "learning_rate": 1.5656068754865388e-05, | |
| "loss": 0.3151, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 1.1376854599406527, | |
| "grad_norm": 0.3252885341644287, | |
| "learning_rate": 1.561028917209291e-05, | |
| "loss": 0.3248, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 1.142433234421365, | |
| "grad_norm": 0.3324680030345917, | |
| "learning_rate": 1.5564337355766412e-05, | |
| "loss": 0.3119, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 1.147181008902077, | |
| "grad_norm": 0.3315082788467407, | |
| "learning_rate": 1.551821471658763e-05, | |
| "loss": 0.3094, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 1.1519287833827894, | |
| "grad_norm": 0.3459841310977936, | |
| "learning_rate": 1.5471922670502472e-05, | |
| "loss": 0.2982, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 1.1566765578635014, | |
| "grad_norm": 0.30599355697631836, | |
| "learning_rate": 1.5425462638657597e-05, | |
| "loss": 0.2802, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 1.1614243323442137, | |
| "grad_norm": 0.3451800048351288, | |
| "learning_rate": 1.5378836047356725e-05, | |
| "loss": 0.3065, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 1.1661721068249258, | |
| "grad_norm": 0.28921040892601013, | |
| "learning_rate": 1.5332044328016916e-05, | |
| "loss": 0.2935, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 1.170919881305638, | |
| "grad_norm": 0.3126339912414551, | |
| "learning_rate": 1.5285088917124555e-05, | |
| "loss": 0.2794, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 1.1756676557863501, | |
| "grad_norm": 0.33224689960479736, | |
| "learning_rate": 1.5237971256191325e-05, | |
| "loss": 0.3068, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 1.1804154302670624, | |
| "grad_norm": 0.33340543508529663, | |
| "learning_rate": 1.5190692791709891e-05, | |
| "loss": 0.2895, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 1.1851632047477745, | |
| "grad_norm": 0.3247314989566803, | |
| "learning_rate": 1.5143254975109538e-05, | |
| "loss": 0.3005, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 1.1899109792284865, | |
| "grad_norm": 0.3434915542602539, | |
| "learning_rate": 1.5095659262711588e-05, | |
| "loss": 0.2991, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 1.1946587537091988, | |
| "grad_norm": 0.28737980127334595, | |
| "learning_rate": 1.5047907115684695e-05, | |
| "loss": 0.2603, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 1.1946587537091988, | |
| "eval_loss": 0.15092316269874573, | |
| "eval_runtime": 136.3285, | |
| "eval_samples_per_second": 13.174, | |
| "eval_steps_per_second": 0.829, | |
| "step": 252 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 630, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 126, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 5.752167478826369e+18, | |
| "train_batch_size": 2, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |