| { |
| "best_metric": 0.007892865687608719, |
| "best_model_checkpoint": "/home/paperspace/Data/models/spinny/llm3br256/checkpoint-250", |
| "epoch": 2.8901734104046244, |
| "eval_steps": 5, |
| "global_step": 250, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.011560693641618497, |
| "grad_norm": 0.15129172801971436, |
| "learning_rate": 2.325581395348837e-06, |
| "loss": 0.068, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.023121387283236993, |
| "grad_norm": 0.14174357056617737, |
| "learning_rate": 4.651162790697674e-06, |
| "loss": 0.0635, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.03468208092485549, |
| "grad_norm": 0.14430448412895203, |
| "learning_rate": 6.976744186046512e-06, |
| "loss": 0.0649, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.046242774566473986, |
| "grad_norm": 0.1410493552684784, |
| "learning_rate": 9.302325581395349e-06, |
| "loss": 0.063, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.057803468208092484, |
| "grad_norm": 0.13995003700256348, |
| "learning_rate": 1.1627906976744187e-05, |
| "loss": 0.0598, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.057803468208092484, |
| "eval_loss": 0.05565303936600685, |
| "eval_runtime": 11.8576, |
| "eval_samples_per_second": 4.217, |
| "eval_steps_per_second": 1.096, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.06936416184971098, |
| "grad_norm": 0.12226737290620804, |
| "learning_rate": 1.3953488372093024e-05, |
| "loss": 0.0533, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.08092485549132948, |
| "grad_norm": 0.10596886277198792, |
| "learning_rate": 1.6279069767441862e-05, |
| "loss": 0.0544, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.09248554913294797, |
| "grad_norm": 0.07781537622213364, |
| "learning_rate": 1.8604651162790697e-05, |
| "loss": 0.0448, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.10404624277456648, |
| "grad_norm": 0.06554879248142242, |
| "learning_rate": 2.0930232558139536e-05, |
| "loss": 0.0373, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.11560693641618497, |
| "grad_norm": 0.08980869501829147, |
| "learning_rate": 2.3255813953488374e-05, |
| "loss": 0.0388, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.11560693641618497, |
| "eval_loss": 0.034586433321237564, |
| "eval_runtime": 10.0247, |
| "eval_samples_per_second": 4.988, |
| "eval_steps_per_second": 1.297, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.12716763005780346, |
| "grad_norm": 0.08015415817499161, |
| "learning_rate": 2.5581395348837212e-05, |
| "loss": 0.0393, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.13872832369942195, |
| "grad_norm": 0.06148159131407738, |
| "learning_rate": 2.7906976744186048e-05, |
| "loss": 0.0377, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.15028901734104047, |
| "grad_norm": 0.04883963614702225, |
| "learning_rate": 3.0232558139534883e-05, |
| "loss": 0.0356, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.16184971098265896, |
| "grad_norm": 0.04954546317458153, |
| "learning_rate": 3.2558139534883724e-05, |
| "loss": 0.0363, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.17341040462427745, |
| "grad_norm": 0.041509971022605896, |
| "learning_rate": 3.488372093023256e-05, |
| "loss": 0.0275, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.17341040462427745, |
| "eval_loss": 0.027632026001811028, |
| "eval_runtime": 10.0235, |
| "eval_samples_per_second": 4.988, |
| "eval_steps_per_second": 1.297, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.18497109826589594, |
| "grad_norm": 0.041887976229190826, |
| "learning_rate": 3.7209302325581394e-05, |
| "loss": 0.0303, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.19653179190751446, |
| "grad_norm": 0.03965295851230621, |
| "learning_rate": 3.953488372093023e-05, |
| "loss": 0.0284, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.20809248554913296, |
| "grad_norm": 0.03458382189273834, |
| "learning_rate": 4.186046511627907e-05, |
| "loss": 0.03, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.21965317919075145, |
| "grad_norm": 0.03138720989227295, |
| "learning_rate": 4.418604651162791e-05, |
| "loss": 0.0242, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.23121387283236994, |
| "grad_norm": 0.02681548520922661, |
| "learning_rate": 4.651162790697675e-05, |
| "loss": 0.0218, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.23121387283236994, |
| "eval_loss": 0.02280299924314022, |
| "eval_runtime": 10.0675, |
| "eval_samples_per_second": 4.967, |
| "eval_steps_per_second": 1.291, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.24277456647398843, |
| "grad_norm": 0.03174450621008873, |
| "learning_rate": 4.883720930232558e-05, |
| "loss": 0.0235, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.2543352601156069, |
| "grad_norm": 0.033451005816459656, |
| "learning_rate": 5.1162790697674425e-05, |
| "loss": 0.0291, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.2658959537572254, |
| "grad_norm": 0.03227928280830383, |
| "learning_rate": 5.348837209302326e-05, |
| "loss": 0.0254, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.2774566473988439, |
| "grad_norm": 0.03783709928393364, |
| "learning_rate": 5.5813953488372095e-05, |
| "loss": 0.0256, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.28901734104046245, |
| "grad_norm": 0.03228365629911423, |
| "learning_rate": 5.8139534883720933e-05, |
| "loss": 0.0236, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.28901734104046245, |
| "eval_loss": 0.020257851108908653, |
| "eval_runtime": 10.0419, |
| "eval_samples_per_second": 4.979, |
| "eval_steps_per_second": 1.295, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.30057803468208094, |
| "grad_norm": 0.02823326177895069, |
| "learning_rate": 6.0465116279069765e-05, |
| "loss": 0.0229, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.31213872832369943, |
| "grad_norm": 0.02882234752178192, |
| "learning_rate": 6.27906976744186e-05, |
| "loss": 0.0226, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.3236994219653179, |
| "grad_norm": 0.02527730166912079, |
| "learning_rate": 6.511627906976745e-05, |
| "loss": 0.0196, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.3352601156069364, |
| "grad_norm": 0.026926985010504723, |
| "learning_rate": 6.744186046511628e-05, |
| "loss": 0.0182, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.3468208092485549, |
| "grad_norm": 0.02383883111178875, |
| "learning_rate": 6.976744186046513e-05, |
| "loss": 0.0182, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.3468208092485549, |
| "eval_loss": 0.01791333220899105, |
| "eval_runtime": 10.1038, |
| "eval_samples_per_second": 4.949, |
| "eval_steps_per_second": 1.287, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.3583815028901734, |
| "grad_norm": 0.020905988290905952, |
| "learning_rate": 7.209302325581396e-05, |
| "loss": 0.0173, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.3699421965317919, |
| "grad_norm": 0.021837348118424416, |
| "learning_rate": 7.441860465116279e-05, |
| "loss": 0.0186, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.3815028901734104, |
| "grad_norm": 0.022534186020493507, |
| "learning_rate": 7.674418604651163e-05, |
| "loss": 0.0192, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.3930635838150289, |
| "grad_norm": 0.021404536440968513, |
| "learning_rate": 7.906976744186047e-05, |
| "loss": 0.0163, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.4046242774566474, |
| "grad_norm": 0.02087543159723282, |
| "learning_rate": 8.139534883720931e-05, |
| "loss": 0.019, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.4046242774566474, |
| "eval_loss": 0.016155000776052475, |
| "eval_runtime": 10.0056, |
| "eval_samples_per_second": 4.997, |
| "eval_steps_per_second": 1.299, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.4161849710982659, |
| "grad_norm": 0.023041941225528717, |
| "learning_rate": 8.372093023255814e-05, |
| "loss": 0.0157, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.4277456647398844, |
| "grad_norm": 0.020684899762272835, |
| "learning_rate": 8.604651162790697e-05, |
| "loss": 0.0154, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.4393063583815029, |
| "grad_norm": 0.019204530864953995, |
| "learning_rate": 8.837209302325582e-05, |
| "loss": 0.0142, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.4508670520231214, |
| "grad_norm": 0.019893651828169823, |
| "learning_rate": 9.069767441860465e-05, |
| "loss": 0.0155, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.4624277456647399, |
| "grad_norm": 0.022773414850234985, |
| "learning_rate": 9.30232558139535e-05, |
| "loss": 0.017, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.4624277456647399, |
| "eval_loss": 0.01472202967852354, |
| "eval_runtime": 10.018, |
| "eval_samples_per_second": 4.991, |
| "eval_steps_per_second": 1.298, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.47398843930635837, |
| "grad_norm": 0.021773461252450943, |
| "learning_rate": 9.534883720930233e-05, |
| "loss": 0.0161, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.48554913294797686, |
| "grad_norm": 0.023095447570085526, |
| "learning_rate": 9.767441860465116e-05, |
| "loss": 0.0146, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.49710982658959535, |
| "grad_norm": 0.0202298853546381, |
| "learning_rate": 0.0001, |
| "loss": 0.015, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.5086705202312138, |
| "grad_norm": 0.01809552311897278, |
| "learning_rate": 9.999835253787473e-05, |
| "loss": 0.0149, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.5202312138728323, |
| "grad_norm": 0.019870778545737267, |
| "learning_rate": 9.999341026006419e-05, |
| "loss": 0.0147, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.5202312138728323, |
| "eval_loss": 0.013716931454837322, |
| "eval_runtime": 10.0201, |
| "eval_samples_per_second": 4.99, |
| "eval_steps_per_second": 1.297, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.5317919075144508, |
| "grad_norm": 0.017238808795809746, |
| "learning_rate": 9.998517349225698e-05, |
| "loss": 0.0107, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.5433526011560693, |
| "grad_norm": 0.022697314620018005, |
| "learning_rate": 9.997364277724361e-05, |
| "loss": 0.0149, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.5549132947976878, |
| "grad_norm": 0.023472866043448448, |
| "learning_rate": 9.99588188748808e-05, |
| "loss": 0.0144, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.5664739884393064, |
| "grad_norm": 0.017129473388195038, |
| "learning_rate": 9.994070276204116e-05, |
| "loss": 0.0126, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.5780346820809249, |
| "grad_norm": 0.01756882295012474, |
| "learning_rate": 9.991929563254914e-05, |
| "loss": 0.0118, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.5780346820809249, |
| "eval_loss": 0.013151152059435844, |
| "eval_runtime": 10.0009, |
| "eval_samples_per_second": 5.0, |
| "eval_steps_per_second": 1.3, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.5895953757225434, |
| "grad_norm": 0.019612474367022514, |
| "learning_rate": 9.989459889710213e-05, |
| "loss": 0.0144, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.6011560693641619, |
| "grad_norm": 0.016709089279174805, |
| "learning_rate": 9.986661418317759e-05, |
| "loss": 0.0119, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.6127167630057804, |
| "grad_norm": 0.023054232820868492, |
| "learning_rate": 9.983534333492575e-05, |
| "loss": 0.0137, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.6242774566473989, |
| "grad_norm": 0.02029217965900898, |
| "learning_rate": 9.980078841304816e-05, |
| "loss": 0.014, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.6358381502890174, |
| "grad_norm": 0.018532264977693558, |
| "learning_rate": 9.976295169466178e-05, |
| "loss": 0.0107, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.6358381502890174, |
| "eval_loss": 0.012713148258626461, |
| "eval_runtime": 10.0087, |
| "eval_samples_per_second": 4.996, |
| "eval_steps_per_second": 1.299, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.6473988439306358, |
| "grad_norm": 0.020468739792704582, |
| "learning_rate": 9.97218356731491e-05, |
| "loss": 0.0146, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.6589595375722543, |
| "grad_norm": 0.01887972466647625, |
| "learning_rate": 9.967744305799357e-05, |
| "loss": 0.0126, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.6705202312138728, |
| "grad_norm": 0.02040836587548256, |
| "learning_rate": 9.962977677460132e-05, |
| "loss": 0.0126, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.6820809248554913, |
| "grad_norm": 0.02685651369392872, |
| "learning_rate": 9.957883996410821e-05, |
| "loss": 0.0122, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.6936416184971098, |
| "grad_norm": 0.02112666517496109, |
| "learning_rate": 9.952463598317285e-05, |
| "loss": 0.016, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.6936416184971098, |
| "eval_loss": 0.012294010259211063, |
| "eval_runtime": 10.0073, |
| "eval_samples_per_second": 4.996, |
| "eval_steps_per_second": 1.299, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.7052023121387283, |
| "grad_norm": 0.020788514986634254, |
| "learning_rate": 9.946716840375551e-05, |
| "loss": 0.0106, |
| "step": 61 |
| }, |
| { |
| "epoch": 0.7167630057803468, |
| "grad_norm": 0.01940411888062954, |
| "learning_rate": 9.940644101288259e-05, |
| "loss": 0.0098, |
| "step": 62 |
| }, |
| { |
| "epoch": 0.7283236994219653, |
| "grad_norm": 0.02072029560804367, |
| "learning_rate": 9.934245781239714e-05, |
| "loss": 0.0142, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.7398843930635838, |
| "grad_norm": 0.017221063375473022, |
| "learning_rate": 9.927522301869515e-05, |
| "loss": 0.0114, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.7514450867052023, |
| "grad_norm": 0.026088256388902664, |
| "learning_rate": 9.920474106244763e-05, |
| "loss": 0.0144, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.7514450867052023, |
| "eval_loss": 0.011602817103266716, |
| "eval_runtime": 10.0191, |
| "eval_samples_per_second": 4.99, |
| "eval_steps_per_second": 1.298, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.7630057803468208, |
| "grad_norm": 0.016981270164251328, |
| "learning_rate": 9.91310165883088e-05, |
| "loss": 0.0136, |
| "step": 66 |
| }, |
| { |
| "epoch": 0.7745664739884393, |
| "grad_norm": 0.017894290387630463, |
| "learning_rate": 9.905405445460972e-05, |
| "loss": 0.0113, |
| "step": 67 |
| }, |
| { |
| "epoch": 0.7861271676300579, |
| "grad_norm": 0.017901500687003136, |
| "learning_rate": 9.897385973303845e-05, |
| "loss": 0.0133, |
| "step": 68 |
| }, |
| { |
| "epoch": 0.7976878612716763, |
| "grad_norm": 0.018053261563181877, |
| "learning_rate": 9.889043770830566e-05, |
| "loss": 0.0108, |
| "step": 69 |
| }, |
| { |
| "epoch": 0.8092485549132948, |
| "grad_norm": 0.023738127201795578, |
| "learning_rate": 9.880379387779637e-05, |
| "loss": 0.0119, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.8092485549132948, |
| "eval_loss": 0.011262123472988605, |
| "eval_runtime": 10.0281, |
| "eval_samples_per_second": 4.986, |
| "eval_steps_per_second": 1.296, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.8208092485549133, |
| "grad_norm": 0.01628166250884533, |
| "learning_rate": 9.871393395120774e-05, |
| "loss": 0.014, |
| "step": 71 |
| }, |
| { |
| "epoch": 0.8323699421965318, |
| "grad_norm": 0.014171602204442024, |
| "learning_rate": 9.862086385017283e-05, |
| "loss": 0.0099, |
| "step": 72 |
| }, |
| { |
| "epoch": 0.8439306358381503, |
| "grad_norm": 0.01973879709839821, |
| "learning_rate": 9.852458970787026e-05, |
| "loss": 0.011, |
| "step": 73 |
| }, |
| { |
| "epoch": 0.8554913294797688, |
| "grad_norm": 0.015728101134300232, |
| "learning_rate": 9.842511786862019e-05, |
| "loss": 0.01, |
| "step": 74 |
| }, |
| { |
| "epoch": 0.8670520231213873, |
| "grad_norm": 0.015097449533641338, |
| "learning_rate": 9.832245488746611e-05, |
| "loss": 0.0111, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.8670520231213873, |
| "eval_loss": 0.010938560590147972, |
| "eval_runtime": 10.0124, |
| "eval_samples_per_second": 4.994, |
| "eval_steps_per_second": 1.298, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.8786127167630058, |
| "grad_norm": 0.02237934060394764, |
| "learning_rate": 9.821660752974293e-05, |
| "loss": 0.0142, |
| "step": 76 |
| }, |
| { |
| "epoch": 0.8901734104046243, |
| "grad_norm": 0.01948779821395874, |
| "learning_rate": 9.810758277063119e-05, |
| "loss": 0.0138, |
| "step": 77 |
| }, |
| { |
| "epoch": 0.9017341040462428, |
| "grad_norm": 0.022487910464406013, |
| "learning_rate": 9.799538779469734e-05, |
| "loss": 0.0151, |
| "step": 78 |
| }, |
| { |
| "epoch": 0.9132947976878613, |
| "grad_norm": 0.017546942457556725, |
| "learning_rate": 9.78800299954203e-05, |
| "loss": 0.0098, |
| "step": 79 |
| }, |
| { |
| "epoch": 0.9248554913294798, |
| "grad_norm": 0.021010980010032654, |
| "learning_rate": 9.77615169747043e-05, |
| "loss": 0.012, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.9248554913294798, |
| "eval_loss": 0.010658971033990383, |
| "eval_runtime": 10.0307, |
| "eval_samples_per_second": 4.985, |
| "eval_steps_per_second": 1.296, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.9364161849710982, |
| "grad_norm": 0.018796751275658607, |
| "learning_rate": 9.763985654237786e-05, |
| "loss": 0.0126, |
| "step": 81 |
| }, |
| { |
| "epoch": 0.9479768786127167, |
| "grad_norm": 0.020107241347432137, |
| "learning_rate": 9.751505671567913e-05, |
| "loss": 0.0129, |
| "step": 82 |
| }, |
| { |
| "epoch": 0.9595375722543352, |
| "grad_norm": 0.015596112236380577, |
| "learning_rate": 9.738712571872763e-05, |
| "loss": 0.0093, |
| "step": 83 |
| }, |
| { |
| "epoch": 0.9710982658959537, |
| "grad_norm": 0.018946580588817596, |
| "learning_rate": 9.725607198198227e-05, |
| "loss": 0.0139, |
| "step": 84 |
| }, |
| { |
| "epoch": 0.9826589595375722, |
| "grad_norm": 0.01942116767168045, |
| "learning_rate": 9.712190414168572e-05, |
| "loss": 0.0139, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.9826589595375722, |
| "eval_loss": 0.010201876051723957, |
| "eval_runtime": 10.0325, |
| "eval_samples_per_second": 4.984, |
| "eval_steps_per_second": 1.296, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.9942196531791907, |
| "grad_norm": 0.01808648556470871, |
| "learning_rate": 9.698463103929542e-05, |
| "loss": 0.0096, |
| "step": 86 |
| }, |
| { |
| "epoch": 1.0057803468208093, |
| "grad_norm": 0.033104732632637024, |
| "learning_rate": 9.684426172090085e-05, |
| "loss": 0.0208, |
| "step": 87 |
| }, |
| { |
| "epoch": 1.0173410404624277, |
| "grad_norm": 0.011730443686246872, |
| "learning_rate": 9.67008054366274e-05, |
| "loss": 0.0079, |
| "step": 88 |
| }, |
| { |
| "epoch": 1.0289017341040463, |
| "grad_norm": 0.017196012660861015, |
| "learning_rate": 9.65542716400269e-05, |
| "loss": 0.0103, |
| "step": 89 |
| }, |
| { |
| "epoch": 1.0404624277456647, |
| "grad_norm": 0.015379775315523148, |
| "learning_rate": 9.640466998745456e-05, |
| "loss": 0.0085, |
| "step": 90 |
| }, |
| { |
| "epoch": 1.0404624277456647, |
| "eval_loss": 0.01041076984256506, |
| "eval_runtime": 10.0354, |
| "eval_samples_per_second": 4.982, |
| "eval_steps_per_second": 1.295, |
| "step": 90 |
| }, |
| { |
| "epoch": 1.0520231213872833, |
| "grad_norm": 0.01682409830391407, |
| "learning_rate": 9.625201033743261e-05, |
| "loss": 0.0102, |
| "step": 91 |
| }, |
| { |
| "epoch": 1.0635838150289016, |
| "grad_norm": 0.017098382115364075, |
| "learning_rate": 9.609630275000072e-05, |
| "loss": 0.0081, |
| "step": 92 |
| }, |
| { |
| "epoch": 1.0751445086705202, |
| "grad_norm": 0.017288707196712494, |
| "learning_rate": 9.5937557486053e-05, |
| "loss": 0.0094, |
| "step": 93 |
| }, |
| { |
| "epoch": 1.0867052023121386, |
| "grad_norm": 0.018024342134594917, |
| "learning_rate": 9.577578500666187e-05, |
| "loss": 0.0116, |
| "step": 94 |
| }, |
| { |
| "epoch": 1.0982658959537572, |
| "grad_norm": 0.017229463905096054, |
| "learning_rate": 9.56109959723886e-05, |
| "loss": 0.01, |
| "step": 95 |
| }, |
| { |
| "epoch": 1.0982658959537572, |
| "eval_loss": 0.010240535251796246, |
| "eval_runtime": 10.0835, |
| "eval_samples_per_second": 4.959, |
| "eval_steps_per_second": 1.289, |
| "step": 95 |
| }, |
| { |
| "epoch": 1.1098265895953756, |
| "grad_norm": 0.015899403020739555, |
| "learning_rate": 9.544320124258092e-05, |
| "loss": 0.0099, |
| "step": 96 |
| }, |
| { |
| "epoch": 1.1213872832369942, |
| "grad_norm": 0.019960559904575348, |
| "learning_rate": 9.527241187465734e-05, |
| "loss": 0.0098, |
| "step": 97 |
| }, |
| { |
| "epoch": 1.1329479768786128, |
| "grad_norm": 0.01425469946116209, |
| "learning_rate": 9.509863912337842e-05, |
| "loss": 0.0084, |
| "step": 98 |
| }, |
| { |
| "epoch": 1.1445086705202312, |
| "grad_norm": 0.015252132900059223, |
| "learning_rate": 9.492189444010521e-05, |
| "loss": 0.01, |
| "step": 99 |
| }, |
| { |
| "epoch": 1.1560693641618498, |
| "grad_norm": 0.014408037066459656, |
| "learning_rate": 9.474218947204459e-05, |
| "loss": 0.009, |
| "step": 100 |
| }, |
| { |
| "epoch": 1.1560693641618498, |
| "eval_loss": 0.009868398308753967, |
| "eval_runtime": 10.0137, |
| "eval_samples_per_second": 4.993, |
| "eval_steps_per_second": 1.298, |
| "step": 100 |
| }, |
| { |
| "epoch": 1.1676300578034682, |
| "grad_norm": 0.01568921096622944, |
| "learning_rate": 9.455953606148172e-05, |
| "loss": 0.0125, |
| "step": 101 |
| }, |
| { |
| "epoch": 1.1791907514450868, |
| "grad_norm": 0.01693912222981453, |
| "learning_rate": 9.437394624499958e-05, |
| "loss": 0.0116, |
| "step": 102 |
| }, |
| { |
| "epoch": 1.1907514450867052, |
| "grad_norm": 0.018201276659965515, |
| "learning_rate": 9.418543225268596e-05, |
| "loss": 0.0117, |
| "step": 103 |
| }, |
| { |
| "epoch": 1.2023121387283238, |
| "grad_norm": 0.015434252098202705, |
| "learning_rate": 9.399400650732735e-05, |
| "loss": 0.011, |
| "step": 104 |
| }, |
| { |
| "epoch": 1.2138728323699421, |
| "grad_norm": 0.01606677658855915, |
| "learning_rate": 9.379968162359034e-05, |
| "loss": 0.0094, |
| "step": 105 |
| }, |
| { |
| "epoch": 1.2138728323699421, |
| "eval_loss": 0.009825030341744423, |
| "eval_runtime": 10.0305, |
| "eval_samples_per_second": 4.985, |
| "eval_steps_per_second": 1.296, |
| "step": 105 |
| }, |
| { |
| "epoch": 1.2254335260115607, |
| "grad_norm": 0.015970559790730476, |
| "learning_rate": 9.360247040719039e-05, |
| "loss": 0.0079, |
| "step": 106 |
| }, |
| { |
| "epoch": 1.2369942196531791, |
| "grad_norm": 0.01553067285567522, |
| "learning_rate": 9.340238585404788e-05, |
| "loss": 0.0077, |
| "step": 107 |
| }, |
| { |
| "epoch": 1.2485549132947977, |
| "grad_norm": 0.017365023493766785, |
| "learning_rate": 9.319944114943171e-05, |
| "loss": 0.0089, |
| "step": 108 |
| }, |
| { |
| "epoch": 1.260115606936416, |
| "grad_norm": 0.01562884822487831, |
| "learning_rate": 9.29936496670905e-05, |
| "loss": 0.0076, |
| "step": 109 |
| }, |
| { |
| "epoch": 1.2716763005780347, |
| "grad_norm": 0.016722485423088074, |
| "learning_rate": 9.278502496837116e-05, |
| "loss": 0.0069, |
| "step": 110 |
| }, |
| { |
| "epoch": 1.2716763005780347, |
| "eval_loss": 0.009917480871081352, |
| "eval_runtime": 9.9972, |
| "eval_samples_per_second": 5.001, |
| "eval_steps_per_second": 1.3, |
| "step": 110 |
| }, |
| { |
| "epoch": 1.2832369942196533, |
| "grad_norm": 0.019191304221749306, |
| "learning_rate": 9.257358080132523e-05, |
| "loss": 0.0108, |
| "step": 111 |
| }, |
| { |
| "epoch": 1.2947976878612717, |
| "grad_norm": 0.02037121169269085, |
| "learning_rate": 9.235933109980301e-05, |
| "loss": 0.011, |
| "step": 112 |
| }, |
| { |
| "epoch": 1.30635838150289, |
| "grad_norm": 0.016349676996469498, |
| "learning_rate": 9.214228998253527e-05, |
| "loss": 0.0084, |
| "step": 113 |
| }, |
| { |
| "epoch": 1.3179190751445087, |
| "grad_norm": 0.01671903021633625, |
| "learning_rate": 9.192247175220276e-05, |
| "loss": 0.0087, |
| "step": 114 |
| }, |
| { |
| "epoch": 1.3294797687861273, |
| "grad_norm": 0.01710616424679756, |
| "learning_rate": 9.16998908944939e-05, |
| "loss": 0.0108, |
| "step": 115 |
| }, |
| { |
| "epoch": 1.3294797687861273, |
| "eval_loss": 0.009632885456085205, |
| "eval_runtime": 10.0191, |
| "eval_samples_per_second": 4.99, |
| "eval_steps_per_second": 1.298, |
| "step": 115 |
| }, |
| { |
| "epoch": 1.3410404624277457, |
| "grad_norm": 0.01672513224184513, |
| "learning_rate": 9.147456207714997e-05, |
| "loss": 0.0077, |
| "step": 116 |
| }, |
| { |
| "epoch": 1.352601156069364, |
| "grad_norm": 0.017153792083263397, |
| "learning_rate": 9.124650014899867e-05, |
| "loss": 0.0101, |
| "step": 117 |
| }, |
| { |
| "epoch": 1.3641618497109826, |
| "grad_norm": 0.018158160150051117, |
| "learning_rate": 9.101572013897555e-05, |
| "loss": 0.0103, |
| "step": 118 |
| }, |
| { |
| "epoch": 1.3757225433526012, |
| "grad_norm": 0.017670975998044014, |
| "learning_rate": 9.078223725513366e-05, |
| "loss": 0.0112, |
| "step": 119 |
| }, |
| { |
| "epoch": 1.3872832369942196, |
| "grad_norm": 0.013258930295705795, |
| "learning_rate": 9.05460668836413e-05, |
| "loss": 0.0066, |
| "step": 120 |
| }, |
| { |
| "epoch": 1.3872832369942196, |
| "eval_loss": 0.009479235857725143, |
| "eval_runtime": 10.0092, |
| "eval_samples_per_second": 4.995, |
| "eval_steps_per_second": 1.299, |
| "step": 120 |
| }, |
| { |
| "epoch": 1.3988439306358382, |
| "grad_norm": 0.016367750242352486, |
| "learning_rate": 9.030722458776814e-05, |
| "loss": 0.0083, |
| "step": 121 |
| }, |
| { |
| "epoch": 1.4104046242774566, |
| "grad_norm": 0.01966327428817749, |
| "learning_rate": 9.006572610685968e-05, |
| "loss": 0.013, |
| "step": 122 |
| }, |
| { |
| "epoch": 1.4219653179190752, |
| "grad_norm": 0.014227735809981823, |
| "learning_rate": 8.98215873552999e-05, |
| "loss": 0.0081, |
| "step": 123 |
| }, |
| { |
| "epoch": 1.4335260115606936, |
| "grad_norm": 0.0126969488337636, |
| "learning_rate": 8.957482442146272e-05, |
| "loss": 0.0069, |
| "step": 124 |
| }, |
| { |
| "epoch": 1.4450867052023122, |
| "grad_norm": 0.0182335264980793, |
| "learning_rate": 8.932545356665157e-05, |
| "loss": 0.0089, |
| "step": 125 |
| }, |
| { |
| "epoch": 1.4450867052023122, |
| "eval_loss": 0.009439178742468357, |
| "eval_runtime": 10.0231, |
| "eval_samples_per_second": 4.988, |
| "eval_steps_per_second": 1.297, |
| "step": 125 |
| }, |
| { |
| "epoch": 1.4566473988439306, |
| "grad_norm": 0.013425913639366627, |
| "learning_rate": 8.907349122402801e-05, |
| "loss": 0.0064, |
| "step": 126 |
| }, |
| { |
| "epoch": 1.4682080924855492, |
| "grad_norm": 0.017706867307424545, |
| "learning_rate": 8.881895399752874e-05, |
| "loss": 0.0085, |
| "step": 127 |
| }, |
| { |
| "epoch": 1.4797687861271676, |
| "grad_norm": 0.015954311937093735, |
| "learning_rate": 8.856185866077129e-05, |
| "loss": 0.0091, |
| "step": 128 |
| }, |
| { |
| "epoch": 1.4913294797687862, |
| "grad_norm": 0.01800154522061348, |
| "learning_rate": 8.83022221559489e-05, |
| "loss": 0.0118, |
| "step": 129 |
| }, |
| { |
| "epoch": 1.5028901734104045, |
| "grad_norm": 0.022386785596609116, |
| "learning_rate": 8.80400615927139e-05, |
| "loss": 0.0084, |
| "step": 130 |
| }, |
| { |
| "epoch": 1.5028901734104045, |
| "eval_loss": 0.009285805746912956, |
| "eval_runtime": 10.0195, |
| "eval_samples_per_second": 4.99, |
| "eval_steps_per_second": 1.297, |
| "step": 130 |
| }, |
| { |
| "epoch": 1.5144508670520231, |
| "grad_norm": 0.015237067826092243, |
| "learning_rate": 8.777539424705023e-05, |
| "loss": 0.0067, |
| "step": 131 |
| }, |
| { |
| "epoch": 1.5260115606936417, |
| "grad_norm": 0.01740083284676075, |
| "learning_rate": 8.750823756013498e-05, |
| "loss": 0.0111, |
| "step": 132 |
| }, |
| { |
| "epoch": 1.5375722543352601, |
| "grad_norm": 0.01614670269191265, |
| "learning_rate": 8.72386091371891e-05, |
| "loss": 0.0066, |
| "step": 133 |
| }, |
| { |
| "epoch": 1.5491329479768785, |
| "grad_norm": 0.018181053921580315, |
| "learning_rate": 8.696652674631717e-05, |
| "loss": 0.0089, |
| "step": 134 |
| }, |
| { |
| "epoch": 1.560693641618497, |
| "grad_norm": 0.01898285746574402, |
| "learning_rate": 8.669200831733655e-05, |
| "loss": 0.0102, |
| "step": 135 |
| }, |
| { |
| "epoch": 1.560693641618497, |
| "eval_loss": 0.00925290398299694, |
| "eval_runtime": 10.0096, |
| "eval_samples_per_second": 4.995, |
| "eval_steps_per_second": 1.299, |
| "step": 135 |
| }, |
| { |
| "epoch": 1.5722543352601157, |
| "grad_norm": 0.01583796739578247, |
| "learning_rate": 8.641507194059579e-05, |
| "loss": 0.0084, |
| "step": 136 |
| }, |
| { |
| "epoch": 1.583815028901734, |
| "grad_norm": 0.020259637385606766, |
| "learning_rate": 8.613573586578262e-05, |
| "loss": 0.0097, |
| "step": 137 |
| }, |
| { |
| "epoch": 1.5953757225433525, |
| "grad_norm": 0.0170609038323164, |
| "learning_rate": 8.585401850072113e-05, |
| "loss": 0.0102, |
| "step": 138 |
| }, |
| { |
| "epoch": 1.606936416184971, |
| "grad_norm": 0.013348150067031384, |
| "learning_rate": 8.55699384101589e-05, |
| "loss": 0.0076, |
| "step": 139 |
| }, |
| { |
| "epoch": 1.6184971098265897, |
| "grad_norm": 0.0179379191249609, |
| "learning_rate": 8.528351431454351e-05, |
| "loss": 0.01, |
| "step": 140 |
| }, |
| { |
| "epoch": 1.6184971098265897, |
| "eval_loss": 0.009090203791856766, |
| "eval_runtime": 10.0116, |
| "eval_samples_per_second": 4.994, |
| "eval_steps_per_second": 1.298, |
| "step": 140 |
| }, |
| { |
| "epoch": 1.630057803468208, |
| "grad_norm": 0.014754540286958218, |
| "learning_rate": 8.499476508878893e-05, |
| "loss": 0.0091, |
| "step": 141 |
| }, |
| { |
| "epoch": 1.6416184971098264, |
| "grad_norm": 0.014680921100080013, |
| "learning_rate": 8.47037097610317e-05, |
| "loss": 0.008, |
| "step": 142 |
| }, |
| { |
| "epoch": 1.653179190751445, |
| "grad_norm": 0.01806177943944931, |
| "learning_rate": 8.441036751137696e-05, |
| "loss": 0.0122, |
| "step": 143 |
| }, |
| { |
| "epoch": 1.6647398843930636, |
| "grad_norm": 0.014901218004524708, |
| "learning_rate": 8.411475767063455e-05, |
| "loss": 0.0075, |
| "step": 144 |
| }, |
| { |
| "epoch": 1.6763005780346822, |
| "grad_norm": 0.01642463356256485, |
| "learning_rate": 8.381689971904514e-05, |
| "loss": 0.0098, |
| "step": 145 |
| }, |
| { |
| "epoch": 1.6763005780346822, |
| "eval_loss": 0.008807710371911526, |
| "eval_runtime": 10.0289, |
| "eval_samples_per_second": 4.986, |
| "eval_steps_per_second": 1.296, |
| "step": 145 |
| }, |
| { |
| "epoch": 1.6878612716763006, |
| "grad_norm": 0.0145337525755167, |
| "learning_rate": 8.35168132849965e-05, |
| "loss": 0.0074, |
| "step": 146 |
| }, |
| { |
| "epoch": 1.699421965317919, |
| "grad_norm": 0.019870057702064514, |
| "learning_rate": 8.321451814372997e-05, |
| "loss": 0.0111, |
| "step": 147 |
| }, |
| { |
| "epoch": 1.7109826589595376, |
| "grad_norm": 0.015163514763116837, |
| "learning_rate": 8.29100342160374e-05, |
| "loss": 0.0097, |
| "step": 148 |
| }, |
| { |
| "epoch": 1.7225433526011562, |
| "grad_norm": 0.01458723470568657, |
| "learning_rate": 8.260338156694836e-05, |
| "loss": 0.007, |
| "step": 149 |
| }, |
| { |
| "epoch": 1.7341040462427746, |
| "grad_norm": 0.015452525578439236, |
| "learning_rate": 8.229458040440783e-05, |
| "loss": 0.0071, |
| "step": 150 |
| }, |
| { |
| "epoch": 1.7341040462427746, |
| "eval_loss": 0.008663667365908623, |
| "eval_runtime": 10.0348, |
| "eval_samples_per_second": 4.983, |
| "eval_steps_per_second": 1.295, |
| "step": 150 |
| }, |
| { |
| "epoch": 1.745664739884393, |
| "grad_norm": 0.017789160832762718, |
| "learning_rate": 8.198365107794457e-05, |
| "loss": 0.0098, |
| "step": 151 |
| }, |
| { |
| "epoch": 1.7572254335260116, |
| "grad_norm": 0.019878627732396126, |
| "learning_rate": 8.167061407733016e-05, |
| "loss": 0.0077, |
| "step": 152 |
| }, |
| { |
| "epoch": 1.7687861271676302, |
| "grad_norm": 0.018271176144480705, |
| "learning_rate": 8.135549003122871e-05, |
| "loss": 0.0096, |
| "step": 153 |
| }, |
| { |
| "epoch": 1.7803468208092486, |
| "grad_norm": 0.017175493761897087, |
| "learning_rate": 8.103829970583742e-05, |
| "loss": 0.0111, |
| "step": 154 |
| }, |
| { |
| "epoch": 1.791907514450867, |
| "grad_norm": 0.014682111330330372, |
| "learning_rate": 8.071906400351822e-05, |
| "loss": 0.0094, |
| "step": 155 |
| }, |
| { |
| "epoch": 1.791907514450867, |
| "eval_loss": 0.008556585758924484, |
| "eval_runtime": 10.0386, |
| "eval_samples_per_second": 4.981, |
| "eval_steps_per_second": 1.295, |
| "step": 155 |
| }, |
| { |
| "epoch": 1.8034682080924855, |
| "grad_norm": 0.016677409410476685, |
| "learning_rate": 8.039780396142022e-05, |
| "loss": 0.0079, |
| "step": 156 |
| }, |
| { |
| "epoch": 1.8150289017341041, |
| "grad_norm": 0.01885165646672249, |
| "learning_rate": 8.007454075009351e-05, |
| "loss": 0.0115, |
| "step": 157 |
| }, |
| { |
| "epoch": 1.8265895953757225, |
| "grad_norm": 0.01522767636924982, |
| "learning_rate": 7.9749295672094e-05, |
| "loss": 0.0076, |
| "step": 158 |
| }, |
| { |
| "epoch": 1.838150289017341, |
| "grad_norm": 0.01811189390718937, |
| "learning_rate": 7.942209016057954e-05, |
| "loss": 0.01, |
| "step": 159 |
| }, |
| { |
| "epoch": 1.8497109826589595, |
| "grad_norm": 0.01717502437531948, |
| "learning_rate": 7.909294577789766e-05, |
| "loss": 0.008, |
| "step": 160 |
| }, |
| { |
| "epoch": 1.8497109826589595, |
| "eval_loss": 0.008559009060263634, |
| "eval_runtime": 10.0652, |
| "eval_samples_per_second": 4.968, |
| "eval_steps_per_second": 1.292, |
| "step": 160 |
| }, |
| { |
| "epoch": 1.861271676300578, |
| "grad_norm": 0.015464117750525475, |
| "learning_rate": 7.876188421416449e-05, |
| "loss": 0.0068, |
| "step": 161 |
| }, |
| { |
| "epoch": 1.8728323699421965, |
| "grad_norm": 0.015217979438602924, |
| "learning_rate": 7.842892728583558e-05, |
| "loss": 0.0074, |
| "step": 162 |
| }, |
| { |
| "epoch": 1.8843930635838149, |
| "grad_norm": 0.01554564293473959, |
| "learning_rate": 7.809409693426803e-05, |
| "loss": 0.0071, |
| "step": 163 |
| }, |
| { |
| "epoch": 1.8959537572254335, |
| "grad_norm": 0.01914595626294613, |
| "learning_rate": 7.775741522427477e-05, |
| "loss": 0.009, |
| "step": 164 |
| }, |
| { |
| "epoch": 1.907514450867052, |
| "grad_norm": 0.01934712752699852, |
| "learning_rate": 7.741890434267043e-05, |
| "loss": 0.01, |
| "step": 165 |
| }, |
| { |
| "epoch": 1.907514450867052, |
| "eval_loss": 0.00854802131652832, |
| "eval_runtime": 10.0226, |
| "eval_samples_per_second": 4.989, |
| "eval_steps_per_second": 1.297, |
| "step": 165 |
| }, |
| { |
| "epoch": 1.9190751445086707, |
| "grad_norm": 0.01625540666282177, |
| "learning_rate": 7.707858659680924e-05, |
| "loss": 0.0071, |
| "step": 166 |
| }, |
| { |
| "epoch": 1.930635838150289, |
| "grad_norm": 0.020419662818312645, |
| "learning_rate": 7.673648441311508e-05, |
| "loss": 0.0076, |
| "step": 167 |
| }, |
| { |
| "epoch": 1.9421965317919074, |
| "grad_norm": 0.012187760323286057, |
| "learning_rate": 7.639262033560359e-05, |
| "loss": 0.0064, |
| "step": 168 |
| }, |
| { |
| "epoch": 1.953757225433526, |
| "grad_norm": 0.015659835189580917, |
| "learning_rate": 7.604701702439651e-05, |
| "loss": 0.0074, |
| "step": 169 |
| }, |
| { |
| "epoch": 1.9653179190751446, |
| "grad_norm": 0.017916489392518997, |
| "learning_rate": 7.56996972542285e-05, |
| "loss": 0.0084, |
| "step": 170 |
| }, |
| { |
| "epoch": 1.9653179190751446, |
| "eval_loss": 0.008561979979276657, |
| "eval_runtime": 10.0323, |
| "eval_samples_per_second": 4.984, |
| "eval_steps_per_second": 1.296, |
| "step": 170 |
| }, |
| { |
| "epoch": 1.976878612716763, |
| "grad_norm": 0.018545281141996384, |
| "learning_rate": 7.535068391294617e-05, |
| "loss": 0.0102, |
| "step": 171 |
| }, |
| { |
| "epoch": 1.9884393063583814, |
| "grad_norm": 0.01769687980413437, |
| "learning_rate": 7.500000000000001e-05, |
| "loss": 0.0086, |
| "step": 172 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 0.021417992189526558, |
| "learning_rate": 7.464766862492855e-05, |
| "loss": 0.0092, |
| "step": 173 |
| }, |
| { |
| "epoch": 2.0115606936416186, |
| "grad_norm": 0.01275323610752821, |
| "learning_rate": 7.42937130058357e-05, |
| "loss": 0.0069, |
| "step": 174 |
| }, |
| { |
| "epoch": 2.023121387283237, |
| "grad_norm": 0.011270470917224884, |
| "learning_rate": 7.393815646786046e-05, |
| "loss": 0.0058, |
| "step": 175 |
| }, |
| { |
| "epoch": 2.023121387283237, |
| "eval_loss": 0.008691162802278996, |
| "eval_runtime": 10.0831, |
| "eval_samples_per_second": 4.959, |
| "eval_steps_per_second": 1.289, |
| "step": 175 |
| }, |
| { |
| "epoch": 2.0346820809248554, |
| "grad_norm": 0.014580226503312588, |
| "learning_rate": 7.358102244164003e-05, |
| "loss": 0.0077, |
| "step": 176 |
| }, |
| { |
| "epoch": 2.046242774566474, |
| "grad_norm": 0.014212528243660927, |
| "learning_rate": 7.322233446176571e-05, |
| "loss": 0.0074, |
| "step": 177 |
| }, |
| { |
| "epoch": 2.0578034682080926, |
| "grad_norm": 0.01766706444323063, |
| "learning_rate": 7.286211616523193e-05, |
| "loss": 0.007, |
| "step": 178 |
| }, |
| { |
| "epoch": 2.069364161849711, |
| "grad_norm": 0.015258056111633778, |
| "learning_rate": 7.250039128987873e-05, |
| "loss": 0.0073, |
| "step": 179 |
| }, |
| { |
| "epoch": 2.0809248554913293, |
| "grad_norm": 0.01634743995964527, |
| "learning_rate": 7.213718367282737e-05, |
| "loss": 0.0056, |
| "step": 180 |
| }, |
| { |
| "epoch": 2.0809248554913293, |
| "eval_loss": 0.008958614431321621, |
| "eval_runtime": 10.0215, |
| "eval_samples_per_second": 4.989, |
| "eval_steps_per_second": 1.297, |
| "step": 180 |
| }, |
| { |
| "epoch": 2.092485549132948, |
| "grad_norm": 0.014233705587685108, |
| "learning_rate": 7.177251724890956e-05, |
| "loss": 0.0056, |
| "step": 181 |
| }, |
| { |
| "epoch": 2.1040462427745665, |
| "grad_norm": 0.013312169350683689, |
| "learning_rate": 7.14064160490902e-05, |
| "loss": 0.0055, |
| "step": 182 |
| }, |
| { |
| "epoch": 2.115606936416185, |
| "grad_norm": 0.012955575250089169, |
| "learning_rate": 7.103890419888367e-05, |
| "loss": 0.0053, |
| "step": 183 |
| }, |
| { |
| "epoch": 2.1271676300578033, |
| "grad_norm": 0.018879901617765427, |
| "learning_rate": 7.067000591676416e-05, |
| "loss": 0.0065, |
| "step": 184 |
| }, |
| { |
| "epoch": 2.138728323699422, |
| "grad_norm": 0.01857951283454895, |
| "learning_rate": 7.029974551256956e-05, |
| "loss": 0.0077, |
| "step": 185 |
| }, |
| { |
| "epoch": 2.138728323699422, |
| "eval_loss": 0.008626853115856647, |
| "eval_runtime": 10.0278, |
| "eval_samples_per_second": 4.986, |
| "eval_steps_per_second": 1.296, |
| "step": 185 |
| }, |
| { |
| "epoch": 2.1502890173410405, |
| "grad_norm": 0.015720047056674957, |
| "learning_rate": 6.992814738589957e-05, |
| "loss": 0.0057, |
| "step": 186 |
| }, |
| { |
| "epoch": 2.161849710982659, |
| "grad_norm": 0.016932690516114235, |
| "learning_rate": 6.95552360245078e-05, |
| "loss": 0.0074, |
| "step": 187 |
| }, |
| { |
| "epoch": 2.1734104046242773, |
| "grad_norm": 0.013904884457588196, |
| "learning_rate": 6.918103600268799e-05, |
| "loss": 0.0054, |
| "step": 188 |
| }, |
| { |
| "epoch": 2.184971098265896, |
| "grad_norm": 0.015230257995426655, |
| "learning_rate": 6.880557197965464e-05, |
| "loss": 0.0053, |
| "step": 189 |
| }, |
| { |
| "epoch": 2.1965317919075145, |
| "grad_norm": 0.017044221982359886, |
| "learning_rate": 6.842886869791809e-05, |
| "loss": 0.0061, |
| "step": 190 |
| }, |
| { |
| "epoch": 2.1965317919075145, |
| "eval_loss": 0.00858032051473856, |
| "eval_runtime": 10.0219, |
| "eval_samples_per_second": 4.989, |
| "eval_steps_per_second": 1.297, |
| "step": 190 |
| }, |
| { |
| "epoch": 2.208092485549133, |
| "grad_norm": 0.01593642681837082, |
| "learning_rate": 6.805095098165389e-05, |
| "loss": 0.0056, |
| "step": 191 |
| }, |
| { |
| "epoch": 2.2196531791907512, |
| "grad_norm": 0.018296780064702034, |
| "learning_rate": 6.767184373506697e-05, |
| "loss": 0.008, |
| "step": 192 |
| }, |
| { |
| "epoch": 2.23121387283237, |
| "grad_norm": 0.02369534969329834, |
| "learning_rate": 6.729157194075057e-05, |
| "loss": 0.0116, |
| "step": 193 |
| }, |
| { |
| "epoch": 2.2427745664739884, |
| "grad_norm": 0.020556915551424026, |
| "learning_rate": 6.691016065803983e-05, |
| "loss": 0.0078, |
| "step": 194 |
| }, |
| { |
| "epoch": 2.254335260115607, |
| "grad_norm": 0.01714545115828514, |
| "learning_rate": 6.652763502136043e-05, |
| "loss": 0.008, |
| "step": 195 |
| }, |
| { |
| "epoch": 2.254335260115607, |
| "eval_loss": 0.008329370059072971, |
| "eval_runtime": 10.0304, |
| "eval_samples_per_second": 4.985, |
| "eval_steps_per_second": 1.296, |
| "step": 195 |
| }, |
| { |
| "epoch": 2.2658959537572256, |
| "grad_norm": 0.013920610770583153, |
| "learning_rate": 6.614402023857232e-05, |
| "loss": 0.0057, |
| "step": 196 |
| }, |
| { |
| "epoch": 2.277456647398844, |
| "grad_norm": 0.014315689913928509, |
| "learning_rate": 6.57593415893085e-05, |
| "loss": 0.0062, |
| "step": 197 |
| }, |
| { |
| "epoch": 2.2890173410404624, |
| "grad_norm": 0.01451371144503355, |
| "learning_rate": 6.537362442330916e-05, |
| "loss": 0.0059, |
| "step": 198 |
| }, |
| { |
| "epoch": 2.300578034682081, |
| "grad_norm": 0.013728981837630272, |
| "learning_rate": 6.498689415875121e-05, |
| "loss": 0.0054, |
| "step": 199 |
| }, |
| { |
| "epoch": 2.3121387283236996, |
| "grad_norm": 0.01417378056794405, |
| "learning_rate": 6.45991762805732e-05, |
| "loss": 0.0058, |
| "step": 200 |
| }, |
| { |
| "epoch": 2.3121387283236996, |
| "eval_loss": 0.00831348542124033, |
| "eval_runtime": 10.0504, |
| "eval_samples_per_second": 4.975, |
| "eval_steps_per_second": 1.293, |
| "step": 200 |
| }, |
| { |
| "epoch": 2.3236994219653178, |
| "grad_norm": 0.013795462436974049, |
| "learning_rate": 6.421049633879588e-05, |
| "loss": 0.005, |
| "step": 201 |
| }, |
| { |
| "epoch": 2.3352601156069364, |
| "grad_norm": 0.01878678984940052, |
| "learning_rate": 6.382087994683859e-05, |
| "loss": 0.0075, |
| "step": 202 |
| }, |
| { |
| "epoch": 2.346820809248555, |
| "grad_norm": 0.01882004365324974, |
| "learning_rate": 6.343035277983127e-05, |
| "loss": 0.0068, |
| "step": 203 |
| }, |
| { |
| "epoch": 2.3583815028901736, |
| "grad_norm": 0.015659485012292862, |
| "learning_rate": 6.303894057292261e-05, |
| "loss": 0.006, |
| "step": 204 |
| }, |
| { |
| "epoch": 2.3699421965317917, |
| "grad_norm": 0.015219546854496002, |
| "learning_rate": 6.264666911958404e-05, |
| "loss": 0.0047, |
| "step": 205 |
| }, |
| { |
| "epoch": 2.3699421965317917, |
| "eval_loss": 0.00836299266666174, |
| "eval_runtime": 10.0293, |
| "eval_samples_per_second": 4.985, |
| "eval_steps_per_second": 1.296, |
| "step": 205 |
| }, |
| { |
| "epoch": 2.3815028901734103, |
| "grad_norm": 0.017362525686621666, |
| "learning_rate": 6.225356426991007e-05, |
| "loss": 0.0069, |
| "step": 206 |
| }, |
| { |
| "epoch": 2.393063583815029, |
| "grad_norm": 0.014902186580002308, |
| "learning_rate": 6.185965192891472e-05, |
| "loss": 0.0055, |
| "step": 207 |
| }, |
| { |
| "epoch": 2.4046242774566475, |
| "grad_norm": 0.016910729929804802, |
| "learning_rate": 6.146495805482451e-05, |
| "loss": 0.0062, |
| "step": 208 |
| }, |
| { |
| "epoch": 2.416184971098266, |
| "grad_norm": 0.014512901194393635, |
| "learning_rate": 6.106950865736777e-05, |
| "loss": 0.0052, |
| "step": 209 |
| }, |
| { |
| "epoch": 2.4277456647398843, |
| "grad_norm": 0.017721183598041534, |
| "learning_rate": 6.0673329796060686e-05, |
| "loss": 0.0066, |
| "step": 210 |
| }, |
| { |
| "epoch": 2.4277456647398843, |
| "eval_loss": 0.008446984924376011, |
| "eval_runtime": 10.0239, |
| "eval_samples_per_second": 4.988, |
| "eval_steps_per_second": 1.297, |
| "step": 210 |
| }, |
| { |
| "epoch": 2.439306358381503, |
| "grad_norm": 0.01806679181754589, |
| "learning_rate": 6.0276447578490035e-05, |
| "loss": 0.0063, |
| "step": 211 |
| }, |
| { |
| "epoch": 2.4508670520231215, |
| "grad_norm": 0.01761380210518837, |
| "learning_rate": 5.987888815859266e-05, |
| "loss": 0.0058, |
| "step": 212 |
| }, |
| { |
| "epoch": 2.4624277456647397, |
| "grad_norm": 0.015691956505179405, |
| "learning_rate": 5.9480677734932045e-05, |
| "loss": 0.0057, |
| "step": 213 |
| }, |
| { |
| "epoch": 2.4739884393063583, |
| "grad_norm": 0.016050070524215698, |
| "learning_rate": 5.908184254897182e-05, |
| "loss": 0.0073, |
| "step": 214 |
| }, |
| { |
| "epoch": 2.485549132947977, |
| "grad_norm": 0.013452823273837566, |
| "learning_rate": 5.868240888334653e-05, |
| "loss": 0.0055, |
| "step": 215 |
| }, |
| { |
| "epoch": 2.485549132947977, |
| "eval_loss": 0.008248448371887207, |
| "eval_runtime": 10.0544, |
| "eval_samples_per_second": 4.973, |
| "eval_steps_per_second": 1.293, |
| "step": 215 |
| }, |
| { |
| "epoch": 2.4971098265895955, |
| "grad_norm": 0.013310995884239674, |
| "learning_rate": 5.8282403060129566e-05, |
| "loss": 0.0064, |
| "step": 216 |
| }, |
| { |
| "epoch": 2.508670520231214, |
| "grad_norm": 0.016307106241583824, |
| "learning_rate": 5.788185143909868e-05, |
| "loss": 0.006, |
| "step": 217 |
| }, |
| { |
| "epoch": 2.520231213872832, |
| "grad_norm": 0.014403578825294971, |
| "learning_rate": 5.7480780415998906e-05, |
| "loss": 0.0053, |
| "step": 218 |
| }, |
| { |
| "epoch": 2.531791907514451, |
| "grad_norm": 0.015278173610568047, |
| "learning_rate": 5.707921642080299e-05, |
| "loss": 0.0067, |
| "step": 219 |
| }, |
| { |
| "epoch": 2.5433526011560694, |
| "grad_norm": 0.01573382504284382, |
| "learning_rate": 5.66771859159699e-05, |
| "loss": 0.0056, |
| "step": 220 |
| }, |
| { |
| "epoch": 2.5433526011560694, |
| "eval_loss": 0.008289137855172157, |
| "eval_runtime": 10.0106, |
| "eval_samples_per_second": 4.995, |
| "eval_steps_per_second": 1.299, |
| "step": 220 |
| }, |
| { |
| "epoch": 2.5549132947976876, |
| "grad_norm": 0.014673051424324512, |
| "learning_rate": 5.6274715394700805e-05, |
| "loss": 0.0057, |
| "step": 221 |
| }, |
| { |
| "epoch": 2.5664739884393066, |
| "grad_norm": 0.017399994656443596, |
| "learning_rate": 5.587183137919332e-05, |
| "loss": 0.0066, |
| "step": 222 |
| }, |
| { |
| "epoch": 2.578034682080925, |
| "grad_norm": 0.015570064075291157, |
| "learning_rate": 5.546856041889373e-05, |
| "loss": 0.0062, |
| "step": 223 |
| }, |
| { |
| "epoch": 2.5895953757225434, |
| "grad_norm": 0.012873631902039051, |
| "learning_rate": 5.5064929088747316e-05, |
| "loss": 0.0048, |
| "step": 224 |
| }, |
| { |
| "epoch": 2.601156069364162, |
| "grad_norm": 0.01452395599335432, |
| "learning_rate": 5.46609639874473e-05, |
| "loss": 0.005, |
| "step": 225 |
| }, |
| { |
| "epoch": 2.601156069364162, |
| "eval_loss": 0.008248119615018368, |
| "eval_runtime": 10.0518, |
| "eval_samples_per_second": 4.974, |
| "eval_steps_per_second": 1.293, |
| "step": 225 |
| }, |
| { |
| "epoch": 2.61271676300578, |
| "grad_norm": 0.015882886946201324, |
| "learning_rate": 5.4256691735681786e-05, |
| "loss": 0.006, |
| "step": 226 |
| }, |
| { |
| "epoch": 2.6242774566473988, |
| "grad_norm": 0.01853213645517826, |
| "learning_rate": 5.385213897437975e-05, |
| "loss": 0.0072, |
| "step": 227 |
| }, |
| { |
| "epoch": 2.6358381502890174, |
| "grad_norm": 0.01727055385708809, |
| "learning_rate": 5.344733236295525e-05, |
| "loss": 0.0056, |
| "step": 228 |
| }, |
| { |
| "epoch": 2.647398843930636, |
| "grad_norm": 0.01715169847011566, |
| "learning_rate": 5.3042298577550696e-05, |
| "loss": 0.007, |
| "step": 229 |
| }, |
| { |
| "epoch": 2.6589595375722546, |
| "grad_norm": 0.01657768152654171, |
| "learning_rate": 5.263706430927895e-05, |
| "loss": 0.0065, |
| "step": 230 |
| }, |
| { |
| "epoch": 2.6589595375722546, |
| "eval_loss": 0.00817356538027525, |
| "eval_runtime": 10.0412, |
| "eval_samples_per_second": 4.979, |
| "eval_steps_per_second": 1.295, |
| "step": 230 |
| }, |
| { |
| "epoch": 2.6705202312138727, |
| "grad_norm": 0.016935985535383224, |
| "learning_rate": 5.223165626246432e-05, |
| "loss": 0.0056, |
| "step": 231 |
| }, |
| { |
| "epoch": 2.6820809248554913, |
| "grad_norm": 0.016983183100819588, |
| "learning_rate": 5.182610115288295e-05, |
| "loss": 0.0054, |
| "step": 232 |
| }, |
| { |
| "epoch": 2.69364161849711, |
| "grad_norm": 0.014110002666711807, |
| "learning_rate": 5.142042570600212e-05, |
| "loss": 0.0048, |
| "step": 233 |
| }, |
| { |
| "epoch": 2.705202312138728, |
| "grad_norm": 0.015584226697683334, |
| "learning_rate": 5.1014656655219197e-05, |
| "loss": 0.0074, |
| "step": 234 |
| }, |
| { |
| "epoch": 2.7167630057803467, |
| "grad_norm": 0.015623894520103931, |
| "learning_rate": 5.060882074009988e-05, |
| "loss": 0.0061, |
| "step": 235 |
| }, |
| { |
| "epoch": 2.7167630057803467, |
| "eval_loss": 0.00805756263434887, |
| "eval_runtime": 10.0337, |
| "eval_samples_per_second": 4.983, |
| "eval_steps_per_second": 1.296, |
| "step": 235 |
| }, |
| { |
| "epoch": 2.7283236994219653, |
| "grad_norm": 0.014514243230223656, |
| "learning_rate": 5.020294470461615e-05, |
| "loss": 0.005, |
| "step": 236 |
| }, |
| { |
| "epoch": 2.739884393063584, |
| "grad_norm": 0.016889233142137527, |
| "learning_rate": 4.979705529538385e-05, |
| "loss": 0.0065, |
| "step": 237 |
| }, |
| { |
| "epoch": 2.7514450867052025, |
| "grad_norm": 0.014339439570903778, |
| "learning_rate": 4.9391179259900125e-05, |
| "loss": 0.0047, |
| "step": 238 |
| }, |
| { |
| "epoch": 2.7630057803468207, |
| "grad_norm": 0.019386202096939087, |
| "learning_rate": 4.8985343344780815e-05, |
| "loss": 0.0055, |
| "step": 239 |
| }, |
| { |
| "epoch": 2.7745664739884393, |
| "grad_norm": 0.016825564205646515, |
| "learning_rate": 4.857957429399788e-05, |
| "loss": 0.0052, |
| "step": 240 |
| }, |
| { |
| "epoch": 2.7745664739884393, |
| "eval_loss": 0.008152703754603863, |
| "eval_runtime": 10.0237, |
| "eval_samples_per_second": 4.988, |
| "eval_steps_per_second": 1.297, |
| "step": 240 |
| }, |
| { |
| "epoch": 2.786127167630058, |
| "grad_norm": 0.018642200157046318, |
| "learning_rate": 4.817389884711705e-05, |
| "loss": 0.0054, |
| "step": 241 |
| }, |
| { |
| "epoch": 2.7976878612716765, |
| "grad_norm": 0.017443155869841576, |
| "learning_rate": 4.776834373753569e-05, |
| "loss": 0.0066, |
| "step": 242 |
| }, |
| { |
| "epoch": 2.809248554913295, |
| "grad_norm": 0.017855728045105934, |
| "learning_rate": 4.7362935690721076e-05, |
| "loss": 0.0064, |
| "step": 243 |
| }, |
| { |
| "epoch": 2.820809248554913, |
| "grad_norm": 0.013038886711001396, |
| "learning_rate": 4.695770142244931e-05, |
| "loss": 0.0054, |
| "step": 244 |
| }, |
| { |
| "epoch": 2.832369942196532, |
| "grad_norm": 0.0145301828160882, |
| "learning_rate": 4.655266763704476e-05, |
| "loss": 0.0053, |
| "step": 245 |
| }, |
| { |
| "epoch": 2.832369942196532, |
| "eval_loss": 0.008140629157423973, |
| "eval_runtime": 10.0264, |
| "eval_samples_per_second": 4.987, |
| "eval_steps_per_second": 1.297, |
| "step": 245 |
| }, |
| { |
| "epoch": 2.8439306358381504, |
| "grad_norm": 0.015260045416653156, |
| "learning_rate": 4.614786102562026e-05, |
| "loss": 0.0048, |
| "step": 246 |
| }, |
| { |
| "epoch": 2.8554913294797686, |
| "grad_norm": 0.014731310307979584, |
| "learning_rate": 4.574330826431821e-05, |
| "loss": 0.005, |
| "step": 247 |
| }, |
| { |
| "epoch": 2.867052023121387, |
| "grad_norm": 0.017430312931537628, |
| "learning_rate": 4.5339036012552716e-05, |
| "loss": 0.0062, |
| "step": 248 |
| }, |
| { |
| "epoch": 2.878612716763006, |
| "grad_norm": 0.01655244641005993, |
| "learning_rate": 4.493507091125269e-05, |
| "loss": 0.0056, |
| "step": 249 |
| }, |
| { |
| "epoch": 2.8901734104046244, |
| "grad_norm": 0.016315851360559464, |
| "learning_rate": 4.4531439581106295e-05, |
| "loss": 0.0058, |
| "step": 250 |
| }, |
| { |
| "epoch": 2.8901734104046244, |
| "eval_loss": 0.007892865687608719, |
| "eval_runtime": 10.0251, |
| "eval_samples_per_second": 4.988, |
| "eval_steps_per_second": 1.297, |
| "step": 250 |
| } |
| ], |
| "logging_steps": 1, |
| "max_steps": 430, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 5, |
| "save_steps": 50, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 5.930577893658132e+17, |
| "train_batch_size": 4, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|