| { |
| "best_global_step": 9770, |
| "best_metric": 4.354217052459717, |
| "best_model_checkpoint": "sindhibert_scratch/checkpoint-9770", |
| "epoch": 5.0, |
| "eval_steps": 977, |
| "global_step": 9770, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0512, |
| "grad_norm": 18.65768051147461, |
| "learning_rate": 9.900000000000002e-06, |
| "loss": 77.516953125, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.1024, |
| "grad_norm": 13.30678939819336, |
| "learning_rate": 1.9900000000000003e-05, |
| "loss": 67.7811669921875, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.1536, |
| "grad_norm": 9.597126007080078, |
| "learning_rate": 2.9900000000000002e-05, |
| "loss": 60.4075341796875, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.2048, |
| "grad_norm": 8.602116584777832, |
| "learning_rate": 3.99e-05, |
| "loss": 57.3657763671875, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.256, |
| "grad_norm": 9.814536094665527, |
| "learning_rate": 4.99e-05, |
| "loss": 56.28333984375, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.3072, |
| "grad_norm": 11.049890518188477, |
| "learning_rate": 5.99e-05, |
| "loss": 55.6001611328125, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.3584, |
| "grad_norm": 11.56328010559082, |
| "learning_rate": 6.99e-05, |
| "loss": 55.2201416015625, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.4096, |
| "grad_norm": 13.450942993164062, |
| "learning_rate": 7.99e-05, |
| "loss": 54.76716796875, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.4608, |
| "grad_norm": 11.986808776855469, |
| "learning_rate": 8.99e-05, |
| "loss": 54.2532373046875, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.500224, |
| "eval_loss": 6.677970886230469, |
| "eval_runtime": 53.9539, |
| "eval_samples_per_second": 370.687, |
| "eval_steps_per_second": 11.584, |
| "step": 977 |
| }, |
| { |
| "epoch": 0.512, |
| "grad_norm": 11.47065258026123, |
| "learning_rate": 9.99e-05, |
| "loss": 53.8267431640625, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.5632, |
| "grad_norm": 14.508301734924316, |
| "learning_rate": 9.887115165336375e-05, |
| "loss": 53.4335546875, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.6144, |
| "grad_norm": 13.515727043151855, |
| "learning_rate": 9.773090079817561e-05, |
| "loss": 53.126240234375, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.6656, |
| "grad_norm": 13.126218795776367, |
| "learning_rate": 9.659064994298746e-05, |
| "loss": 52.5950341796875, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.7168, |
| "grad_norm": 21.854541778564453, |
| "learning_rate": 9.545039908779932e-05, |
| "loss": 52.1828076171875, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.768, |
| "grad_norm": 18.417003631591797, |
| "learning_rate": 9.431014823261119e-05, |
| "loss": 51.821767578125, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.8192, |
| "grad_norm": 12.346606254577637, |
| "learning_rate": 9.316989737742304e-05, |
| "loss": 51.36373046875, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.8704, |
| "grad_norm": 22.9099063873291, |
| "learning_rate": 9.202964652223489e-05, |
| "loss": 50.9531591796875, |
| "step": 1700 |
| }, |
| { |
| "epoch": 0.9216, |
| "grad_norm": 18.55419158935547, |
| "learning_rate": 9.088939566704675e-05, |
| "loss": 50.487490234375, |
| "step": 1800 |
| }, |
| { |
| "epoch": 0.9728, |
| "grad_norm": 16.247209548950195, |
| "learning_rate": 8.974914481185861e-05, |
| "loss": 50.037265625, |
| "step": 1900 |
| }, |
| { |
| "epoch": 1.0, |
| "eval_loss": 6.126930236816406, |
| "eval_runtime": 54.5716, |
| "eval_samples_per_second": 366.491, |
| "eval_steps_per_second": 11.453, |
| "step": 1954 |
| }, |
| { |
| "epoch": 1.023552, |
| "grad_norm": 15.577176094055176, |
| "learning_rate": 8.860889395667046e-05, |
| "loss": 49.1747216796875, |
| "step": 2000 |
| }, |
| { |
| "epoch": 1.074752, |
| "grad_norm": 14.534530639648438, |
| "learning_rate": 8.746864310148233e-05, |
| "loss": 48.833271484375, |
| "step": 2100 |
| }, |
| { |
| "epoch": 1.125952, |
| "grad_norm": 21.150440216064453, |
| "learning_rate": 8.632839224629419e-05, |
| "loss": 47.9638330078125, |
| "step": 2200 |
| }, |
| { |
| "epoch": 1.177152, |
| "grad_norm": 14.35093879699707, |
| "learning_rate": 8.518814139110604e-05, |
| "loss": 47.6843505859375, |
| "step": 2300 |
| }, |
| { |
| "epoch": 1.228352, |
| "grad_norm": 18.37192726135254, |
| "learning_rate": 8.40478905359179e-05, |
| "loss": 47.1163623046875, |
| "step": 2400 |
| }, |
| { |
| "epoch": 1.279552, |
| "grad_norm": 15.366902351379395, |
| "learning_rate": 8.290763968072977e-05, |
| "loss": 46.6763720703125, |
| "step": 2500 |
| }, |
| { |
| "epoch": 1.330752, |
| "grad_norm": 18.58373260498047, |
| "learning_rate": 8.176738882554162e-05, |
| "loss": 46.071142578125, |
| "step": 2600 |
| }, |
| { |
| "epoch": 1.381952, |
| "grad_norm": 16.35076141357422, |
| "learning_rate": 8.062713797035348e-05, |
| "loss": 45.5451220703125, |
| "step": 2700 |
| }, |
| { |
| "epoch": 1.433152, |
| "grad_norm": 18.8570556640625, |
| "learning_rate": 7.948688711516535e-05, |
| "loss": 45.1469384765625, |
| "step": 2800 |
| }, |
| { |
| "epoch": 1.484352, |
| "grad_norm": 17.72637367248535, |
| "learning_rate": 7.83466362599772e-05, |
| "loss": 44.6100927734375, |
| "step": 2900 |
| }, |
| { |
| "epoch": 1.500224, |
| "eval_loss": 5.475055694580078, |
| "eval_runtime": 54.4157, |
| "eval_samples_per_second": 367.541, |
| "eval_steps_per_second": 11.486, |
| "step": 2931 |
| }, |
| { |
| "epoch": 1.535552, |
| "grad_norm": 14.798540115356445, |
| "learning_rate": 7.720638540478906e-05, |
| "loss": 44.670400390625, |
| "step": 3000 |
| }, |
| { |
| "epoch": 1.5867520000000002, |
| "grad_norm": 17.05501365661621, |
| "learning_rate": 7.606613454960093e-05, |
| "loss": 44.11970703125, |
| "step": 3100 |
| }, |
| { |
| "epoch": 1.6379519999999999, |
| "grad_norm": 15.10769271850586, |
| "learning_rate": 7.492588369441278e-05, |
| "loss": 43.6269775390625, |
| "step": 3200 |
| }, |
| { |
| "epoch": 1.689152, |
| "grad_norm": 16.199745178222656, |
| "learning_rate": 7.378563283922463e-05, |
| "loss": 43.4040771484375, |
| "step": 3300 |
| }, |
| { |
| "epoch": 1.7403520000000001, |
| "grad_norm": 19.104358673095703, |
| "learning_rate": 7.264538198403649e-05, |
| "loss": 42.9161328125, |
| "step": 3400 |
| }, |
| { |
| "epoch": 1.791552, |
| "grad_norm": 17.44623374938965, |
| "learning_rate": 7.150513112884834e-05, |
| "loss": 42.778564453125, |
| "step": 3500 |
| }, |
| { |
| "epoch": 1.842752, |
| "grad_norm": 16.97149658203125, |
| "learning_rate": 7.03648802736602e-05, |
| "loss": 42.4598193359375, |
| "step": 3600 |
| }, |
| { |
| "epoch": 1.893952, |
| "grad_norm": 14.990788459777832, |
| "learning_rate": 6.922462941847207e-05, |
| "loss": 42.2105419921875, |
| "step": 3700 |
| }, |
| { |
| "epoch": 1.945152, |
| "grad_norm": 17.827381134033203, |
| "learning_rate": 6.808437856328392e-05, |
| "loss": 42.019033203125, |
| "step": 3800 |
| }, |
| { |
| "epoch": 1.996352, |
| "grad_norm": 20.906902313232422, |
| "learning_rate": 6.694412770809578e-05, |
| "loss": 41.5656201171875, |
| "step": 3900 |
| }, |
| { |
| "epoch": 2.0, |
| "eval_loss": 5.107455253601074, |
| "eval_runtime": 54.6023, |
| "eval_samples_per_second": 366.285, |
| "eval_steps_per_second": 11.446, |
| "step": 3908 |
| }, |
| { |
| "epoch": 2.047104, |
| "grad_norm": 16.346454620361328, |
| "learning_rate": 6.580387685290765e-05, |
| "loss": 40.89203369140625, |
| "step": 4000 |
| }, |
| { |
| "epoch": 2.098304, |
| "grad_norm": 16.84693145751953, |
| "learning_rate": 6.46636259977195e-05, |
| "loss": 41.0606884765625, |
| "step": 4100 |
| }, |
| { |
| "epoch": 2.149504, |
| "grad_norm": 18.569360733032227, |
| "learning_rate": 6.352337514253136e-05, |
| "loss": 40.92314697265625, |
| "step": 4200 |
| }, |
| { |
| "epoch": 2.200704, |
| "grad_norm": 15.775079727172852, |
| "learning_rate": 6.238312428734322e-05, |
| "loss": 40.7518408203125, |
| "step": 4300 |
| }, |
| { |
| "epoch": 2.251904, |
| "grad_norm": 18.271591186523438, |
| "learning_rate": 6.124287343215507e-05, |
| "loss": 40.4889990234375, |
| "step": 4400 |
| }, |
| { |
| "epoch": 2.303104, |
| "grad_norm": 20.265701293945312, |
| "learning_rate": 6.010262257696694e-05, |
| "loss": 40.1488427734375, |
| "step": 4500 |
| }, |
| { |
| "epoch": 2.354304, |
| "grad_norm": 19.53594398498535, |
| "learning_rate": 5.8962371721778794e-05, |
| "loss": 40.10376708984375, |
| "step": 4600 |
| }, |
| { |
| "epoch": 2.405504, |
| "grad_norm": 19.707582473754883, |
| "learning_rate": 5.782212086659066e-05, |
| "loss": 39.953662109375, |
| "step": 4700 |
| }, |
| { |
| "epoch": 2.456704, |
| "grad_norm": 15.16901683807373, |
| "learning_rate": 5.6681870011402515e-05, |
| "loss": 39.77469970703125, |
| "step": 4800 |
| }, |
| { |
| "epoch": 2.5002240000000002, |
| "eval_loss": 4.868845462799072, |
| "eval_runtime": 54.8341, |
| "eval_samples_per_second": 364.737, |
| "eval_steps_per_second": 11.398, |
| "step": 4885 |
| }, |
| { |
| "epoch": 2.507904, |
| "grad_norm": 17.12852668762207, |
| "learning_rate": 5.554161915621437e-05, |
| "loss": 39.64916015625, |
| "step": 4900 |
| }, |
| { |
| "epoch": 2.559104, |
| "grad_norm": 19.869260787963867, |
| "learning_rate": 5.440136830102622e-05, |
| "loss": 39.51341552734375, |
| "step": 5000 |
| }, |
| { |
| "epoch": 2.610304, |
| "grad_norm": 17.342073440551758, |
| "learning_rate": 5.326111744583808e-05, |
| "loss": 39.350986328125, |
| "step": 5100 |
| }, |
| { |
| "epoch": 2.661504, |
| "grad_norm": 19.635601043701172, |
| "learning_rate": 5.212086659064994e-05, |
| "loss": 39.2056787109375, |
| "step": 5200 |
| }, |
| { |
| "epoch": 2.712704, |
| "grad_norm": 16.2427921295166, |
| "learning_rate": 5.09806157354618e-05, |
| "loss": 38.98402587890625, |
| "step": 5300 |
| }, |
| { |
| "epoch": 2.763904, |
| "grad_norm": 21.025632858276367, |
| "learning_rate": 4.984036488027366e-05, |
| "loss": 38.84795166015625, |
| "step": 5400 |
| }, |
| { |
| "epoch": 2.815104, |
| "grad_norm": 16.225173950195312, |
| "learning_rate": 4.870011402508552e-05, |
| "loss": 38.63864013671875, |
| "step": 5500 |
| }, |
| { |
| "epoch": 2.866304, |
| "grad_norm": 19.175825119018555, |
| "learning_rate": 4.755986316989738e-05, |
| "loss": 38.5738232421875, |
| "step": 5600 |
| }, |
| { |
| "epoch": 2.917504, |
| "grad_norm": 18.190704345703125, |
| "learning_rate": 4.6419612314709235e-05, |
| "loss": 38.47998046875, |
| "step": 5700 |
| }, |
| { |
| "epoch": 2.968704, |
| "grad_norm": 17.764493942260742, |
| "learning_rate": 4.52793614595211e-05, |
| "loss": 38.36706298828125, |
| "step": 5800 |
| }, |
| { |
| "epoch": 3.0, |
| "eval_loss": 4.690184116363525, |
| "eval_runtime": 54.4789, |
| "eval_samples_per_second": 367.115, |
| "eval_steps_per_second": 11.472, |
| "step": 5862 |
| }, |
| { |
| "epoch": 3.019456, |
| "grad_norm": 16.051603317260742, |
| "learning_rate": 4.4139110604332956e-05, |
| "loss": 37.86203369140625, |
| "step": 5900 |
| }, |
| { |
| "epoch": 3.070656, |
| "grad_norm": 16.511327743530273, |
| "learning_rate": 4.299885974914481e-05, |
| "loss": 38.09601318359375, |
| "step": 6000 |
| }, |
| { |
| "epoch": 3.121856, |
| "grad_norm": 16.089813232421875, |
| "learning_rate": 4.1858608893956676e-05, |
| "loss": 37.88030029296875, |
| "step": 6100 |
| }, |
| { |
| "epoch": 3.173056, |
| "grad_norm": 18.612686157226562, |
| "learning_rate": 4.0718358038768533e-05, |
| "loss": 37.66849853515625, |
| "step": 6200 |
| }, |
| { |
| "epoch": 3.224256, |
| "grad_norm": 17.79659652709961, |
| "learning_rate": 3.957810718358039e-05, |
| "loss": 37.43648681640625, |
| "step": 6300 |
| }, |
| { |
| "epoch": 3.275456, |
| "grad_norm": 16.967939376831055, |
| "learning_rate": 3.843785632839225e-05, |
| "loss": 37.43171142578125, |
| "step": 6400 |
| }, |
| { |
| "epoch": 3.326656, |
| "grad_norm": 16.4842529296875, |
| "learning_rate": 3.7297605473204104e-05, |
| "loss": 37.26192138671875, |
| "step": 6500 |
| }, |
| { |
| "epoch": 3.377856, |
| "grad_norm": 16.261512756347656, |
| "learning_rate": 3.615735461801597e-05, |
| "loss": 37.26702392578125, |
| "step": 6600 |
| }, |
| { |
| "epoch": 3.429056, |
| "grad_norm": 17.182903289794922, |
| "learning_rate": 3.5017103762827825e-05, |
| "loss": 37.16359619140625, |
| "step": 6700 |
| }, |
| { |
| "epoch": 3.480256, |
| "grad_norm": 17.765832901000977, |
| "learning_rate": 3.387685290763968e-05, |
| "loss": 37.05822998046875, |
| "step": 6800 |
| }, |
| { |
| "epoch": 3.5002240000000002, |
| "eval_loss": 4.552316188812256, |
| "eval_runtime": 54.6295, |
| "eval_samples_per_second": 366.103, |
| "eval_steps_per_second": 11.441, |
| "step": 6839 |
| }, |
| { |
| "epoch": 3.531456, |
| "grad_norm": 18.267107009887695, |
| "learning_rate": 3.2736602052451546e-05, |
| "loss": 36.907744140625, |
| "step": 6900 |
| }, |
| { |
| "epoch": 3.582656, |
| "grad_norm": 20.562347412109375, |
| "learning_rate": 3.15963511972634e-05, |
| "loss": 36.8392919921875, |
| "step": 7000 |
| }, |
| { |
| "epoch": 3.6338559999999998, |
| "grad_norm": 16.440811157226562, |
| "learning_rate": 3.0456100342075257e-05, |
| "loss": 36.639384765625, |
| "step": 7100 |
| }, |
| { |
| "epoch": 3.685056, |
| "grad_norm": 17.070350646972656, |
| "learning_rate": 2.9315849486887114e-05, |
| "loss": 36.63670166015625, |
| "step": 7200 |
| }, |
| { |
| "epoch": 3.736256, |
| "grad_norm": 17.15755844116211, |
| "learning_rate": 2.8175598631698974e-05, |
| "loss": 36.6428759765625, |
| "step": 7300 |
| }, |
| { |
| "epoch": 3.787456, |
| "grad_norm": 18.988977432250977, |
| "learning_rate": 2.7035347776510834e-05, |
| "loss": 36.578349609375, |
| "step": 7400 |
| }, |
| { |
| "epoch": 3.838656, |
| "grad_norm": 17.87462615966797, |
| "learning_rate": 2.589509692132269e-05, |
| "loss": 36.4393994140625, |
| "step": 7500 |
| }, |
| { |
| "epoch": 3.889856, |
| "grad_norm": 18.244951248168945, |
| "learning_rate": 2.4754846066134552e-05, |
| "loss": 36.40805908203125, |
| "step": 7600 |
| }, |
| { |
| "epoch": 3.941056, |
| "grad_norm": 16.166940689086914, |
| "learning_rate": 2.361459521094641e-05, |
| "loss": 36.23387451171875, |
| "step": 7700 |
| }, |
| { |
| "epoch": 3.9922560000000002, |
| "grad_norm": 19.10250473022461, |
| "learning_rate": 2.2474344355758266e-05, |
| "loss": 36.08807373046875, |
| "step": 7800 |
| }, |
| { |
| "epoch": 4.0, |
| "eval_loss": 4.433136940002441, |
| "eval_runtime": 54.6448, |
| "eval_samples_per_second": 366.0, |
| "eval_steps_per_second": 11.438, |
| "step": 7816 |
| }, |
| { |
| "epoch": 4.043008, |
| "grad_norm": 17.705839157104492, |
| "learning_rate": 2.1334093500570126e-05, |
| "loss": 35.8281103515625, |
| "step": 7900 |
| }, |
| { |
| "epoch": 4.094208, |
| "grad_norm": 16.15859031677246, |
| "learning_rate": 2.0193842645381987e-05, |
| "loss": 35.8823193359375, |
| "step": 8000 |
| }, |
| { |
| "epoch": 4.145408, |
| "grad_norm": 17.96906852722168, |
| "learning_rate": 1.9053591790193844e-05, |
| "loss": 35.968681640625, |
| "step": 8100 |
| }, |
| { |
| "epoch": 4.196608, |
| "grad_norm": 16.74631118774414, |
| "learning_rate": 1.79133409350057e-05, |
| "loss": 35.894248046875, |
| "step": 8200 |
| }, |
| { |
| "epoch": 4.247808, |
| "grad_norm": 16.067609786987305, |
| "learning_rate": 1.677309007981756e-05, |
| "loss": 35.78274169921875, |
| "step": 8300 |
| }, |
| { |
| "epoch": 4.299008, |
| "grad_norm": 17.308250427246094, |
| "learning_rate": 1.563283922462942e-05, |
| "loss": 35.8246630859375, |
| "step": 8400 |
| }, |
| { |
| "epoch": 4.350208, |
| "grad_norm": 17.30617332458496, |
| "learning_rate": 1.4492588369441278e-05, |
| "loss": 35.7324560546875, |
| "step": 8500 |
| }, |
| { |
| "epoch": 4.401408, |
| "grad_norm": 16.38850975036621, |
| "learning_rate": 1.3352337514253135e-05, |
| "loss": 35.88875244140625, |
| "step": 8600 |
| }, |
| { |
| "epoch": 4.452608, |
| "grad_norm": 20.229633331298828, |
| "learning_rate": 1.2212086659064994e-05, |
| "loss": 35.7011962890625, |
| "step": 8700 |
| }, |
| { |
| "epoch": 4.500224, |
| "eval_loss": 4.369416236877441, |
| "eval_runtime": 54.3995, |
| "eval_samples_per_second": 367.651, |
| "eval_steps_per_second": 11.489, |
| "step": 8793 |
| }, |
| { |
| "epoch": 4.503808, |
| "grad_norm": 18.37810707092285, |
| "learning_rate": 1.1071835803876854e-05, |
| "loss": 35.69582763671875, |
| "step": 8800 |
| }, |
| { |
| "epoch": 4.555008, |
| "grad_norm": 16.187055587768555, |
| "learning_rate": 9.931584948688711e-06, |
| "loss": 35.73132080078125, |
| "step": 8900 |
| }, |
| { |
| "epoch": 4.606208, |
| "grad_norm": 16.89447021484375, |
| "learning_rate": 8.79133409350057e-06, |
| "loss": 35.47484375, |
| "step": 9000 |
| }, |
| { |
| "epoch": 4.657408, |
| "grad_norm": 17.172351837158203, |
| "learning_rate": 7.651083238312429e-06, |
| "loss": 35.4910205078125, |
| "step": 9100 |
| }, |
| { |
| "epoch": 4.708608, |
| "grad_norm": 18.75235939025879, |
| "learning_rate": 6.5108323831242875e-06, |
| "loss": 35.401552734375, |
| "step": 9200 |
| }, |
| { |
| "epoch": 4.759808, |
| "grad_norm": 17.79449462890625, |
| "learning_rate": 5.370581527936146e-06, |
| "loss": 35.37335693359375, |
| "step": 9300 |
| }, |
| { |
| "epoch": 4.811008, |
| "grad_norm": 16.077747344970703, |
| "learning_rate": 4.230330672748005e-06, |
| "loss": 35.395380859375, |
| "step": 9400 |
| }, |
| { |
| "epoch": 4.862208, |
| "grad_norm": 17.572834014892578, |
| "learning_rate": 3.0900798175598636e-06, |
| "loss": 35.3202685546875, |
| "step": 9500 |
| }, |
| { |
| "epoch": 4.913408, |
| "grad_norm": 17.8610897064209, |
| "learning_rate": 1.949828962371722e-06, |
| "loss": 35.46013916015625, |
| "step": 9600 |
| }, |
| { |
| "epoch": 4.964608, |
| "grad_norm": 18.012996673583984, |
| "learning_rate": 8.095781071835805e-07, |
| "loss": 35.363046875, |
| "step": 9700 |
| }, |
| { |
| "epoch": 5.0, |
| "eval_loss": 4.354217052459717, |
| "eval_runtime": 54.5977, |
| "eval_samples_per_second": 366.316, |
| "eval_steps_per_second": 11.447, |
| "step": 9770 |
| } |
| ], |
| "logging_steps": 100, |
| "max_steps": 9770, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 5, |
| "save_steps": 977, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 6.5802339072e+17, |
| "train_batch_size": 32, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|