| { | |
| "best_global_step": 9000, | |
| "best_metric": 0.5520676374435425, | |
| "best_model_checkpoint": "/home/work/.workspace/exaone_RL_test/outputs/checkpoint-9000", | |
| "epoch": 2.0, | |
| "eval_steps": 1000, | |
| "global_step": 9766, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0020479213598197828, | |
| "grad_norm": 154.0, | |
| "learning_rate": 3.071672354948806e-07, | |
| "loss": 5.2426, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.0040958427196395655, | |
| "grad_norm": 134.0, | |
| "learning_rate": 6.484641638225256e-07, | |
| "loss": 5.0177, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.006143764079459349, | |
| "grad_norm": 101.0, | |
| "learning_rate": 9.897610921501708e-07, | |
| "loss": 4.6525, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.008191685439279131, | |
| "grad_norm": 64.0, | |
| "learning_rate": 1.331058020477816e-06, | |
| "loss": 3.8346, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.010239606799098914, | |
| "grad_norm": 44.0, | |
| "learning_rate": 1.6723549488054607e-06, | |
| "loss": 2.898, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.012287528158918698, | |
| "grad_norm": 35.75, | |
| "learning_rate": 2.013651877133106e-06, | |
| "loss": 1.9498, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.014335449518738481, | |
| "grad_norm": 29.375, | |
| "learning_rate": 2.354948805460751e-06, | |
| "loss": 1.3878, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.016383370878558262, | |
| "grad_norm": 10.6875, | |
| "learning_rate": 2.696245733788396e-06, | |
| "loss": 1.1501, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.018431292238378045, | |
| "grad_norm": 9.1875, | |
| "learning_rate": 3.0375426621160415e-06, | |
| "loss": 1.0169, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.020479213598197828, | |
| "grad_norm": 9.6875, | |
| "learning_rate": 3.378839590443686e-06, | |
| "loss": 0.8679, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.02252713495801761, | |
| "grad_norm": 7.59375, | |
| "learning_rate": 3.7201365187713314e-06, | |
| "loss": 0.8122, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.024575056317837397, | |
| "grad_norm": 8.25, | |
| "learning_rate": 4.061433447098976e-06, | |
| "loss": 0.7636, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.02662297767765718, | |
| "grad_norm": 7.9375, | |
| "learning_rate": 4.402730375426622e-06, | |
| "loss": 0.7755, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.028670899037476962, | |
| "grad_norm": 7.90625, | |
| "learning_rate": 4.744027303754267e-06, | |
| "loss": 0.749, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.030718820397296745, | |
| "grad_norm": 8.9375, | |
| "learning_rate": 5.0853242320819115e-06, | |
| "loss": 0.7375, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.032766741757116524, | |
| "grad_norm": 7.25, | |
| "learning_rate": 5.426621160409556e-06, | |
| "loss": 0.7443, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.03481466311693631, | |
| "grad_norm": 7.53125, | |
| "learning_rate": 5.767918088737202e-06, | |
| "loss": 0.7866, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.03686258447675609, | |
| "grad_norm": 6.5, | |
| "learning_rate": 6.109215017064847e-06, | |
| "loss": 0.719, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.038910505836575876, | |
| "grad_norm": 6.125, | |
| "learning_rate": 6.450511945392492e-06, | |
| "loss": 0.6936, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.040958427196395655, | |
| "grad_norm": 5.75, | |
| "learning_rate": 6.7918088737201375e-06, | |
| "loss": 0.7347, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.04300634855621544, | |
| "grad_norm": 5.96875, | |
| "learning_rate": 7.133105802047782e-06, | |
| "loss": 0.6707, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.04505426991603522, | |
| "grad_norm": 7.9375, | |
| "learning_rate": 7.474402730375427e-06, | |
| "loss": 0.6804, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.04710219127585501, | |
| "grad_norm": 5.71875, | |
| "learning_rate": 7.815699658703072e-06, | |
| "loss": 0.6831, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.04915011263567479, | |
| "grad_norm": 5.40625, | |
| "learning_rate": 8.156996587030718e-06, | |
| "loss": 0.634, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.05119803399549457, | |
| "grad_norm": 5.5625, | |
| "learning_rate": 8.498293515358363e-06, | |
| "loss": 0.6739, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.05324595535531436, | |
| "grad_norm": 4.4375, | |
| "learning_rate": 8.839590443686009e-06, | |
| "loss": 0.65, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.05529387671513414, | |
| "grad_norm": 5.5625, | |
| "learning_rate": 9.180887372013653e-06, | |
| "loss": 0.6803, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.057341798074953924, | |
| "grad_norm": 5.0625, | |
| "learning_rate": 9.522184300341298e-06, | |
| "loss": 0.7035, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.059389719434773704, | |
| "grad_norm": 5.875, | |
| "learning_rate": 9.863481228668942e-06, | |
| "loss": 0.7119, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.06143764079459349, | |
| "grad_norm": 6.4375, | |
| "learning_rate": 9.999990101553574e-06, | |
| "loss": 0.7058, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.06348556215441327, | |
| "grad_norm": 5.34375, | |
| "learning_rate": 9.999929611189557e-06, | |
| "loss": 0.6344, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.06553348351423305, | |
| "grad_norm": 3.6875, | |
| "learning_rate": 9.999814130262912e-06, | |
| "loss": 0.6832, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.06758140487405284, | |
| "grad_norm": 4.375, | |
| "learning_rate": 9.999643660043727e-06, | |
| "loss": 0.6186, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.06962932623387262, | |
| "grad_norm": 4.90625, | |
| "learning_rate": 9.999418202406882e-06, | |
| "loss": 0.7257, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.0716772475936924, | |
| "grad_norm": 5.0, | |
| "learning_rate": 9.999137759832017e-06, | |
| "loss": 0.6864, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.07372516895351218, | |
| "grad_norm": 4.53125, | |
| "learning_rate": 9.998802335403514e-06, | |
| "loss": 0.6495, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.07577309031333197, | |
| "grad_norm": 4.875, | |
| "learning_rate": 9.998411932810465e-06, | |
| "loss": 0.6305, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.07782101167315175, | |
| "grad_norm": 4.25, | |
| "learning_rate": 9.997966556346617e-06, | |
| "loss": 0.6956, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.07986893303297153, | |
| "grad_norm": 4.34375, | |
| "learning_rate": 9.997466210910344e-06, | |
| "loss": 0.5943, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.08191685439279131, | |
| "grad_norm": 6.71875, | |
| "learning_rate": 9.996910902004576e-06, | |
| "loss": 0.6226, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.0839647757526111, | |
| "grad_norm": 4.8125, | |
| "learning_rate": 9.99630063573675e-06, | |
| "loss": 0.6206, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.08601269711243088, | |
| "grad_norm": 4.625, | |
| "learning_rate": 9.995635418818734e-06, | |
| "loss": 0.6269, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.08806061847225066, | |
| "grad_norm": 4.5625, | |
| "learning_rate": 9.994915258566766e-06, | |
| "loss": 0.5965, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.09010853983207044, | |
| "grad_norm": 4.5625, | |
| "learning_rate": 9.994140162901362e-06, | |
| "loss": 0.6204, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.09215646119189023, | |
| "grad_norm": 4.40625, | |
| "learning_rate": 9.993310140347227e-06, | |
| "loss": 0.687, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.09420438255171001, | |
| "grad_norm": 4.625, | |
| "learning_rate": 9.992425200033173e-06, | |
| "loss": 0.5611, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.0962523039115298, | |
| "grad_norm": 4.21875, | |
| "learning_rate": 9.99148535169201e-06, | |
| "loss": 0.6094, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.09830022527134959, | |
| "grad_norm": 4.90625, | |
| "learning_rate": 9.990490605660441e-06, | |
| "loss": 0.6283, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.10034814663116937, | |
| "grad_norm": 4.96875, | |
| "learning_rate": 9.98944097287895e-06, | |
| "loss": 0.6276, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.10239606799098915, | |
| "grad_norm": 4.65625, | |
| "learning_rate": 9.988336464891677e-06, | |
| "loss": 0.6535, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.10444398935080892, | |
| "grad_norm": 5.03125, | |
| "learning_rate": 9.987177093846299e-06, | |
| "loss": 0.6377, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.10649191071062872, | |
| "grad_norm": 3.71875, | |
| "learning_rate": 9.985962872493885e-06, | |
| "loss": 0.6542, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.1085398320704485, | |
| "grad_norm": 4.40625, | |
| "learning_rate": 9.984693814188769e-06, | |
| "loss": 0.6088, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.11058775343026828, | |
| "grad_norm": 4.34375, | |
| "learning_rate": 9.983369932888393e-06, | |
| "loss": 0.6549, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.11263567479008806, | |
| "grad_norm": 3.75, | |
| "learning_rate": 9.981991243153156e-06, | |
| "loss": 0.5653, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.11468359614990785, | |
| "grad_norm": 3.484375, | |
| "learning_rate": 9.980557760146259e-06, | |
| "loss": 0.6158, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.11673151750972763, | |
| "grad_norm": 3.875, | |
| "learning_rate": 9.979069499633528e-06, | |
| "loss": 0.6301, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.11877943886954741, | |
| "grad_norm": 3.375, | |
| "learning_rate": 9.977526477983252e-06, | |
| "loss": 0.5643, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.12082736022936719, | |
| "grad_norm": 4.21875, | |
| "learning_rate": 9.975928712165997e-06, | |
| "loss": 0.6924, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.12287528158918698, | |
| "grad_norm": 4.6875, | |
| "learning_rate": 9.974276219754416e-06, | |
| "loss": 0.6238, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.12492320294900676, | |
| "grad_norm": 4.34375, | |
| "learning_rate": 9.972569018923064e-06, | |
| "loss": 0.6142, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.12697112430882654, | |
| "grad_norm": 4.03125, | |
| "learning_rate": 9.97080712844819e-06, | |
| "loss": 0.5602, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.12901904566864633, | |
| "grad_norm": 4.96875, | |
| "learning_rate": 9.96899056770754e-06, | |
| "loss": 0.5933, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.1310669670284661, | |
| "grad_norm": 4.34375, | |
| "learning_rate": 9.967119356680131e-06, | |
| "loss": 0.591, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.1331148883882859, | |
| "grad_norm": 4.46875, | |
| "learning_rate": 9.965193515946045e-06, | |
| "loss": 0.594, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.13516280974810568, | |
| "grad_norm": 4.40625, | |
| "learning_rate": 9.963213066686193e-06, | |
| "loss": 0.6271, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.13721073110792545, | |
| "grad_norm": 4.4375, | |
| "learning_rate": 9.961178030682083e-06, | |
| "loss": 0.6486, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.13925865246774524, | |
| "grad_norm": 4.25, | |
| "learning_rate": 9.959088430315587e-06, | |
| "loss": 0.641, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.14130657382756504, | |
| "grad_norm": 4.375, | |
| "learning_rate": 9.956944288568689e-06, | |
| "loss": 0.6058, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.1433544951873848, | |
| "grad_norm": 4.3125, | |
| "learning_rate": 9.95474562902323e-06, | |
| "loss": 0.6139, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.1454024165472046, | |
| "grad_norm": 4.53125, | |
| "learning_rate": 9.952492475860662e-06, | |
| "loss": 0.5889, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.14745033790702436, | |
| "grad_norm": 3.796875, | |
| "learning_rate": 9.95018485386176e-06, | |
| "loss": 0.5904, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.14949825926684415, | |
| "grad_norm": 4.21875, | |
| "learning_rate": 9.947822788406367e-06, | |
| "loss": 0.6628, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.15154618062666395, | |
| "grad_norm": 3.59375, | |
| "learning_rate": 9.945406305473116e-06, | |
| "loss": 0.555, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.1535941019864837, | |
| "grad_norm": 4.3125, | |
| "learning_rate": 9.942935431639128e-06, | |
| "loss": 0.6128, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.1556420233463035, | |
| "grad_norm": 4.1875, | |
| "learning_rate": 9.940410194079736e-06, | |
| "loss": 0.5984, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.1576899447061233, | |
| "grad_norm": 3.53125, | |
| "learning_rate": 9.937830620568176e-06, | |
| "loss": 0.6214, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.15973786606594306, | |
| "grad_norm": 3.671875, | |
| "learning_rate": 9.935196739475287e-06, | |
| "loss": 0.5355, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.16178578742576286, | |
| "grad_norm": 4.5625, | |
| "learning_rate": 9.932508579769203e-06, | |
| "loss": 0.6303, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.16383370878558262, | |
| "grad_norm": 3.9375, | |
| "learning_rate": 9.929766171015015e-06, | |
| "loss": 0.5743, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.16588163014540241, | |
| "grad_norm": 3.625, | |
| "learning_rate": 9.926969543374473e-06, | |
| "loss": 0.5829, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.1679295515052222, | |
| "grad_norm": 4.75, | |
| "learning_rate": 9.924118727605633e-06, | |
| "loss": 0.6498, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.16997747286504197, | |
| "grad_norm": 4.21875, | |
| "learning_rate": 9.921213755062531e-06, | |
| "loss": 0.6007, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.17202539422486177, | |
| "grad_norm": 4.46875, | |
| "learning_rate": 9.918254657694831e-06, | |
| "loss": 0.639, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.17407331558468156, | |
| "grad_norm": 3.28125, | |
| "learning_rate": 9.915241468047476e-06, | |
| "loss": 0.5943, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.17612123694450132, | |
| "grad_norm": 3.65625, | |
| "learning_rate": 9.912174219260332e-06, | |
| "loss": 0.5616, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.17816915830432112, | |
| "grad_norm": 4.84375, | |
| "learning_rate": 9.90905294506782e-06, | |
| "loss": 0.5864, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.18021707966414088, | |
| "grad_norm": 3.796875, | |
| "learning_rate": 9.905877679798552e-06, | |
| "loss": 0.6337, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.18226500102396068, | |
| "grad_norm": 4.25, | |
| "learning_rate": 9.90264845837494e-06, | |
| "loss": 0.6289, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.18431292238378047, | |
| "grad_norm": 3.328125, | |
| "learning_rate": 9.899365316312826e-06, | |
| "loss": 0.5829, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.18636084374360024, | |
| "grad_norm": 4.90625, | |
| "learning_rate": 9.896028289721085e-06, | |
| "loss": 0.6398, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.18840876510342003, | |
| "grad_norm": 3.765625, | |
| "learning_rate": 9.892637415301227e-06, | |
| "loss": 0.5661, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.19045668646323982, | |
| "grad_norm": 3.765625, | |
| "learning_rate": 9.889192730346994e-06, | |
| "loss": 0.6144, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.1925046078230596, | |
| "grad_norm": 4.375, | |
| "learning_rate": 9.885694272743955e-06, | |
| "loss": 0.5456, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.19455252918287938, | |
| "grad_norm": 5.53125, | |
| "learning_rate": 9.882142080969082e-06, | |
| "loss": 0.6665, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.19660045054269917, | |
| "grad_norm": 3.515625, | |
| "learning_rate": 9.878536194090326e-06, | |
| "loss": 0.6107, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.19864837190251894, | |
| "grad_norm": 3.359375, | |
| "learning_rate": 9.874876651766198e-06, | |
| "loss": 0.5805, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.20069629326233873, | |
| "grad_norm": 3.984375, | |
| "learning_rate": 9.871163494245324e-06, | |
| "loss": 0.6071, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.2027442146221585, | |
| "grad_norm": 4.25, | |
| "learning_rate": 9.867396762366e-06, | |
| "loss": 0.6215, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.2047921359819783, | |
| "grad_norm": 5.1875, | |
| "learning_rate": 9.863576497555752e-06, | |
| "loss": 0.6455, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.2047921359819783, | |
| "eval_loss": 0.5988336801528931, | |
| "eval_runtime": 49.5583, | |
| "eval_samples_per_second": 82.973, | |
| "eval_steps_per_second": 41.487, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.20684005734179808, | |
| "grad_norm": 3.609375, | |
| "learning_rate": 9.859702741830873e-06, | |
| "loss": 0.546, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.20888797870161785, | |
| "grad_norm": 4.125, | |
| "learning_rate": 9.855775537795965e-06, | |
| "loss": 0.6586, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.21093590006143764, | |
| "grad_norm": 4.625, | |
| "learning_rate": 9.851794928643465e-06, | |
| "loss": 0.6279, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.21298382142125744, | |
| "grad_norm": 3.296875, | |
| "learning_rate": 9.847760958153183e-06, | |
| "loss": 0.5474, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.2150317427810772, | |
| "grad_norm": 3.640625, | |
| "learning_rate": 9.8436736706918e-06, | |
| "loss": 0.58, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.217079664140897, | |
| "grad_norm": 3.515625, | |
| "learning_rate": 9.839533111212395e-06, | |
| "loss": 0.6043, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.21912758550071676, | |
| "grad_norm": 4.3125, | |
| "learning_rate": 9.835339325253954e-06, | |
| "loss": 0.6341, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.22117550686053655, | |
| "grad_norm": 4.0625, | |
| "learning_rate": 9.831092358940853e-06, | |
| "loss": 0.5638, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.22322342822035635, | |
| "grad_norm": 3.3125, | |
| "learning_rate": 9.82679225898236e-06, | |
| "loss": 0.6114, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.2252713495801761, | |
| "grad_norm": 4.90625, | |
| "learning_rate": 9.82243907267213e-06, | |
| "loss": 0.6526, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.2273192709399959, | |
| "grad_norm": 4.625, | |
| "learning_rate": 9.818032847887664e-06, | |
| "loss": 0.5896, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.2293671922998157, | |
| "grad_norm": 3.453125, | |
| "learning_rate": 9.813573633089804e-06, | |
| "loss": 0.5732, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.23141511365963546, | |
| "grad_norm": 3.71875, | |
| "learning_rate": 9.809061477322186e-06, | |
| "loss": 0.6223, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.23346303501945526, | |
| "grad_norm": 4.21875, | |
| "learning_rate": 9.804496430210704e-06, | |
| "loss": 0.6012, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.23551095637927505, | |
| "grad_norm": 4.0, | |
| "learning_rate": 9.799878541962968e-06, | |
| "loss": 0.6064, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.23755887773909481, | |
| "grad_norm": 3.828125, | |
| "learning_rate": 9.795207863367745e-06, | |
| "loss": 0.553, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.2396067990989146, | |
| "grad_norm": 4.53125, | |
| "learning_rate": 9.790484445794413e-06, | |
| "loss": 0.6526, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.24165472045873437, | |
| "grad_norm": 4.25, | |
| "learning_rate": 9.785708341192375e-06, | |
| "loss": 0.6117, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.24370264181855417, | |
| "grad_norm": 3.828125, | |
| "learning_rate": 9.78087960209051e-06, | |
| "loss": 0.5773, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.24575056317837396, | |
| "grad_norm": 4.71875, | |
| "learning_rate": 9.77599828159658e-06, | |
| "loss": 0.6004, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.24779848453819373, | |
| "grad_norm": 4.40625, | |
| "learning_rate": 9.771064433396652e-06, | |
| "loss": 0.5644, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 0.24984640589801352, | |
| "grad_norm": 3.75, | |
| "learning_rate": 9.76607811175451e-06, | |
| "loss": 0.5899, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.2518943272578333, | |
| "grad_norm": 4.78125, | |
| "learning_rate": 9.761039371511051e-06, | |
| "loss": 0.5757, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.2539422486176531, | |
| "grad_norm": 3.6875, | |
| "learning_rate": 9.755948268083689e-06, | |
| "loss": 0.5837, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.25599016997747287, | |
| "grad_norm": 4.21875, | |
| "learning_rate": 9.750804857465736e-06, | |
| "loss": 0.5767, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.25803809133729266, | |
| "grad_norm": 3.0625, | |
| "learning_rate": 9.745609196225804e-06, | |
| "loss": 0.6256, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.26008601269711246, | |
| "grad_norm": 3.9375, | |
| "learning_rate": 9.740361341507158e-06, | |
| "loss": 0.5644, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 0.2621339340569322, | |
| "grad_norm": 3.859375, | |
| "learning_rate": 9.735061351027114e-06, | |
| "loss": 0.6039, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.264181855416752, | |
| "grad_norm": 3.75, | |
| "learning_rate": 9.729709283076375e-06, | |
| "loss": 0.5902, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.2662297767765718, | |
| "grad_norm": 3.921875, | |
| "learning_rate": 9.724305196518425e-06, | |
| "loss": 0.5738, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.2682776981363916, | |
| "grad_norm": 3.75, | |
| "learning_rate": 9.718849150788848e-06, | |
| "loss": 0.5882, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 0.27032561949621137, | |
| "grad_norm": 3.65625, | |
| "learning_rate": 9.713341205894691e-06, | |
| "loss": 0.5549, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.2723735408560311, | |
| "grad_norm": 3.578125, | |
| "learning_rate": 9.707781422413811e-06, | |
| "loss": 0.5866, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 0.2744214622158509, | |
| "grad_norm": 3.671875, | |
| "learning_rate": 9.702169861494189e-06, | |
| "loss": 0.6095, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.2764693835756707, | |
| "grad_norm": 3.890625, | |
| "learning_rate": 9.696506584853271e-06, | |
| "loss": 0.5493, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.2785173049354905, | |
| "grad_norm": 4.125, | |
| "learning_rate": 9.690791654777293e-06, | |
| "loss": 0.6172, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.2805652262953103, | |
| "grad_norm": 3.046875, | |
| "learning_rate": 9.685025134120577e-06, | |
| "loss": 0.5721, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 0.28261314765513007, | |
| "grad_norm": 4.84375, | |
| "learning_rate": 9.679207086304864e-06, | |
| "loss": 0.5712, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.2846610690149498, | |
| "grad_norm": 3.890625, | |
| "learning_rate": 9.673337575318597e-06, | |
| "loss": 0.6429, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 0.2867089903747696, | |
| "grad_norm": 4.3125, | |
| "learning_rate": 9.66741666571623e-06, | |
| "loss": 0.6728, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.2887569117345894, | |
| "grad_norm": 4.375, | |
| "learning_rate": 9.661444422617505e-06, | |
| "loss": 0.6315, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 0.2908048330944092, | |
| "grad_norm": 4.3125, | |
| "learning_rate": 9.655420911706752e-06, | |
| "loss": 0.5544, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.292852754454229, | |
| "grad_norm": 3.796875, | |
| "learning_rate": 9.649346199232154e-06, | |
| "loss": 0.5458, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 0.2949006758140487, | |
| "grad_norm": 4.28125, | |
| "learning_rate": 9.643220352005023e-06, | |
| "loss": 0.5573, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.2969485971738685, | |
| "grad_norm": 3.75, | |
| "learning_rate": 9.637043437399063e-06, | |
| "loss": 0.6293, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.2989965185336883, | |
| "grad_norm": 3.296875, | |
| "learning_rate": 9.630815523349636e-06, | |
| "loss": 0.5829, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.3010444398935081, | |
| "grad_norm": 3.46875, | |
| "learning_rate": 9.624536678352998e-06, | |
| "loss": 0.5512, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 0.3030923612533279, | |
| "grad_norm": 3.65625, | |
| "learning_rate": 9.618206971465574e-06, | |
| "loss": 0.5756, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.30514028261314763, | |
| "grad_norm": 3.5625, | |
| "learning_rate": 9.611826472303165e-06, | |
| "loss": 0.5815, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 0.3071882039729674, | |
| "grad_norm": 3.515625, | |
| "learning_rate": 9.605395251040212e-06, | |
| "loss": 0.575, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.3092361253327872, | |
| "grad_norm": 3.484375, | |
| "learning_rate": 9.598913378409006e-06, | |
| "loss": 0.661, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 0.311284046692607, | |
| "grad_norm": 3.21875, | |
| "learning_rate": 9.592380925698911e-06, | |
| "loss": 0.5835, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.3133319680524268, | |
| "grad_norm": 4.53125, | |
| "learning_rate": 9.585797964755593e-06, | |
| "loss": 0.5867, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 0.3153798894122466, | |
| "grad_norm": 3.875, | |
| "learning_rate": 9.579164567980212e-06, | |
| "loss": 0.5765, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.31742781077206633, | |
| "grad_norm": 3.328125, | |
| "learning_rate": 9.572480808328643e-06, | |
| "loss": 0.5486, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.3194757321318861, | |
| "grad_norm": 3.65625, | |
| "learning_rate": 9.565746759310659e-06, | |
| "loss": 0.6179, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.3215236534917059, | |
| "grad_norm": 4.28125, | |
| "learning_rate": 9.558962494989133e-06, | |
| "loss": 0.6122, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 0.3235715748515257, | |
| "grad_norm": 4.0625, | |
| "learning_rate": 9.552128089979215e-06, | |
| "loss": 0.5539, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.3256194962113455, | |
| "grad_norm": 3.84375, | |
| "learning_rate": 9.545243619447524e-06, | |
| "loss": 0.5897, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 0.32766741757116524, | |
| "grad_norm": 4.15625, | |
| "learning_rate": 9.538309159111303e-06, | |
| "loss": 0.5231, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.32971533893098504, | |
| "grad_norm": 4.15625, | |
| "learning_rate": 9.531324785237603e-06, | |
| "loss": 0.6409, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 0.33176326029080483, | |
| "grad_norm": 3.109375, | |
| "learning_rate": 9.524290574642432e-06, | |
| "loss": 0.5647, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.3338111816506246, | |
| "grad_norm": 3.3125, | |
| "learning_rate": 9.517206604689924e-06, | |
| "loss": 0.6115, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 0.3358591030104444, | |
| "grad_norm": 4.21875, | |
| "learning_rate": 9.51007295329147e-06, | |
| "loss": 0.5784, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.3379070243702642, | |
| "grad_norm": 3.125, | |
| "learning_rate": 9.502889698904877e-06, | |
| "loss": 0.5442, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.33995494573008395, | |
| "grad_norm": 3.8125, | |
| "learning_rate": 9.495656920533499e-06, | |
| "loss": 0.5771, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.34200286708990374, | |
| "grad_norm": 3.78125, | |
| "learning_rate": 9.488374697725361e-06, | |
| "loss": 0.5675, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 0.34405078844972353, | |
| "grad_norm": 4.1875, | |
| "learning_rate": 9.481043110572301e-06, | |
| "loss": 0.6298, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.3460987098095433, | |
| "grad_norm": 5.0625, | |
| "learning_rate": 9.473662239709074e-06, | |
| "loss": 0.6504, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 0.3481466311693631, | |
| "grad_norm": 3.828125, | |
| "learning_rate": 9.46623216631247e-06, | |
| "loss": 0.606, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.35019455252918286, | |
| "grad_norm": 3.625, | |
| "learning_rate": 9.458752972100425e-06, | |
| "loss": 0.6153, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 0.35224247388900265, | |
| "grad_norm": 3.484375, | |
| "learning_rate": 9.451224739331113e-06, | |
| "loss": 0.5412, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 0.35429039524882244, | |
| "grad_norm": 3.5, | |
| "learning_rate": 9.443647550802052e-06, | |
| "loss": 0.5356, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 0.35633831660864224, | |
| "grad_norm": 5.1875, | |
| "learning_rate": 9.436021489849189e-06, | |
| "loss": 0.6106, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 0.35838623796846203, | |
| "grad_norm": 3.921875, | |
| "learning_rate": 9.42834664034598e-06, | |
| "loss": 0.5864, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.36043415932828177, | |
| "grad_norm": 4.03125, | |
| "learning_rate": 9.420623086702469e-06, | |
| "loss": 0.5616, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 0.36248208068810156, | |
| "grad_norm": 3.4375, | |
| "learning_rate": 9.41285091386437e-06, | |
| "loss": 0.5831, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 0.36453000204792135, | |
| "grad_norm": 3.640625, | |
| "learning_rate": 9.405030207312113e-06, | |
| "loss": 0.5746, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 0.36657792340774115, | |
| "grad_norm": 3.5, | |
| "learning_rate": 9.397161053059923e-06, | |
| "loss": 0.5429, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 0.36862584476756094, | |
| "grad_norm": 2.953125, | |
| "learning_rate": 9.389243537654857e-06, | |
| "loss": 0.5584, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.37067376612738073, | |
| "grad_norm": 3.90625, | |
| "learning_rate": 9.381277748175872e-06, | |
| "loss": 0.6123, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 0.37272168748720047, | |
| "grad_norm": 4.25, | |
| "learning_rate": 9.373263772232848e-06, | |
| "loss": 0.597, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 0.37476960884702026, | |
| "grad_norm": 4.0625, | |
| "learning_rate": 9.365201697965634e-06, | |
| "loss": 0.5578, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 0.37681753020684006, | |
| "grad_norm": 3.890625, | |
| "learning_rate": 9.357091614043078e-06, | |
| "loss": 0.6015, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 0.37886545156665985, | |
| "grad_norm": 3.515625, | |
| "learning_rate": 9.348933609662055e-06, | |
| "loss": 0.5694, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 0.38091337292647964, | |
| "grad_norm": 3.453125, | |
| "learning_rate": 9.340727774546476e-06, | |
| "loss": 0.5754, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 0.3829612942862994, | |
| "grad_norm": 3.609375, | |
| "learning_rate": 9.332474198946308e-06, | |
| "loss": 0.5598, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 0.3850092156461192, | |
| "grad_norm": 3.3125, | |
| "learning_rate": 9.324172973636583e-06, | |
| "loss": 0.5535, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 0.38705713700593897, | |
| "grad_norm": 3.890625, | |
| "learning_rate": 9.315824189916398e-06, | |
| "loss": 0.5996, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 0.38910505836575876, | |
| "grad_norm": 4.25, | |
| "learning_rate": 9.307427939607906e-06, | |
| "loss": 0.5701, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.39115297972557855, | |
| "grad_norm": 3.53125, | |
| "learning_rate": 9.298984315055316e-06, | |
| "loss": 0.5873, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 0.39320090108539835, | |
| "grad_norm": 3.515625, | |
| "learning_rate": 9.290493409123864e-06, | |
| "loss": 0.5693, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 0.3952488224452181, | |
| "grad_norm": 3.84375, | |
| "learning_rate": 9.281955315198805e-06, | |
| "loss": 0.5408, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 0.3972967438050379, | |
| "grad_norm": 3.203125, | |
| "learning_rate": 9.273370127184382e-06, | |
| "loss": 0.582, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 0.39934466516485767, | |
| "grad_norm": 3.28125, | |
| "learning_rate": 9.264737939502784e-06, | |
| "loss": 0.5886, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 0.40139258652467746, | |
| "grad_norm": 4.34375, | |
| "learning_rate": 9.256058847093122e-06, | |
| "loss": 0.5457, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 0.40344050788449726, | |
| "grad_norm": 4.0, | |
| "learning_rate": 9.24733294541037e-06, | |
| "loss": 0.5653, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 0.405488429244317, | |
| "grad_norm": 4.03125, | |
| "learning_rate": 9.238560330424333e-06, | |
| "loss": 0.6059, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 0.4075363506041368, | |
| "grad_norm": 4.1875, | |
| "learning_rate": 9.229741098618573e-06, | |
| "loss": 0.5621, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 0.4095842719639566, | |
| "grad_norm": 3.578125, | |
| "learning_rate": 9.220875346989363e-06, | |
| "loss": 0.5432, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.4095842719639566, | |
| "eval_loss": 0.5784266591072083, | |
| "eval_runtime": 49.5386, | |
| "eval_samples_per_second": 83.006, | |
| "eval_steps_per_second": 41.503, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.4116321933237764, | |
| "grad_norm": 3.09375, | |
| "learning_rate": 9.211963173044604e-06, | |
| "loss": 0.5743, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 0.41368011468359617, | |
| "grad_norm": 4.1875, | |
| "learning_rate": 9.203004674802774e-06, | |
| "loss": 0.613, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 0.41572803604341596, | |
| "grad_norm": 3.4375, | |
| "learning_rate": 9.193999950791825e-06, | |
| "loss": 0.5318, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 0.4177759574032357, | |
| "grad_norm": 3.1875, | |
| "learning_rate": 9.184949100048123e-06, | |
| "loss": 0.5718, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 0.4198238787630555, | |
| "grad_norm": 3.578125, | |
| "learning_rate": 9.175852222115346e-06, | |
| "loss": 0.5876, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 0.4218718001228753, | |
| "grad_norm": 4.15625, | |
| "learning_rate": 9.166709417043386e-06, | |
| "loss": 0.5436, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 0.4239197214826951, | |
| "grad_norm": 4.21875, | |
| "learning_rate": 9.15752078538726e-06, | |
| "loss": 0.5473, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 0.42596764284251487, | |
| "grad_norm": 3.5, | |
| "learning_rate": 9.148286428205994e-06, | |
| "loss": 0.6101, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 0.4280155642023346, | |
| "grad_norm": 4.625, | |
| "learning_rate": 9.139006447061518e-06, | |
| "loss": 0.5197, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 0.4300634855621544, | |
| "grad_norm": 4.25, | |
| "learning_rate": 9.129680944017544e-06, | |
| "loss": 0.5946, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.4321114069219742, | |
| "grad_norm": 4.5, | |
| "learning_rate": 9.120310021638451e-06, | |
| "loss": 0.5866, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 0.434159328281794, | |
| "grad_norm": 3.65625, | |
| "learning_rate": 9.110893782988148e-06, | |
| "loss": 0.5906, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 0.4362072496416138, | |
| "grad_norm": 3.28125, | |
| "learning_rate": 9.101432331628943e-06, | |
| "loss": 0.5084, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 0.4382551710014335, | |
| "grad_norm": 3.875, | |
| "learning_rate": 9.09192577162041e-06, | |
| "loss": 0.5464, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 0.4403030923612533, | |
| "grad_norm": 4.0, | |
| "learning_rate": 9.082374207518234e-06, | |
| "loss": 0.5649, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 0.4423510137210731, | |
| "grad_norm": 3.515625, | |
| "learning_rate": 9.072777744373073e-06, | |
| "loss": 0.5508, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 0.4443989350808929, | |
| "grad_norm": 3.265625, | |
| "learning_rate": 9.063136487729397e-06, | |
| "loss": 0.5607, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 0.4464468564407127, | |
| "grad_norm": 3.546875, | |
| "learning_rate": 9.05345054362432e-06, | |
| "loss": 0.5629, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 0.4484947778005325, | |
| "grad_norm": 4.09375, | |
| "learning_rate": 9.043720018586447e-06, | |
| "loss": 0.574, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 0.4505426991603522, | |
| "grad_norm": 4.125, | |
| "learning_rate": 9.033945019634693e-06, | |
| "loss": 0.5668, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.452590620520172, | |
| "grad_norm": 3.609375, | |
| "learning_rate": 9.024125654277111e-06, | |
| "loss": 0.5291, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 0.4546385418799918, | |
| "grad_norm": 4.1875, | |
| "learning_rate": 9.014262030509704e-06, | |
| "loss": 0.6214, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 0.4566864632398116, | |
| "grad_norm": 4.125, | |
| "learning_rate": 9.004354256815247e-06, | |
| "loss": 0.6003, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 0.4587343845996314, | |
| "grad_norm": 3.375, | |
| "learning_rate": 8.994402442162083e-06, | |
| "loss": 0.5938, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 0.46078230595945113, | |
| "grad_norm": 3.3125, | |
| "learning_rate": 8.984406696002925e-06, | |
| "loss": 0.5887, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 0.4628302273192709, | |
| "grad_norm": 3.84375, | |
| "learning_rate": 8.97436712827367e-06, | |
| "loss": 0.6317, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 0.4648781486790907, | |
| "grad_norm": 3.59375, | |
| "learning_rate": 8.964283849392163e-06, | |
| "loss": 0.5547, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 0.4669260700389105, | |
| "grad_norm": 3.515625, | |
| "learning_rate": 8.954156970257001e-06, | |
| "loss": 0.6202, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 0.4689739913987303, | |
| "grad_norm": 3.9375, | |
| "learning_rate": 8.94398660224631e-06, | |
| "loss": 0.6002, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 0.4710219127585501, | |
| "grad_norm": 3.9375, | |
| "learning_rate": 8.933772857216517e-06, | |
| "loss": 0.5653, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.47306983411836984, | |
| "grad_norm": 3.203125, | |
| "learning_rate": 8.923515847501117e-06, | |
| "loss": 0.624, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 0.47511775547818963, | |
| "grad_norm": 3.5, | |
| "learning_rate": 8.913215685909448e-06, | |
| "loss": 0.5667, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 0.4771656768380094, | |
| "grad_norm": 3.8125, | |
| "learning_rate": 8.90287248572544e-06, | |
| "loss": 0.5623, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 0.4792135981978292, | |
| "grad_norm": 3.640625, | |
| "learning_rate": 8.892486360706377e-06, | |
| "loss": 0.5846, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 0.481261519557649, | |
| "grad_norm": 4.25, | |
| "learning_rate": 8.882057425081635e-06, | |
| "loss": 0.5705, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 0.48330944091746875, | |
| "grad_norm": 3.96875, | |
| "learning_rate": 8.871585793551435e-06, | |
| "loss": 0.5893, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 0.48535736227728854, | |
| "grad_norm": 4.84375, | |
| "learning_rate": 8.861071581285584e-06, | |
| "loss": 0.6106, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 0.48740528363710833, | |
| "grad_norm": 3.71875, | |
| "learning_rate": 8.850514903922198e-06, | |
| "loss": 0.5863, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 0.4894532049969281, | |
| "grad_norm": 3.4375, | |
| "learning_rate": 8.839915877566438e-06, | |
| "loss": 0.5634, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 0.4915011263567479, | |
| "grad_norm": 3.765625, | |
| "learning_rate": 8.829274618789228e-06, | |
| "loss": 0.5406, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.49354904771656766, | |
| "grad_norm": 3.671875, | |
| "learning_rate": 8.818591244625977e-06, | |
| "loss": 0.5016, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 0.49559696907638745, | |
| "grad_norm": 3.484375, | |
| "learning_rate": 8.807865872575294e-06, | |
| "loss": 0.6582, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 0.49764489043620724, | |
| "grad_norm": 3.515625, | |
| "learning_rate": 8.797098620597683e-06, | |
| "loss": 0.573, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 0.49969281179602704, | |
| "grad_norm": 4.0625, | |
| "learning_rate": 8.786289607114262e-06, | |
| "loss": 0.6161, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 0.5017407331558468, | |
| "grad_norm": 3.953125, | |
| "learning_rate": 8.775438951005453e-06, | |
| "loss": 0.5801, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 0.5037886545156666, | |
| "grad_norm": 3.046875, | |
| "learning_rate": 8.764546771609672e-06, | |
| "loss": 0.5537, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 0.5058365758754864, | |
| "grad_norm": 4.03125, | |
| "learning_rate": 8.753613188722018e-06, | |
| "loss": 0.5411, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 0.5078844972353062, | |
| "grad_norm": 3.828125, | |
| "learning_rate": 8.74263832259296e-06, | |
| "loss": 0.531, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 0.509932418595126, | |
| "grad_norm": 3.78125, | |
| "learning_rate": 8.73162229392701e-06, | |
| "loss": 0.5652, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 0.5119803399549457, | |
| "grad_norm": 4.03125, | |
| "learning_rate": 8.7205652238814e-06, | |
| "loss": 0.6105, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.5140282613147655, | |
| "grad_norm": 4.125, | |
| "learning_rate": 8.709467234064738e-06, | |
| "loss": 0.5952, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 0.5160761826745853, | |
| "grad_norm": 4.125, | |
| "learning_rate": 8.698328446535683e-06, | |
| "loss": 0.5135, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 0.5181241040344051, | |
| "grad_norm": 4.28125, | |
| "learning_rate": 8.687148983801599e-06, | |
| "loss": 0.6145, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 0.5201720253942249, | |
| "grad_norm": 3.9375, | |
| "learning_rate": 8.675928968817206e-06, | |
| "loss": 0.5855, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 0.5222199467540446, | |
| "grad_norm": 3.875, | |
| "learning_rate": 8.664668524983232e-06, | |
| "loss": 0.6286, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 0.5242678681138644, | |
| "grad_norm": 3.0, | |
| "learning_rate": 8.653367776145045e-06, | |
| "loss": 0.5642, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 0.5263157894736842, | |
| "grad_norm": 3.0625, | |
| "learning_rate": 8.642026846591303e-06, | |
| "loss": 0.5474, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 0.528363710833504, | |
| "grad_norm": 3.734375, | |
| "learning_rate": 8.63064586105258e-06, | |
| "loss": 0.6154, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 0.5304116321933238, | |
| "grad_norm": 4.375, | |
| "learning_rate": 8.619224944700003e-06, | |
| "loss": 0.5802, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 0.5324595535531436, | |
| "grad_norm": 3.546875, | |
| "learning_rate": 8.607764223143854e-06, | |
| "loss": 0.5435, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.5345074749129634, | |
| "grad_norm": 3.015625, | |
| "learning_rate": 8.596263822432222e-06, | |
| "loss": 0.585, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 0.5365553962727831, | |
| "grad_norm": 3.171875, | |
| "learning_rate": 8.584723869049586e-06, | |
| "loss": 0.5417, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 0.5386033176326029, | |
| "grad_norm": 4.1875, | |
| "learning_rate": 8.573144489915437e-06, | |
| "loss": 0.6276, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 0.5406512389924227, | |
| "grad_norm": 3.5625, | |
| "learning_rate": 8.561525812382882e-06, | |
| "loss": 0.597, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 0.5426991603522425, | |
| "grad_norm": 3.265625, | |
| "learning_rate": 8.549867964237246e-06, | |
| "loss": 0.5478, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 0.5447470817120622, | |
| "grad_norm": 3.859375, | |
| "learning_rate": 8.538171073694656e-06, | |
| "loss": 0.5764, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 0.546795003071882, | |
| "grad_norm": 3.78125, | |
| "learning_rate": 8.526435269400644e-06, | |
| "loss": 0.5692, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 0.5488429244317018, | |
| "grad_norm": 3.359375, | |
| "learning_rate": 8.514660680428723e-06, | |
| "loss": 0.5599, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 0.5508908457915216, | |
| "grad_norm": 3.671875, | |
| "learning_rate": 8.50284743627897e-06, | |
| "loss": 0.6127, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 0.5529387671513414, | |
| "grad_norm": 4.1875, | |
| "learning_rate": 8.490995666876602e-06, | |
| "loss": 0.654, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.5549866885111612, | |
| "grad_norm": 5.1875, | |
| "learning_rate": 8.47910550257055e-06, | |
| "loss": 0.5806, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 0.557034609870981, | |
| "grad_norm": 3.984375, | |
| "learning_rate": 8.467177074132017e-06, | |
| "loss": 0.5577, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 0.5590825312308008, | |
| "grad_norm": 2.984375, | |
| "learning_rate": 8.455210512753047e-06, | |
| "loss": 0.5119, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 0.5611304525906206, | |
| "grad_norm": 3.59375, | |
| "learning_rate": 8.443205950045086e-06, | |
| "loss": 0.5889, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 0.5631783739504403, | |
| "grad_norm": 3.3125, | |
| "learning_rate": 8.431163518037523e-06, | |
| "loss": 0.5612, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 0.5652262953102601, | |
| "grad_norm": 3.21875, | |
| "learning_rate": 8.41908334917624e-06, | |
| "loss": 0.5988, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 0.5672742166700798, | |
| "grad_norm": 3.9375, | |
| "learning_rate": 8.40696557632217e-06, | |
| "loss": 0.5484, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 0.5693221380298996, | |
| "grad_norm": 4.34375, | |
| "learning_rate": 8.394810332749814e-06, | |
| "loss": 0.5484, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 0.5713700593897194, | |
| "grad_norm": 4.9375, | |
| "learning_rate": 8.382617752145794e-06, | |
| "loss": 0.5683, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 0.5734179807495392, | |
| "grad_norm": 5.03125, | |
| "learning_rate": 8.370387968607369e-06, | |
| "loss": 0.5616, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.575465902109359, | |
| "grad_norm": 3.515625, | |
| "learning_rate": 8.358121116640971e-06, | |
| "loss": 0.5814, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 0.5775138234691788, | |
| "grad_norm": 3.65625, | |
| "learning_rate": 8.345817331160714e-06, | |
| "loss": 0.5634, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 0.5795617448289986, | |
| "grad_norm": 3.6875, | |
| "learning_rate": 8.333476747486922e-06, | |
| "loss": 0.6005, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 0.5816096661888184, | |
| "grad_norm": 3.390625, | |
| "learning_rate": 8.321099501344634e-06, | |
| "loss": 0.5763, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 0.5836575875486382, | |
| "grad_norm": 3.46875, | |
| "learning_rate": 8.308685728862111e-06, | |
| "loss": 0.5533, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 0.585705508908458, | |
| "grad_norm": 4.09375, | |
| "learning_rate": 8.296235566569343e-06, | |
| "loss": 0.5745, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 0.5877534302682776, | |
| "grad_norm": 3.640625, | |
| "learning_rate": 8.283749151396542e-06, | |
| "loss": 0.5044, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 0.5898013516280974, | |
| "grad_norm": 3.453125, | |
| "learning_rate": 8.27122662067264e-06, | |
| "loss": 0.5412, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 0.5918492729879172, | |
| "grad_norm": 3.765625, | |
| "learning_rate": 8.258668112123781e-06, | |
| "loss": 0.5889, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 0.593897194347737, | |
| "grad_norm": 4.375, | |
| "learning_rate": 8.246073763871796e-06, | |
| "loss": 0.5874, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.5959451157075568, | |
| "grad_norm": 3.265625, | |
| "learning_rate": 8.233443714432698e-06, | |
| "loss": 0.5976, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 0.5979930370673766, | |
| "grad_norm": 4.21875, | |
| "learning_rate": 8.220778102715145e-06, | |
| "loss": 0.6134, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 0.6000409584271964, | |
| "grad_norm": 4.03125, | |
| "learning_rate": 8.20807706801892e-06, | |
| "loss": 0.5816, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 0.6020888797870162, | |
| "grad_norm": 3.578125, | |
| "learning_rate": 8.195340750033401e-06, | |
| "loss": 0.5833, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 0.604136801146836, | |
| "grad_norm": 3.9375, | |
| "learning_rate": 8.182569288836015e-06, | |
| "loss": 0.5768, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 0.6061847225066558, | |
| "grad_norm": 3.65625, | |
| "learning_rate": 8.16976282489071e-06, | |
| "loss": 0.578, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 0.6082326438664756, | |
| "grad_norm": 3.21875, | |
| "learning_rate": 8.1569214990464e-06, | |
| "loss": 0.461, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 0.6102805652262953, | |
| "grad_norm": 3.09375, | |
| "learning_rate": 8.144045452535418e-06, | |
| "loss": 0.5564, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 0.612328486586115, | |
| "grad_norm": 3.828125, | |
| "learning_rate": 8.131134826971967e-06, | |
| "loss": 0.5592, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 0.6143764079459348, | |
| "grad_norm": 4.4375, | |
| "learning_rate": 8.118189764350557e-06, | |
| "loss": 0.6086, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.6143764079459348, | |
| "eval_loss": 0.564007043838501, | |
| "eval_runtime": 49.3837, | |
| "eval_samples_per_second": 83.266, | |
| "eval_steps_per_second": 41.633, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.6164243293057546, | |
| "grad_norm": 3.890625, | |
| "learning_rate": 8.105210407044452e-06, | |
| "loss": 0.6133, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 0.6184722506655744, | |
| "grad_norm": 3.734375, | |
| "learning_rate": 8.09219689780409e-06, | |
| "loss": 0.5601, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 0.6205201720253942, | |
| "grad_norm": 3.828125, | |
| "learning_rate": 8.079149379755525e-06, | |
| "loss": 0.569, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 0.622568093385214, | |
| "grad_norm": 4.375, | |
| "learning_rate": 8.066067996398853e-06, | |
| "loss": 0.5513, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 0.6246160147450338, | |
| "grad_norm": 4.4375, | |
| "learning_rate": 8.052952891606617e-06, | |
| "loss": 0.5664, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 0.6266639361048536, | |
| "grad_norm": 4.09375, | |
| "learning_rate": 8.039804209622255e-06, | |
| "loss": 0.5716, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 0.6287118574646734, | |
| "grad_norm": 3.6875, | |
| "learning_rate": 8.026622095058483e-06, | |
| "loss": 0.5099, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 0.6307597788244932, | |
| "grad_norm": 4.125, | |
| "learning_rate": 8.01340669289572e-06, | |
| "loss": 0.5344, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 0.6328077001843129, | |
| "grad_norm": 3.78125, | |
| "learning_rate": 8.000158148480497e-06, | |
| "loss": 0.5317, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 0.6348556215441327, | |
| "grad_norm": 3.25, | |
| "learning_rate": 7.986876607523842e-06, | |
| "loss": 0.5768, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.6369035429039525, | |
| "grad_norm": 6.375, | |
| "learning_rate": 7.973562216099694e-06, | |
| "loss": 0.5755, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 0.6389514642637723, | |
| "grad_norm": 3.78125, | |
| "learning_rate": 7.960215120643294e-06, | |
| "loss": 0.5833, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 0.640999385623592, | |
| "grad_norm": 4.1875, | |
| "learning_rate": 7.946835467949561e-06, | |
| "loss": 0.5504, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 0.6430473069834118, | |
| "grad_norm": 4.0, | |
| "learning_rate": 7.933423405171496e-06, | |
| "loss": 0.5595, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 0.6450952283432316, | |
| "grad_norm": 3.984375, | |
| "learning_rate": 7.91997907981855e-06, | |
| "loss": 0.5771, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 0.6471431497030514, | |
| "grad_norm": 3.953125, | |
| "learning_rate": 7.906502639755008e-06, | |
| "loss": 0.5923, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 0.6491910710628712, | |
| "grad_norm": 3.0, | |
| "learning_rate": 7.892994233198361e-06, | |
| "loss": 0.5906, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 0.651238992422691, | |
| "grad_norm": 3.671875, | |
| "learning_rate": 7.879454008717675e-06, | |
| "loss": 0.5744, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 0.6532869137825108, | |
| "grad_norm": 3.78125, | |
| "learning_rate": 7.865882115231959e-06, | |
| "loss": 0.5737, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 0.6553348351423305, | |
| "grad_norm": 4.1875, | |
| "learning_rate": 7.852278702008527e-06, | |
| "loss": 0.5927, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.6573827565021503, | |
| "grad_norm": 3.625, | |
| "learning_rate": 7.838643918661354e-06, | |
| "loss": 0.5658, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 0.6594306778619701, | |
| "grad_norm": 3.046875, | |
| "learning_rate": 7.824977915149432e-06, | |
| "loss": 0.5321, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 0.6614785992217899, | |
| "grad_norm": 3.390625, | |
| "learning_rate": 7.811280841775124e-06, | |
| "loss": 0.5156, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 0.6635265205816097, | |
| "grad_norm": 3.578125, | |
| "learning_rate": 7.797552849182502e-06, | |
| "loss": 0.5588, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 0.6655744419414295, | |
| "grad_norm": 3.90625, | |
| "learning_rate": 7.783794088355698e-06, | |
| "loss": 0.6302, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 0.6676223633012492, | |
| "grad_norm": 3.1875, | |
| "learning_rate": 7.770004710617247e-06, | |
| "loss": 0.5842, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 0.669670284661069, | |
| "grad_norm": 3.015625, | |
| "learning_rate": 7.756184867626407e-06, | |
| "loss": 0.5592, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 0.6717182060208888, | |
| "grad_norm": 3.625, | |
| "learning_rate": 7.742334711377509e-06, | |
| "loss": 0.5674, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 0.6737661273807086, | |
| "grad_norm": 3.90625, | |
| "learning_rate": 7.728454394198271e-06, | |
| "loss": 0.6026, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 0.6758140487405284, | |
| "grad_norm": 3.265625, | |
| "learning_rate": 7.714544068748137e-06, | |
| "loss": 0.5144, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.6778619701003481, | |
| "grad_norm": 3.625, | |
| "learning_rate": 7.700603888016583e-06, | |
| "loss": 0.5396, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 0.6799098914601679, | |
| "grad_norm": 3.796875, | |
| "learning_rate": 7.686634005321442e-06, | |
| "loss": 0.5889, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 0.6819578128199877, | |
| "grad_norm": 2.921875, | |
| "learning_rate": 7.672634574307223e-06, | |
| "loss": 0.5414, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 0.6840057341798075, | |
| "grad_norm": 2.78125, | |
| "learning_rate": 7.658605748943407e-06, | |
| "loss": 0.4579, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 0.6860536555396273, | |
| "grad_norm": 4.5625, | |
| "learning_rate": 7.644547683522767e-06, | |
| "loss": 0.5635, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 0.6881015768994471, | |
| "grad_norm": 3.59375, | |
| "learning_rate": 7.630460532659667e-06, | |
| "loss": 0.555, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 0.6901494982592669, | |
| "grad_norm": 4.375, | |
| "learning_rate": 7.616344451288357e-06, | |
| "loss": 0.5808, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 0.6921974196190867, | |
| "grad_norm": 3.28125, | |
| "learning_rate": 7.602199594661273e-06, | |
| "loss": 0.5623, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 0.6942453409789064, | |
| "grad_norm": 3.46875, | |
| "learning_rate": 7.588026118347334e-06, | |
| "loss": 0.5326, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 0.6962932623387262, | |
| "grad_norm": 2.84375, | |
| "learning_rate": 7.5738241782302156e-06, | |
| "loss": 0.5501, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.698341183698546, | |
| "grad_norm": 3.96875, | |
| "learning_rate": 7.559593930506658e-06, | |
| "loss": 0.5593, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 0.7003891050583657, | |
| "grad_norm": 2.90625, | |
| "learning_rate": 7.545335531684725e-06, | |
| "loss": 0.5492, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 0.7024370264181855, | |
| "grad_norm": 4.96875, | |
| "learning_rate": 7.531049138582105e-06, | |
| "loss": 0.5853, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 0.7044849477780053, | |
| "grad_norm": 3.59375, | |
| "learning_rate": 7.516734908324363e-06, | |
| "loss": 0.5875, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 0.7065328691378251, | |
| "grad_norm": 3.453125, | |
| "learning_rate": 7.5023929983432344e-06, | |
| "loss": 0.5459, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 0.7085807904976449, | |
| "grad_norm": 3.171875, | |
| "learning_rate": 7.488023566374878e-06, | |
| "loss": 0.4829, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 0.7106287118574647, | |
| "grad_norm": 4.71875, | |
| "learning_rate": 7.4736267704581485e-06, | |
| "loss": 0.6025, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 0.7126766332172845, | |
| "grad_norm": 2.796875, | |
| "learning_rate": 7.459202768932856e-06, | |
| "loss": 0.5715, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 0.7147245545771043, | |
| "grad_norm": 3.21875, | |
| "learning_rate": 7.444751720438026e-06, | |
| "loss": 0.5352, | |
| "step": 3490 | |
| }, | |
| { | |
| "epoch": 0.7167724759369241, | |
| "grad_norm": 3.953125, | |
| "learning_rate": 7.430273783910153e-06, | |
| "loss": 0.554, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.7188203972967439, | |
| "grad_norm": 3.375, | |
| "learning_rate": 7.415769118581456e-06, | |
| "loss": 0.4954, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 0.7208683186565635, | |
| "grad_norm": 3.859375, | |
| "learning_rate": 7.40123788397812e-06, | |
| "loss": 0.5202, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 0.7229162400163833, | |
| "grad_norm": 3.828125, | |
| "learning_rate": 7.386680239918548e-06, | |
| "loss": 0.5294, | |
| "step": 3530 | |
| }, | |
| { | |
| "epoch": 0.7249641613762031, | |
| "grad_norm": 4.09375, | |
| "learning_rate": 7.372096346511602e-06, | |
| "loss": 0.5507, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 0.7270120827360229, | |
| "grad_norm": 3.0625, | |
| "learning_rate": 7.357486364154842e-06, | |
| "loss": 0.5977, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 0.7290600040958427, | |
| "grad_norm": 2.34375, | |
| "learning_rate": 7.3428504535327585e-06, | |
| "loss": 0.5502, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 0.7311079254556625, | |
| "grad_norm": 3.453125, | |
| "learning_rate": 7.328188775615009e-06, | |
| "loss": 0.5709, | |
| "step": 3570 | |
| }, | |
| { | |
| "epoch": 0.7331558468154823, | |
| "grad_norm": 3.265625, | |
| "learning_rate": 7.313501491654649e-06, | |
| "loss": 0.5813, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 0.7352037681753021, | |
| "grad_norm": 3.375, | |
| "learning_rate": 7.298788763186353e-06, | |
| "loss": 0.5048, | |
| "step": 3590 | |
| }, | |
| { | |
| "epoch": 0.7372516895351219, | |
| "grad_norm": 3.3125, | |
| "learning_rate": 7.284050752024643e-06, | |
| "loss": 0.5842, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.7392996108949417, | |
| "grad_norm": 3.515625, | |
| "learning_rate": 7.269287620262105e-06, | |
| "loss": 0.5087, | |
| "step": 3610 | |
| }, | |
| { | |
| "epoch": 0.7413475322547615, | |
| "grad_norm": 3.578125, | |
| "learning_rate": 7.25449953026761e-06, | |
| "loss": 0.5671, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 0.7433954536145811, | |
| "grad_norm": 3.90625, | |
| "learning_rate": 7.239686644684525e-06, | |
| "loss": 0.5711, | |
| "step": 3630 | |
| }, | |
| { | |
| "epoch": 0.7454433749744009, | |
| "grad_norm": 3.703125, | |
| "learning_rate": 7.224849126428925e-06, | |
| "loss": 0.5572, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 0.7474912963342207, | |
| "grad_norm": 4.5, | |
| "learning_rate": 7.209987138687806e-06, | |
| "loss": 0.543, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 0.7495392176940405, | |
| "grad_norm": 3.5, | |
| "learning_rate": 7.195100844917281e-06, | |
| "loss": 0.5518, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 0.7515871390538603, | |
| "grad_norm": 2.875, | |
| "learning_rate": 7.180190408840788e-06, | |
| "loss": 0.5479, | |
| "step": 3670 | |
| }, | |
| { | |
| "epoch": 0.7536350604136801, | |
| "grad_norm": 3.953125, | |
| "learning_rate": 7.165255994447288e-06, | |
| "loss": 0.5755, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 0.7556829817734999, | |
| "grad_norm": 4.0, | |
| "learning_rate": 7.150297765989467e-06, | |
| "loss": 0.6021, | |
| "step": 3690 | |
| }, | |
| { | |
| "epoch": 0.7577309031333197, | |
| "grad_norm": 3.734375, | |
| "learning_rate": 7.135315887981918e-06, | |
| "loss": 0.5618, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.7597788244931395, | |
| "grad_norm": 3.375, | |
| "learning_rate": 7.120310525199341e-06, | |
| "loss": 0.5302, | |
| "step": 3710 | |
| }, | |
| { | |
| "epoch": 0.7618267458529593, | |
| "grad_norm": 3.015625, | |
| "learning_rate": 7.105281842674729e-06, | |
| "loss": 0.4933, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 0.7638746672127791, | |
| "grad_norm": 3.140625, | |
| "learning_rate": 7.090230005697552e-06, | |
| "loss": 0.588, | |
| "step": 3730 | |
| }, | |
| { | |
| "epoch": 0.7659225885725988, | |
| "grad_norm": 3.546875, | |
| "learning_rate": 7.075155179811931e-06, | |
| "loss": 0.5628, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 0.7679705099324186, | |
| "grad_norm": 3.265625, | |
| "learning_rate": 7.0600575308148375e-06, | |
| "loss": 0.5523, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 0.7700184312922383, | |
| "grad_norm": 3.359375, | |
| "learning_rate": 7.044937224754249e-06, | |
| "loss": 0.5648, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 0.7720663526520581, | |
| "grad_norm": 3.46875, | |
| "learning_rate": 7.029794427927337e-06, | |
| "loss": 0.5495, | |
| "step": 3770 | |
| }, | |
| { | |
| "epoch": 0.7741142740118779, | |
| "grad_norm": 3.65625, | |
| "learning_rate": 7.0146293068786255e-06, | |
| "loss": 0.6386, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 0.7761621953716977, | |
| "grad_norm": 3.828125, | |
| "learning_rate": 6.999442028398171e-06, | |
| "loss": 0.5634, | |
| "step": 3790 | |
| }, | |
| { | |
| "epoch": 0.7782101167315175, | |
| "grad_norm": 4.34375, | |
| "learning_rate": 6.984232759519725e-06, | |
| "loss": 0.4956, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.7802580380913373, | |
| "grad_norm": 3.84375, | |
| "learning_rate": 6.969001667518891e-06, | |
| "loss": 0.5745, | |
| "step": 3810 | |
| }, | |
| { | |
| "epoch": 0.7823059594511571, | |
| "grad_norm": 3.171875, | |
| "learning_rate": 6.9537489199112915e-06, | |
| "loss": 0.5519, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 0.7843538808109769, | |
| "grad_norm": 3.1875, | |
| "learning_rate": 6.938474684450725e-06, | |
| "loss": 0.5338, | |
| "step": 3830 | |
| }, | |
| { | |
| "epoch": 0.7864018021707967, | |
| "grad_norm": 3.859375, | |
| "learning_rate": 6.923179129127312e-06, | |
| "loss": 0.641, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 0.7884497235306164, | |
| "grad_norm": 3.25, | |
| "learning_rate": 6.907862422165661e-06, | |
| "loss": 0.4959, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 0.7904976448904362, | |
| "grad_norm": 4.09375, | |
| "learning_rate": 6.892524732023013e-06, | |
| "loss": 0.5589, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 0.792545566250256, | |
| "grad_norm": 4.28125, | |
| "learning_rate": 6.877166227387382e-06, | |
| "loss": 0.6053, | |
| "step": 3870 | |
| }, | |
| { | |
| "epoch": 0.7945934876100758, | |
| "grad_norm": 3.125, | |
| "learning_rate": 6.86178707717571e-06, | |
| "loss": 0.5434, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 0.7966414089698955, | |
| "grad_norm": 4.0625, | |
| "learning_rate": 6.8463874505320004e-06, | |
| "loss": 0.5558, | |
| "step": 3890 | |
| }, | |
| { | |
| "epoch": 0.7986893303297153, | |
| "grad_norm": 4.15625, | |
| "learning_rate": 6.8309675168254664e-06, | |
| "loss": 0.5358, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.8007372516895351, | |
| "grad_norm": 3.921875, | |
| "learning_rate": 6.81552744564866e-06, | |
| "loss": 0.5228, | |
| "step": 3910 | |
| }, | |
| { | |
| "epoch": 0.8027851730493549, | |
| "grad_norm": 3.328125, | |
| "learning_rate": 6.800067406815612e-06, | |
| "loss": 0.5956, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 0.8048330944091747, | |
| "grad_norm": 3.140625, | |
| "learning_rate": 6.784587570359965e-06, | |
| "loss": 0.5785, | |
| "step": 3930 | |
| }, | |
| { | |
| "epoch": 0.8068810157689945, | |
| "grad_norm": 4.21875, | |
| "learning_rate": 6.769088106533097e-06, | |
| "loss": 0.5481, | |
| "step": 3940 | |
| }, | |
| { | |
| "epoch": 0.8089289371288143, | |
| "grad_norm": 4.21875, | |
| "learning_rate": 6.753569185802256e-06, | |
| "loss": 0.5554, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 0.810976858488634, | |
| "grad_norm": 4.125, | |
| "learning_rate": 6.73803097884868e-06, | |
| "loss": 0.5738, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 0.8130247798484538, | |
| "grad_norm": 2.984375, | |
| "learning_rate": 6.722473656565724e-06, | |
| "loss": 0.5275, | |
| "step": 3970 | |
| }, | |
| { | |
| "epoch": 0.8150727012082736, | |
| "grad_norm": 4.0625, | |
| "learning_rate": 6.706897390056979e-06, | |
| "loss": 0.597, | |
| "step": 3980 | |
| }, | |
| { | |
| "epoch": 0.8171206225680934, | |
| "grad_norm": 4.40625, | |
| "learning_rate": 6.691302350634386e-06, | |
| "loss": 0.5808, | |
| "step": 3990 | |
| }, | |
| { | |
| "epoch": 0.8191685439279132, | |
| "grad_norm": 2.9375, | |
| "learning_rate": 6.675688709816362e-06, | |
| "loss": 0.5217, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.8191685439279132, | |
| "eval_loss": 0.5564442873001099, | |
| "eval_runtime": 48.7272, | |
| "eval_samples_per_second": 84.388, | |
| "eval_steps_per_second": 42.194, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.821216465287733, | |
| "grad_norm": 3.203125, | |
| "learning_rate": 6.660056639325898e-06, | |
| "loss": 0.5558, | |
| "step": 4010 | |
| }, | |
| { | |
| "epoch": 0.8232643866475527, | |
| "grad_norm": 3.5, | |
| "learning_rate": 6.644406311088687e-06, | |
| "loss": 0.5654, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 0.8253123080073725, | |
| "grad_norm": 3.640625, | |
| "learning_rate": 6.628737897231226e-06, | |
| "loss": 0.5095, | |
| "step": 4030 | |
| }, | |
| { | |
| "epoch": 0.8273602293671923, | |
| "grad_norm": 2.9375, | |
| "learning_rate": 6.613051570078914e-06, | |
| "loss": 0.5459, | |
| "step": 4040 | |
| }, | |
| { | |
| "epoch": 0.8294081507270121, | |
| "grad_norm": 4.96875, | |
| "learning_rate": 6.597347502154177e-06, | |
| "loss": 0.5534, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 0.8314560720868319, | |
| "grad_norm": 3.875, | |
| "learning_rate": 6.581625866174548e-06, | |
| "loss": 0.598, | |
| "step": 4060 | |
| }, | |
| { | |
| "epoch": 0.8335039934466516, | |
| "grad_norm": 3.84375, | |
| "learning_rate": 6.565886835050785e-06, | |
| "loss": 0.5429, | |
| "step": 4070 | |
| }, | |
| { | |
| "epoch": 0.8355519148064714, | |
| "grad_norm": 3.4375, | |
| "learning_rate": 6.550130581884958e-06, | |
| "loss": 0.5431, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 0.8375998361662912, | |
| "grad_norm": 3.65625, | |
| "learning_rate": 6.534357279968554e-06, | |
| "loss": 0.5966, | |
| "step": 4090 | |
| }, | |
| { | |
| "epoch": 0.839647757526111, | |
| "grad_norm": 3.71875, | |
| "learning_rate": 6.5185671027805655e-06, | |
| "loss": 0.5669, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 0.8416956788859308, | |
| "grad_norm": 3.28125, | |
| "learning_rate": 6.5027602239855805e-06, | |
| "loss": 0.5071, | |
| "step": 4110 | |
| }, | |
| { | |
| "epoch": 0.8437436002457506, | |
| "grad_norm": 3.75, | |
| "learning_rate": 6.486936817431878e-06, | |
| "loss": 0.5571, | |
| "step": 4120 | |
| }, | |
| { | |
| "epoch": 0.8457915216055704, | |
| "grad_norm": 4.03125, | |
| "learning_rate": 6.471097057149516e-06, | |
| "loss": 0.6112, | |
| "step": 4130 | |
| }, | |
| { | |
| "epoch": 0.8478394429653902, | |
| "grad_norm": 3.21875, | |
| "learning_rate": 6.45524111734841e-06, | |
| "loss": 0.5302, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 0.84988736432521, | |
| "grad_norm": 4.4375, | |
| "learning_rate": 6.439369172416428e-06, | |
| "loss": 0.5725, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 0.8519352856850297, | |
| "grad_norm": 3.5625, | |
| "learning_rate": 6.4234813969174615e-06, | |
| "loss": 0.5502, | |
| "step": 4160 | |
| }, | |
| { | |
| "epoch": 0.8539832070448494, | |
| "grad_norm": 3.453125, | |
| "learning_rate": 6.407577965589515e-06, | |
| "loss": 0.5261, | |
| "step": 4170 | |
| }, | |
| { | |
| "epoch": 0.8560311284046692, | |
| "grad_norm": 3.46875, | |
| "learning_rate": 6.391659053342778e-06, | |
| "loss": 0.5446, | |
| "step": 4180 | |
| }, | |
| { | |
| "epoch": 0.858079049764489, | |
| "grad_norm": 3.65625, | |
| "learning_rate": 6.3757248352577015e-06, | |
| "loss": 0.6036, | |
| "step": 4190 | |
| }, | |
| { | |
| "epoch": 0.8601269711243088, | |
| "grad_norm": 3.203125, | |
| "learning_rate": 6.359775486583077e-06, | |
| "loss": 0.5096, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.8621748924841286, | |
| "grad_norm": 4.0, | |
| "learning_rate": 6.343811182734108e-06, | |
| "loss": 0.5413, | |
| "step": 4210 | |
| }, | |
| { | |
| "epoch": 0.8642228138439484, | |
| "grad_norm": 3.953125, | |
| "learning_rate": 6.327832099290473e-06, | |
| "loss": 0.5094, | |
| "step": 4220 | |
| }, | |
| { | |
| "epoch": 0.8662707352037682, | |
| "grad_norm": 3.90625, | |
| "learning_rate": 6.311838411994404e-06, | |
| "loss": 0.5392, | |
| "step": 4230 | |
| }, | |
| { | |
| "epoch": 0.868318656563588, | |
| "grad_norm": 3.484375, | |
| "learning_rate": 6.295830296748753e-06, | |
| "loss": 0.6021, | |
| "step": 4240 | |
| }, | |
| { | |
| "epoch": 0.8703665779234078, | |
| "grad_norm": 3.171875, | |
| "learning_rate": 6.279807929615047e-06, | |
| "loss": 0.5416, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 0.8724144992832276, | |
| "grad_norm": 3.796875, | |
| "learning_rate": 6.263771486811567e-06, | |
| "loss": 0.5066, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 0.8744624206430474, | |
| "grad_norm": 3.890625, | |
| "learning_rate": 6.2477211447113975e-06, | |
| "loss": 0.5029, | |
| "step": 4270 | |
| }, | |
| { | |
| "epoch": 0.876510342002867, | |
| "grad_norm": 4.03125, | |
| "learning_rate": 6.231657079840491e-06, | |
| "loss": 0.5448, | |
| "step": 4280 | |
| }, | |
| { | |
| "epoch": 0.8785582633626868, | |
| "grad_norm": 3.4375, | |
| "learning_rate": 6.215579468875729e-06, | |
| "loss": 0.5299, | |
| "step": 4290 | |
| }, | |
| { | |
| "epoch": 0.8806061847225066, | |
| "grad_norm": 3.53125, | |
| "learning_rate": 6.199488488642975e-06, | |
| "loss": 0.5806, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 0.8826541060823264, | |
| "grad_norm": 3.375, | |
| "learning_rate": 6.18338431611513e-06, | |
| "loss": 0.5305, | |
| "step": 4310 | |
| }, | |
| { | |
| "epoch": 0.8847020274421462, | |
| "grad_norm": 4.15625, | |
| "learning_rate": 6.167267128410191e-06, | |
| "loss": 0.5216, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 0.886749948801966, | |
| "grad_norm": 3.375, | |
| "learning_rate": 6.151137102789297e-06, | |
| "loss": 0.5515, | |
| "step": 4330 | |
| }, | |
| { | |
| "epoch": 0.8887978701617858, | |
| "grad_norm": 3.0625, | |
| "learning_rate": 6.1349944166547825e-06, | |
| "loss": 0.5276, | |
| "step": 4340 | |
| }, | |
| { | |
| "epoch": 0.8908457915216056, | |
| "grad_norm": 3.171875, | |
| "learning_rate": 6.118839247548226e-06, | |
| "loss": 0.5427, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 0.8928937128814254, | |
| "grad_norm": 3.171875, | |
| "learning_rate": 6.102671773148494e-06, | |
| "loss": 0.5531, | |
| "step": 4360 | |
| }, | |
| { | |
| "epoch": 0.8949416342412452, | |
| "grad_norm": 3.4375, | |
| "learning_rate": 6.086492171269794e-06, | |
| "loss": 0.509, | |
| "step": 4370 | |
| }, | |
| { | |
| "epoch": 0.896989555601065, | |
| "grad_norm": 4.5625, | |
| "learning_rate": 6.070300619859715e-06, | |
| "loss": 0.5731, | |
| "step": 4380 | |
| }, | |
| { | |
| "epoch": 0.8990374769608847, | |
| "grad_norm": 3.1875, | |
| "learning_rate": 6.054097296997266e-06, | |
| "loss": 0.5113, | |
| "step": 4390 | |
| }, | |
| { | |
| "epoch": 0.9010853983207044, | |
| "grad_norm": 5.03125, | |
| "learning_rate": 6.037882380890924e-06, | |
| "loss": 0.5583, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 0.9031333196805242, | |
| "grad_norm": 3.421875, | |
| "learning_rate": 6.021656049876672e-06, | |
| "loss": 0.5341, | |
| "step": 4410 | |
| }, | |
| { | |
| "epoch": 0.905181241040344, | |
| "grad_norm": 4.0, | |
| "learning_rate": 6.005418482416036e-06, | |
| "loss": 0.5671, | |
| "step": 4420 | |
| }, | |
| { | |
| "epoch": 0.9072291624001638, | |
| "grad_norm": 3.96875, | |
| "learning_rate": 5.989169857094126e-06, | |
| "loss": 0.5846, | |
| "step": 4430 | |
| }, | |
| { | |
| "epoch": 0.9092770837599836, | |
| "grad_norm": 3.203125, | |
| "learning_rate": 5.972910352617667e-06, | |
| "loss": 0.552, | |
| "step": 4440 | |
| }, | |
| { | |
| "epoch": 0.9113250051198034, | |
| "grad_norm": 3.78125, | |
| "learning_rate": 5.956640147813035e-06, | |
| "loss": 0.5681, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 0.9133729264796232, | |
| "grad_norm": 4.0, | |
| "learning_rate": 5.940359421624292e-06, | |
| "loss": 0.5831, | |
| "step": 4460 | |
| }, | |
| { | |
| "epoch": 0.915420847839443, | |
| "grad_norm": 3.21875, | |
| "learning_rate": 5.92406835311122e-06, | |
| "loss": 0.5516, | |
| "step": 4470 | |
| }, | |
| { | |
| "epoch": 0.9174687691992628, | |
| "grad_norm": 3.703125, | |
| "learning_rate": 5.907767121447343e-06, | |
| "loss": 0.5779, | |
| "step": 4480 | |
| }, | |
| { | |
| "epoch": 0.9195166905590826, | |
| "grad_norm": 4.09375, | |
| "learning_rate": 5.891455905917968e-06, | |
| "loss": 0.6099, | |
| "step": 4490 | |
| }, | |
| { | |
| "epoch": 0.9215646119189023, | |
| "grad_norm": 3.34375, | |
| "learning_rate": 5.8751348859182025e-06, | |
| "loss": 0.5506, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.9236125332787221, | |
| "grad_norm": 4.03125, | |
| "learning_rate": 5.858804240950988e-06, | |
| "loss": 0.533, | |
| "step": 4510 | |
| }, | |
| { | |
| "epoch": 0.9256604546385419, | |
| "grad_norm": 4.03125, | |
| "learning_rate": 5.842464150625125e-06, | |
| "loss": 0.5653, | |
| "step": 4520 | |
| }, | |
| { | |
| "epoch": 0.9277083759983616, | |
| "grad_norm": 3.484375, | |
| "learning_rate": 5.826114794653294e-06, | |
| "loss": 0.5573, | |
| "step": 4530 | |
| }, | |
| { | |
| "epoch": 0.9297562973581814, | |
| "grad_norm": 3.6875, | |
| "learning_rate": 5.809756352850083e-06, | |
| "loss": 0.5227, | |
| "step": 4540 | |
| }, | |
| { | |
| "epoch": 0.9318042187180012, | |
| "grad_norm": 3.734375, | |
| "learning_rate": 5.793389005130011e-06, | |
| "loss": 0.538, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 0.933852140077821, | |
| "grad_norm": 3.765625, | |
| "learning_rate": 5.7770129315055435e-06, | |
| "loss": 0.514, | |
| "step": 4560 | |
| }, | |
| { | |
| "epoch": 0.9359000614376408, | |
| "grad_norm": 3.75, | |
| "learning_rate": 5.760628312085114e-06, | |
| "loss": 0.5736, | |
| "step": 4570 | |
| }, | |
| { | |
| "epoch": 0.9379479827974606, | |
| "grad_norm": 4.28125, | |
| "learning_rate": 5.744235327071151e-06, | |
| "loss": 0.5473, | |
| "step": 4580 | |
| }, | |
| { | |
| "epoch": 0.9399959041572804, | |
| "grad_norm": 4.1875, | |
| "learning_rate": 5.727834156758082e-06, | |
| "loss": 0.5422, | |
| "step": 4590 | |
| }, | |
| { | |
| "epoch": 0.9420438255171002, | |
| "grad_norm": 3.203125, | |
| "learning_rate": 5.711424981530367e-06, | |
| "loss": 0.5134, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 0.9440917468769199, | |
| "grad_norm": 3.453125, | |
| "learning_rate": 5.6950079818605e-06, | |
| "loss": 0.5406, | |
| "step": 4610 | |
| }, | |
| { | |
| "epoch": 0.9461396682367397, | |
| "grad_norm": 4.5, | |
| "learning_rate": 5.678583338307033e-06, | |
| "loss": 0.6124, | |
| "step": 4620 | |
| }, | |
| { | |
| "epoch": 0.9481875895965595, | |
| "grad_norm": 3.59375, | |
| "learning_rate": 5.662151231512588e-06, | |
| "loss": 0.5095, | |
| "step": 4630 | |
| }, | |
| { | |
| "epoch": 0.9502355109563793, | |
| "grad_norm": 5.15625, | |
| "learning_rate": 5.645711842201869e-06, | |
| "loss": 0.5337, | |
| "step": 4640 | |
| }, | |
| { | |
| "epoch": 0.952283432316199, | |
| "grad_norm": 3.4375, | |
| "learning_rate": 5.629265351179673e-06, | |
| "loss": 0.5252, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 0.9543313536760188, | |
| "grad_norm": 3.390625, | |
| "learning_rate": 5.612811939328907e-06, | |
| "loss": 0.5631, | |
| "step": 4660 | |
| }, | |
| { | |
| "epoch": 0.9563792750358386, | |
| "grad_norm": 3.78125, | |
| "learning_rate": 5.596351787608592e-06, | |
| "loss": 0.5531, | |
| "step": 4670 | |
| }, | |
| { | |
| "epoch": 0.9584271963956584, | |
| "grad_norm": 3.671875, | |
| "learning_rate": 5.579885077051879e-06, | |
| "loss": 0.5673, | |
| "step": 4680 | |
| }, | |
| { | |
| "epoch": 0.9604751177554782, | |
| "grad_norm": 4.0625, | |
| "learning_rate": 5.5634119887640515e-06, | |
| "loss": 0.5721, | |
| "step": 4690 | |
| }, | |
| { | |
| "epoch": 0.962523039115298, | |
| "grad_norm": 3.875, | |
| "learning_rate": 5.546932703920539e-06, | |
| "loss": 0.577, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 0.9645709604751177, | |
| "grad_norm": 3.5625, | |
| "learning_rate": 5.530447403764924e-06, | |
| "loss": 0.4924, | |
| "step": 4710 | |
| }, | |
| { | |
| "epoch": 0.9666188818349375, | |
| "grad_norm": 3.875, | |
| "learning_rate": 5.513956269606944e-06, | |
| "loss": 0.5552, | |
| "step": 4720 | |
| }, | |
| { | |
| "epoch": 0.9686668031947573, | |
| "grad_norm": 3.8125, | |
| "learning_rate": 5.497459482820499e-06, | |
| "loss": 0.6076, | |
| "step": 4730 | |
| }, | |
| { | |
| "epoch": 0.9707147245545771, | |
| "grad_norm": 3.65625, | |
| "learning_rate": 5.480957224841666e-06, | |
| "loss": 0.5824, | |
| "step": 4740 | |
| }, | |
| { | |
| "epoch": 0.9727626459143969, | |
| "grad_norm": 4.59375, | |
| "learning_rate": 5.4644496771666845e-06, | |
| "loss": 0.5634, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 0.9748105672742167, | |
| "grad_norm": 4.3125, | |
| "learning_rate": 5.447937021349979e-06, | |
| "loss": 0.5331, | |
| "step": 4760 | |
| }, | |
| { | |
| "epoch": 0.9768584886340365, | |
| "grad_norm": 3.4375, | |
| "learning_rate": 5.431419439002155e-06, | |
| "loss": 0.5859, | |
| "step": 4770 | |
| }, | |
| { | |
| "epoch": 0.9789064099938563, | |
| "grad_norm": 2.765625, | |
| "learning_rate": 5.414897111787998e-06, | |
| "loss": 0.5416, | |
| "step": 4780 | |
| }, | |
| { | |
| "epoch": 0.980954331353676, | |
| "grad_norm": 3.3125, | |
| "learning_rate": 5.3983702214244805e-06, | |
| "loss": 0.5383, | |
| "step": 4790 | |
| }, | |
| { | |
| "epoch": 0.9830022527134958, | |
| "grad_norm": 2.859375, | |
| "learning_rate": 5.381838949678759e-06, | |
| "loss": 0.5247, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 0.9850501740733156, | |
| "grad_norm": 4.21875, | |
| "learning_rate": 5.365303478366184e-06, | |
| "loss": 0.5983, | |
| "step": 4810 | |
| }, | |
| { | |
| "epoch": 0.9870980954331353, | |
| "grad_norm": 3.96875, | |
| "learning_rate": 5.348763989348285e-06, | |
| "loss": 0.5884, | |
| "step": 4820 | |
| }, | |
| { | |
| "epoch": 0.9891460167929551, | |
| "grad_norm": 3.5, | |
| "learning_rate": 5.332220664530788e-06, | |
| "loss": 0.5408, | |
| "step": 4830 | |
| }, | |
| { | |
| "epoch": 0.9911939381527749, | |
| "grad_norm": 3.421875, | |
| "learning_rate": 5.315673685861602e-06, | |
| "loss": 0.4796, | |
| "step": 4840 | |
| }, | |
| { | |
| "epoch": 0.9932418595125947, | |
| "grad_norm": 4.0625, | |
| "learning_rate": 5.29912323532882e-06, | |
| "loss": 0.5175, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 0.9952897808724145, | |
| "grad_norm": 3.515625, | |
| "learning_rate": 5.282569494958723e-06, | |
| "loss": 0.5127, | |
| "step": 4860 | |
| }, | |
| { | |
| "epoch": 0.9973377022322343, | |
| "grad_norm": 3.796875, | |
| "learning_rate": 5.266012646813774e-06, | |
| "loss": 0.5912, | |
| "step": 4870 | |
| }, | |
| { | |
| "epoch": 0.9993856235920541, | |
| "grad_norm": 4.21875, | |
| "learning_rate": 5.249452872990616e-06, | |
| "loss": 0.5986, | |
| "step": 4880 | |
| }, | |
| { | |
| "epoch": 1.0014335449518739, | |
| "grad_norm": 4.21875, | |
| "learning_rate": 5.2328903556180666e-06, | |
| "loss": 0.5311, | |
| "step": 4890 | |
| }, | |
| { | |
| "epoch": 1.0034814663116935, | |
| "grad_norm": 3.15625, | |
| "learning_rate": 5.216325276855122e-06, | |
| "loss": 0.5112, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 1.0055293876715135, | |
| "grad_norm": 3.53125, | |
| "learning_rate": 5.1997578188889465e-06, | |
| "loss": 0.4965, | |
| "step": 4910 | |
| }, | |
| { | |
| "epoch": 1.0075773090313331, | |
| "grad_norm": 3.46875, | |
| "learning_rate": 5.183188163932873e-06, | |
| "loss": 0.4898, | |
| "step": 4920 | |
| }, | |
| { | |
| "epoch": 1.009625230391153, | |
| "grad_norm": 3.359375, | |
| "learning_rate": 5.166616494224393e-06, | |
| "loss": 0.5006, | |
| "step": 4930 | |
| }, | |
| { | |
| "epoch": 1.0116731517509727, | |
| "grad_norm": 3.625, | |
| "learning_rate": 5.150042992023166e-06, | |
| "loss": 0.5103, | |
| "step": 4940 | |
| }, | |
| { | |
| "epoch": 1.0137210731107926, | |
| "grad_norm": 3.609375, | |
| "learning_rate": 5.133467839608998e-06, | |
| "loss": 0.5185, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 1.0157689944706123, | |
| "grad_norm": 3.140625, | |
| "learning_rate": 5.116891219279846e-06, | |
| "loss": 0.517, | |
| "step": 4960 | |
| }, | |
| { | |
| "epoch": 1.0178169158304322, | |
| "grad_norm": 4.25, | |
| "learning_rate": 5.1003133133498115e-06, | |
| "loss": 0.5011, | |
| "step": 4970 | |
| }, | |
| { | |
| "epoch": 1.019864837190252, | |
| "grad_norm": 3.578125, | |
| "learning_rate": 5.083734304147139e-06, | |
| "loss": 0.5249, | |
| "step": 4980 | |
| }, | |
| { | |
| "epoch": 1.0219127585500716, | |
| "grad_norm": 3.875, | |
| "learning_rate": 5.067154374012201e-06, | |
| "loss": 0.5371, | |
| "step": 4990 | |
| }, | |
| { | |
| "epoch": 1.0239606799098915, | |
| "grad_norm": 3.21875, | |
| "learning_rate": 5.050573705295504e-06, | |
| "loss": 0.5436, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 1.0239606799098915, | |
| "eval_loss": 0.5540497303009033, | |
| "eval_runtime": 48.9146, | |
| "eval_samples_per_second": 84.065, | |
| "eval_steps_per_second": 42.032, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 1.0260086012697112, | |
| "grad_norm": 3.75, | |
| "learning_rate": 5.033992480355675e-06, | |
| "loss": 0.5154, | |
| "step": 5010 | |
| }, | |
| { | |
| "epoch": 1.028056522629531, | |
| "grad_norm": 3.4375, | |
| "learning_rate": 5.01741088155746e-06, | |
| "loss": 0.4631, | |
| "step": 5020 | |
| }, | |
| { | |
| "epoch": 1.0301044439893507, | |
| "grad_norm": 3.75, | |
| "learning_rate": 5.000829091269713e-06, | |
| "loss": 0.4978, | |
| "step": 5030 | |
| }, | |
| { | |
| "epoch": 1.0321523653491707, | |
| "grad_norm": 3.828125, | |
| "learning_rate": 4.984247291863399e-06, | |
| "loss": 0.5375, | |
| "step": 5040 | |
| }, | |
| { | |
| "epoch": 1.0342002867089903, | |
| "grad_norm": 3.796875, | |
| "learning_rate": 4.967665665709582e-06, | |
| "loss": 0.4977, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 1.0362482080688102, | |
| "grad_norm": 3.390625, | |
| "learning_rate": 4.951084395177416e-06, | |
| "loss": 0.4551, | |
| "step": 5060 | |
| }, | |
| { | |
| "epoch": 1.03829612942863, | |
| "grad_norm": 4.125, | |
| "learning_rate": 4.93450366263215e-06, | |
| "loss": 0.4873, | |
| "step": 5070 | |
| }, | |
| { | |
| "epoch": 1.0403440507884498, | |
| "grad_norm": 3.90625, | |
| "learning_rate": 4.917923650433116e-06, | |
| "loss": 0.4607, | |
| "step": 5080 | |
| }, | |
| { | |
| "epoch": 1.0423919721482695, | |
| "grad_norm": 4.34375, | |
| "learning_rate": 4.901344540931715e-06, | |
| "loss": 0.5109, | |
| "step": 5090 | |
| }, | |
| { | |
| "epoch": 1.0444398935080892, | |
| "grad_norm": 4.25, | |
| "learning_rate": 4.884766516469433e-06, | |
| "loss": 0.4653, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 1.046487814867909, | |
| "grad_norm": 3.625, | |
| "learning_rate": 4.868189759375807e-06, | |
| "loss": 0.5088, | |
| "step": 5110 | |
| }, | |
| { | |
| "epoch": 1.0485357362277288, | |
| "grad_norm": 3.921875, | |
| "learning_rate": 4.851614451966451e-06, | |
| "loss": 0.498, | |
| "step": 5120 | |
| }, | |
| { | |
| "epoch": 1.0505836575875487, | |
| "grad_norm": 3.53125, | |
| "learning_rate": 4.835040776541022e-06, | |
| "loss": 0.5035, | |
| "step": 5130 | |
| }, | |
| { | |
| "epoch": 1.0526315789473684, | |
| "grad_norm": 3.3125, | |
| "learning_rate": 4.818468915381236e-06, | |
| "loss": 0.4899, | |
| "step": 5140 | |
| }, | |
| { | |
| "epoch": 1.0546795003071883, | |
| "grad_norm": 3.734375, | |
| "learning_rate": 4.801899050748854e-06, | |
| "loss": 0.524, | |
| "step": 5150 | |
| }, | |
| { | |
| "epoch": 1.056727421667008, | |
| "grad_norm": 4.125, | |
| "learning_rate": 4.785331364883673e-06, | |
| "loss": 0.5365, | |
| "step": 5160 | |
| }, | |
| { | |
| "epoch": 1.0587753430268279, | |
| "grad_norm": 3.375, | |
| "learning_rate": 4.768766040001536e-06, | |
| "loss": 0.5169, | |
| "step": 5170 | |
| }, | |
| { | |
| "epoch": 1.0608232643866475, | |
| "grad_norm": 4.78125, | |
| "learning_rate": 4.752203258292312e-06, | |
| "loss": 0.5385, | |
| "step": 5180 | |
| }, | |
| { | |
| "epoch": 1.0628711857464674, | |
| "grad_norm": 3.234375, | |
| "learning_rate": 4.735643201917903e-06, | |
| "loss": 0.4745, | |
| "step": 5190 | |
| }, | |
| { | |
| "epoch": 1.0649191071062871, | |
| "grad_norm": 3.84375, | |
| "learning_rate": 4.7190860530102385e-06, | |
| "loss": 0.5175, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 1.0669670284661068, | |
| "grad_norm": 3.9375, | |
| "learning_rate": 4.702531993669265e-06, | |
| "loss": 0.5022, | |
| "step": 5210 | |
| }, | |
| { | |
| "epoch": 1.0690149498259267, | |
| "grad_norm": 3.921875, | |
| "learning_rate": 4.685981205960958e-06, | |
| "loss": 0.4885, | |
| "step": 5220 | |
| }, | |
| { | |
| "epoch": 1.0710628711857464, | |
| "grad_norm": 3.703125, | |
| "learning_rate": 4.669433871915301e-06, | |
| "loss": 0.5109, | |
| "step": 5230 | |
| }, | |
| { | |
| "epoch": 1.0731107925455663, | |
| "grad_norm": 3.78125, | |
| "learning_rate": 4.6528901735243e-06, | |
| "loss": 0.4801, | |
| "step": 5240 | |
| }, | |
| { | |
| "epoch": 1.075158713905386, | |
| "grad_norm": 4.125, | |
| "learning_rate": 4.636350292739974e-06, | |
| "loss": 0.5051, | |
| "step": 5250 | |
| }, | |
| { | |
| "epoch": 1.0772066352652059, | |
| "grad_norm": 3.71875, | |
| "learning_rate": 4.6198144114723506e-06, | |
| "loss": 0.488, | |
| "step": 5260 | |
| }, | |
| { | |
| "epoch": 1.0792545566250256, | |
| "grad_norm": 3.875, | |
| "learning_rate": 4.603282711587478e-06, | |
| "loss": 0.5006, | |
| "step": 5270 | |
| }, | |
| { | |
| "epoch": 1.0813024779848455, | |
| "grad_norm": 4.25, | |
| "learning_rate": 4.586755374905405e-06, | |
| "loss": 0.4973, | |
| "step": 5280 | |
| }, | |
| { | |
| "epoch": 1.0833503993446651, | |
| "grad_norm": 3.8125, | |
| "learning_rate": 4.570232583198206e-06, | |
| "loss": 0.5142, | |
| "step": 5290 | |
| }, | |
| { | |
| "epoch": 1.0853983207044848, | |
| "grad_norm": 3.625, | |
| "learning_rate": 4.553714518187955e-06, | |
| "loss": 0.4319, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 1.0874462420643047, | |
| "grad_norm": 4.03125, | |
| "learning_rate": 4.537201361544751e-06, | |
| "loss": 0.49, | |
| "step": 5310 | |
| }, | |
| { | |
| "epoch": 1.0894941634241244, | |
| "grad_norm": 3.609375, | |
| "learning_rate": 4.5206932948847075e-06, | |
| "loss": 0.5164, | |
| "step": 5320 | |
| }, | |
| { | |
| "epoch": 1.0915420847839443, | |
| "grad_norm": 4.09375, | |
| "learning_rate": 4.50419049976795e-06, | |
| "loss": 0.5184, | |
| "step": 5330 | |
| }, | |
| { | |
| "epoch": 1.093590006143764, | |
| "grad_norm": 3.4375, | |
| "learning_rate": 4.487693157696637e-06, | |
| "loss": 0.4799, | |
| "step": 5340 | |
| }, | |
| { | |
| "epoch": 1.095637927503584, | |
| "grad_norm": 3.625, | |
| "learning_rate": 4.471201450112942e-06, | |
| "loss": 0.4984, | |
| "step": 5350 | |
| }, | |
| { | |
| "epoch": 1.0976858488634036, | |
| "grad_norm": 4.21875, | |
| "learning_rate": 4.454715558397076e-06, | |
| "loss": 0.5458, | |
| "step": 5360 | |
| }, | |
| { | |
| "epoch": 1.0997337702232235, | |
| "grad_norm": 3.5, | |
| "learning_rate": 4.438235663865288e-06, | |
| "loss": 0.4502, | |
| "step": 5370 | |
| }, | |
| { | |
| "epoch": 1.1017816915830432, | |
| "grad_norm": 3.125, | |
| "learning_rate": 4.421761947767856e-06, | |
| "loss": 0.4981, | |
| "step": 5380 | |
| }, | |
| { | |
| "epoch": 1.103829612942863, | |
| "grad_norm": 3.40625, | |
| "learning_rate": 4.405294591287122e-06, | |
| "loss": 0.4796, | |
| "step": 5390 | |
| }, | |
| { | |
| "epoch": 1.1058775343026828, | |
| "grad_norm": 3.359375, | |
| "learning_rate": 4.388833775535469e-06, | |
| "loss": 0.4936, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 1.1079254556625027, | |
| "grad_norm": 3.265625, | |
| "learning_rate": 4.372379681553352e-06, | |
| "loss": 0.4897, | |
| "step": 5410 | |
| }, | |
| { | |
| "epoch": 1.1099733770223223, | |
| "grad_norm": 3.328125, | |
| "learning_rate": 4.3559324903072985e-06, | |
| "loss": 0.506, | |
| "step": 5420 | |
| }, | |
| { | |
| "epoch": 1.112021298382142, | |
| "grad_norm": 3.65625, | |
| "learning_rate": 4.3394923826879096e-06, | |
| "loss": 0.5046, | |
| "step": 5430 | |
| }, | |
| { | |
| "epoch": 1.114069219741962, | |
| "grad_norm": 3.734375, | |
| "learning_rate": 4.32305953950789e-06, | |
| "loss": 0.5027, | |
| "step": 5440 | |
| }, | |
| { | |
| "epoch": 1.1161171411017816, | |
| "grad_norm": 3.671875, | |
| "learning_rate": 4.306634141500037e-06, | |
| "loss": 0.5006, | |
| "step": 5450 | |
| }, | |
| { | |
| "epoch": 1.1181650624616015, | |
| "grad_norm": 4.15625, | |
| "learning_rate": 4.290216369315274e-06, | |
| "loss": 0.5329, | |
| "step": 5460 | |
| }, | |
| { | |
| "epoch": 1.1202129838214212, | |
| "grad_norm": 4.1875, | |
| "learning_rate": 4.273806403520644e-06, | |
| "loss": 0.4763, | |
| "step": 5470 | |
| }, | |
| { | |
| "epoch": 1.122260905181241, | |
| "grad_norm": 3.6875, | |
| "learning_rate": 4.257404424597342e-06, | |
| "loss": 0.5092, | |
| "step": 5480 | |
| }, | |
| { | |
| "epoch": 1.1243088265410608, | |
| "grad_norm": 3.03125, | |
| "learning_rate": 4.241010612938719e-06, | |
| "loss": 0.4719, | |
| "step": 5490 | |
| }, | |
| { | |
| "epoch": 1.1263567479008807, | |
| "grad_norm": 3.21875, | |
| "learning_rate": 4.224625148848292e-06, | |
| "loss": 0.4901, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 1.1284046692607004, | |
| "grad_norm": 3.296875, | |
| "learning_rate": 4.208248212537783e-06, | |
| "loss": 0.4323, | |
| "step": 5510 | |
| }, | |
| { | |
| "epoch": 1.13045259062052, | |
| "grad_norm": 4.40625, | |
| "learning_rate": 4.19187998412511e-06, | |
| "loss": 0.4783, | |
| "step": 5520 | |
| }, | |
| { | |
| "epoch": 1.13250051198034, | |
| "grad_norm": 3.65625, | |
| "learning_rate": 4.175520643632428e-06, | |
| "loss": 0.5215, | |
| "step": 5530 | |
| }, | |
| { | |
| "epoch": 1.1345484333401596, | |
| "grad_norm": 5.09375, | |
| "learning_rate": 4.15917037098414e-06, | |
| "loss": 0.4825, | |
| "step": 5540 | |
| }, | |
| { | |
| "epoch": 1.1365963546999795, | |
| "grad_norm": 4.0625, | |
| "learning_rate": 4.142829346004911e-06, | |
| "loss": 0.4754, | |
| "step": 5550 | |
| }, | |
| { | |
| "epoch": 1.1386442760597992, | |
| "grad_norm": 3.96875, | |
| "learning_rate": 4.126497748417708e-06, | |
| "loss": 0.4967, | |
| "step": 5560 | |
| }, | |
| { | |
| "epoch": 1.1406921974196191, | |
| "grad_norm": 3.4375, | |
| "learning_rate": 4.110175757841802e-06, | |
| "loss": 0.5174, | |
| "step": 5570 | |
| }, | |
| { | |
| "epoch": 1.1427401187794388, | |
| "grad_norm": 4.25, | |
| "learning_rate": 4.093863553790813e-06, | |
| "loss": 0.5289, | |
| "step": 5580 | |
| }, | |
| { | |
| "epoch": 1.1447880401392587, | |
| "grad_norm": 4.09375, | |
| "learning_rate": 4.077561315670721e-06, | |
| "loss": 0.5478, | |
| "step": 5590 | |
| }, | |
| { | |
| "epoch": 1.1468359614990784, | |
| "grad_norm": 3.59375, | |
| "learning_rate": 4.061269222777898e-06, | |
| "loss": 0.4968, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 1.1488838828588983, | |
| "grad_norm": 4.3125, | |
| "learning_rate": 4.044987454297142e-06, | |
| "loss": 0.4841, | |
| "step": 5610 | |
| }, | |
| { | |
| "epoch": 1.150931804218718, | |
| "grad_norm": 3.625, | |
| "learning_rate": 4.028716189299691e-06, | |
| "loss": 0.4725, | |
| "step": 5620 | |
| }, | |
| { | |
| "epoch": 1.152979725578538, | |
| "grad_norm": 4.4375, | |
| "learning_rate": 4.01245560674127e-06, | |
| "loss": 0.5094, | |
| "step": 5630 | |
| }, | |
| { | |
| "epoch": 1.1550276469383576, | |
| "grad_norm": 3.265625, | |
| "learning_rate": 3.996205885460112e-06, | |
| "loss": 0.5184, | |
| "step": 5640 | |
| }, | |
| { | |
| "epoch": 1.1570755682981773, | |
| "grad_norm": 3.9375, | |
| "learning_rate": 3.979967204174998e-06, | |
| "loss": 0.4978, | |
| "step": 5650 | |
| }, | |
| { | |
| "epoch": 1.1591234896579972, | |
| "grad_norm": 4.09375, | |
| "learning_rate": 3.963739741483285e-06, | |
| "loss": 0.4841, | |
| "step": 5660 | |
| }, | |
| { | |
| "epoch": 1.1611714110178168, | |
| "grad_norm": 3.9375, | |
| "learning_rate": 3.947523675858945e-06, | |
| "loss": 0.5245, | |
| "step": 5670 | |
| }, | |
| { | |
| "epoch": 1.1632193323776367, | |
| "grad_norm": 3.703125, | |
| "learning_rate": 3.931319185650606e-06, | |
| "loss": 0.4784, | |
| "step": 5680 | |
| }, | |
| { | |
| "epoch": 1.1652672537374564, | |
| "grad_norm": 3.703125, | |
| "learning_rate": 3.915126449079582e-06, | |
| "loss": 0.5205, | |
| "step": 5690 | |
| }, | |
| { | |
| "epoch": 1.1673151750972763, | |
| "grad_norm": 4.6875, | |
| "learning_rate": 3.898945644237919e-06, | |
| "loss": 0.5475, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 1.169363096457096, | |
| "grad_norm": 3.96875, | |
| "learning_rate": 3.882776949086436e-06, | |
| "loss": 0.5308, | |
| "step": 5710 | |
| }, | |
| { | |
| "epoch": 1.171411017816916, | |
| "grad_norm": 4.28125, | |
| "learning_rate": 3.866620541452764e-06, | |
| "loss": 0.516, | |
| "step": 5720 | |
| }, | |
| { | |
| "epoch": 1.1734589391767356, | |
| "grad_norm": 3.125, | |
| "learning_rate": 3.850476599029394e-06, | |
| "loss": 0.4846, | |
| "step": 5730 | |
| }, | |
| { | |
| "epoch": 1.1755068605365553, | |
| "grad_norm": 4.0625, | |
| "learning_rate": 3.834345299371719e-06, | |
| "loss": 0.481, | |
| "step": 5740 | |
| }, | |
| { | |
| "epoch": 1.1775547818963752, | |
| "grad_norm": 4.0, | |
| "learning_rate": 3.818226819896089e-06, | |
| "loss": 0.4901, | |
| "step": 5750 | |
| }, | |
| { | |
| "epoch": 1.1796027032561949, | |
| "grad_norm": 3.953125, | |
| "learning_rate": 3.802121337877848e-06, | |
| "loss": 0.5132, | |
| "step": 5760 | |
| }, | |
| { | |
| "epoch": 1.1816506246160148, | |
| "grad_norm": 4.15625, | |
| "learning_rate": 3.7860290304493953e-06, | |
| "loss": 0.4929, | |
| "step": 5770 | |
| }, | |
| { | |
| "epoch": 1.1836985459758345, | |
| "grad_norm": 4.84375, | |
| "learning_rate": 3.7699500745982287e-06, | |
| "loss": 0.5142, | |
| "step": 5780 | |
| }, | |
| { | |
| "epoch": 1.1857464673356544, | |
| "grad_norm": 3.5, | |
| "learning_rate": 3.7538846471650038e-06, | |
| "loss": 0.5093, | |
| "step": 5790 | |
| }, | |
| { | |
| "epoch": 1.187794388695474, | |
| "grad_norm": 2.859375, | |
| "learning_rate": 3.737832924841587e-06, | |
| "loss": 0.4884, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 1.189842310055294, | |
| "grad_norm": 3.625, | |
| "learning_rate": 3.72179508416911e-06, | |
| "loss": 0.4842, | |
| "step": 5810 | |
| }, | |
| { | |
| "epoch": 1.1918902314151136, | |
| "grad_norm": 4.09375, | |
| "learning_rate": 3.7057713015360365e-06, | |
| "loss": 0.505, | |
| "step": 5820 | |
| }, | |
| { | |
| "epoch": 1.1939381527749335, | |
| "grad_norm": 4.0625, | |
| "learning_rate": 3.6897617531762086e-06, | |
| "loss": 0.5513, | |
| "step": 5830 | |
| }, | |
| { | |
| "epoch": 1.1959860741347532, | |
| "grad_norm": 3.46875, | |
| "learning_rate": 3.6737666151669206e-06, | |
| "loss": 0.5029, | |
| "step": 5840 | |
| }, | |
| { | |
| "epoch": 1.1980339954945731, | |
| "grad_norm": 4.53125, | |
| "learning_rate": 3.657786063426977e-06, | |
| "loss": 0.4956, | |
| "step": 5850 | |
| }, | |
| { | |
| "epoch": 1.2000819168543928, | |
| "grad_norm": 3.1875, | |
| "learning_rate": 3.6418202737147566e-06, | |
| "loss": 0.4914, | |
| "step": 5860 | |
| }, | |
| { | |
| "epoch": 1.2021298382142125, | |
| "grad_norm": 3.671875, | |
| "learning_rate": 3.6258694216262845e-06, | |
| "loss": 0.5057, | |
| "step": 5870 | |
| }, | |
| { | |
| "epoch": 1.2041777595740324, | |
| "grad_norm": 3.796875, | |
| "learning_rate": 3.609933682593299e-06, | |
| "loss": 0.4909, | |
| "step": 5880 | |
| }, | |
| { | |
| "epoch": 1.206225680933852, | |
| "grad_norm": 3.609375, | |
| "learning_rate": 3.5940132318813137e-06, | |
| "loss": 0.5094, | |
| "step": 5890 | |
| }, | |
| { | |
| "epoch": 1.208273602293672, | |
| "grad_norm": 3.328125, | |
| "learning_rate": 3.578108244587705e-06, | |
| "loss": 0.5162, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 1.2103215236534917, | |
| "grad_norm": 3.625, | |
| "learning_rate": 3.562218895639775e-06, | |
| "loss": 0.5171, | |
| "step": 5910 | |
| }, | |
| { | |
| "epoch": 1.2123694450133116, | |
| "grad_norm": 3.875, | |
| "learning_rate": 3.5463453597928306e-06, | |
| "loss": 0.5411, | |
| "step": 5920 | |
| }, | |
| { | |
| "epoch": 1.2144173663731312, | |
| "grad_norm": 3.96875, | |
| "learning_rate": 3.5304878116282654e-06, | |
| "loss": 0.4972, | |
| "step": 5930 | |
| }, | |
| { | |
| "epoch": 1.2164652877329512, | |
| "grad_norm": 4.375, | |
| "learning_rate": 3.51464642555163e-06, | |
| "loss": 0.5189, | |
| "step": 5940 | |
| }, | |
| { | |
| "epoch": 1.2185132090927708, | |
| "grad_norm": 3.78125, | |
| "learning_rate": 3.498821375790728e-06, | |
| "loss": 0.5105, | |
| "step": 5950 | |
| }, | |
| { | |
| "epoch": 1.2205611304525905, | |
| "grad_norm": 4.59375, | |
| "learning_rate": 3.4830128363936835e-06, | |
| "loss": 0.4979, | |
| "step": 5960 | |
| }, | |
| { | |
| "epoch": 1.2226090518124104, | |
| "grad_norm": 4.65625, | |
| "learning_rate": 3.467220981227042e-06, | |
| "loss": 0.5311, | |
| "step": 5970 | |
| }, | |
| { | |
| "epoch": 1.22465697317223, | |
| "grad_norm": 3.359375, | |
| "learning_rate": 3.451445983973848e-06, | |
| "loss": 0.4564, | |
| "step": 5980 | |
| }, | |
| { | |
| "epoch": 1.22670489453205, | |
| "grad_norm": 4.125, | |
| "learning_rate": 3.4356880181317377e-06, | |
| "loss": 0.5113, | |
| "step": 5990 | |
| }, | |
| { | |
| "epoch": 1.2287528158918697, | |
| "grad_norm": 4.1875, | |
| "learning_rate": 3.419947257011036e-06, | |
| "loss": 0.5245, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 1.2287528158918697, | |
| "eval_loss": 0.5531573295593262, | |
| "eval_runtime": 48.5342, | |
| "eval_samples_per_second": 84.724, | |
| "eval_steps_per_second": 42.362, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 1.2308007372516896, | |
| "grad_norm": 4.34375, | |
| "learning_rate": 3.404223873732837e-06, | |
| "loss": 0.5338, | |
| "step": 6010 | |
| }, | |
| { | |
| "epoch": 1.2328486586115093, | |
| "grad_norm": 3.90625, | |
| "learning_rate": 3.3885180412271183e-06, | |
| "loss": 0.5234, | |
| "step": 6020 | |
| }, | |
| { | |
| "epoch": 1.2348965799713292, | |
| "grad_norm": 3.328125, | |
| "learning_rate": 3.372829932230821e-06, | |
| "loss": 0.5337, | |
| "step": 6030 | |
| }, | |
| { | |
| "epoch": 1.2369445013311489, | |
| "grad_norm": 3.96875, | |
| "learning_rate": 3.357159719285964e-06, | |
| "loss": 0.5321, | |
| "step": 6040 | |
| }, | |
| { | |
| "epoch": 1.2389924226909685, | |
| "grad_norm": 4.15625, | |
| "learning_rate": 3.3415075747377414e-06, | |
| "loss": 0.5118, | |
| "step": 6050 | |
| }, | |
| { | |
| "epoch": 1.2410403440507884, | |
| "grad_norm": 3.453125, | |
| "learning_rate": 3.325873670732619e-06, | |
| "loss": 0.5142, | |
| "step": 6060 | |
| }, | |
| { | |
| "epoch": 1.2430882654106084, | |
| "grad_norm": 4.1875, | |
| "learning_rate": 3.3102581792164566e-06, | |
| "loss": 0.4996, | |
| "step": 6070 | |
| }, | |
| { | |
| "epoch": 1.245136186770428, | |
| "grad_norm": 4.9375, | |
| "learning_rate": 3.294661271932601e-06, | |
| "loss": 0.5165, | |
| "step": 6080 | |
| }, | |
| { | |
| "epoch": 1.2471841081302477, | |
| "grad_norm": 4.5, | |
| "learning_rate": 3.2790831204200113e-06, | |
| "loss": 0.5101, | |
| "step": 6090 | |
| }, | |
| { | |
| "epoch": 1.2492320294900676, | |
| "grad_norm": 3.21875, | |
| "learning_rate": 3.2635238960113634e-06, | |
| "loss": 0.5147, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 1.2512799508498873, | |
| "grad_norm": 3.75, | |
| "learning_rate": 3.2479837698311646e-06, | |
| "loss": 0.5224, | |
| "step": 6110 | |
| }, | |
| { | |
| "epoch": 1.2533278722097072, | |
| "grad_norm": 3.375, | |
| "learning_rate": 3.23246291279388e-06, | |
| "loss": 0.4954, | |
| "step": 6120 | |
| }, | |
| { | |
| "epoch": 1.2553757935695269, | |
| "grad_norm": 4.3125, | |
| "learning_rate": 3.2169614956020423e-06, | |
| "loss": 0.4957, | |
| "step": 6130 | |
| }, | |
| { | |
| "epoch": 1.2574237149293468, | |
| "grad_norm": 3.390625, | |
| "learning_rate": 3.2014796887443854e-06, | |
| "loss": 0.4907, | |
| "step": 6140 | |
| }, | |
| { | |
| "epoch": 1.2594716362891665, | |
| "grad_norm": 3.46875, | |
| "learning_rate": 3.1860176624939566e-06, | |
| "loss": 0.4758, | |
| "step": 6150 | |
| }, | |
| { | |
| "epoch": 1.2615195576489864, | |
| "grad_norm": 3.65625, | |
| "learning_rate": 3.1705755869062553e-06, | |
| "loss": 0.4639, | |
| "step": 6160 | |
| }, | |
| { | |
| "epoch": 1.263567479008806, | |
| "grad_norm": 3.6875, | |
| "learning_rate": 3.1551536318173613e-06, | |
| "loss": 0.5144, | |
| "step": 6170 | |
| }, | |
| { | |
| "epoch": 1.2656154003686257, | |
| "grad_norm": 4.25, | |
| "learning_rate": 3.139751966842054e-06, | |
| "loss": 0.5526, | |
| "step": 6180 | |
| }, | |
| { | |
| "epoch": 1.2676633217284456, | |
| "grad_norm": 4.625, | |
| "learning_rate": 3.124370761371968e-06, | |
| "loss": 0.4946, | |
| "step": 6190 | |
| }, | |
| { | |
| "epoch": 1.2697112430882653, | |
| "grad_norm": 3.984375, | |
| "learning_rate": 3.1090101845737084e-06, | |
| "loss": 0.4808, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 1.2717591644480852, | |
| "grad_norm": 3.875, | |
| "learning_rate": 3.0936704053870083e-06, | |
| "loss": 0.5139, | |
| "step": 6210 | |
| }, | |
| { | |
| "epoch": 1.273807085807905, | |
| "grad_norm": 3.625, | |
| "learning_rate": 3.0783515925228626e-06, | |
| "loss": 0.4892, | |
| "step": 6220 | |
| }, | |
| { | |
| "epoch": 1.2758550071677248, | |
| "grad_norm": 4.53125, | |
| "learning_rate": 3.0630539144616646e-06, | |
| "loss": 0.4363, | |
| "step": 6230 | |
| }, | |
| { | |
| "epoch": 1.2779029285275445, | |
| "grad_norm": 3.828125, | |
| "learning_rate": 3.0477775394513743e-06, | |
| "loss": 0.4816, | |
| "step": 6240 | |
| }, | |
| { | |
| "epoch": 1.2799508498873644, | |
| "grad_norm": 4.09375, | |
| "learning_rate": 3.0325226355056425e-06, | |
| "loss": 0.5106, | |
| "step": 6250 | |
| }, | |
| { | |
| "epoch": 1.281998771247184, | |
| "grad_norm": 3.9375, | |
| "learning_rate": 3.0172893704019846e-06, | |
| "loss": 0.4663, | |
| "step": 6260 | |
| }, | |
| { | |
| "epoch": 1.2840466926070038, | |
| "grad_norm": 3.90625, | |
| "learning_rate": 3.0020779116799236e-06, | |
| "loss": 0.5006, | |
| "step": 6270 | |
| }, | |
| { | |
| "epoch": 1.2860946139668237, | |
| "grad_norm": 3.28125, | |
| "learning_rate": 2.9868884266391464e-06, | |
| "loss": 0.511, | |
| "step": 6280 | |
| }, | |
| { | |
| "epoch": 1.2881425353266436, | |
| "grad_norm": 4.65625, | |
| "learning_rate": 2.9717210823376742e-06, | |
| "loss": 0.4877, | |
| "step": 6290 | |
| }, | |
| { | |
| "epoch": 1.2901904566864633, | |
| "grad_norm": 3.90625, | |
| "learning_rate": 2.9565760455900106e-06, | |
| "loss": 0.4974, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 1.292238378046283, | |
| "grad_norm": 3.09375, | |
| "learning_rate": 2.941453482965323e-06, | |
| "loss": 0.5354, | |
| "step": 6310 | |
| }, | |
| { | |
| "epoch": 1.2942862994061028, | |
| "grad_norm": 4.125, | |
| "learning_rate": 2.926353560785594e-06, | |
| "loss": 0.5125, | |
| "step": 6320 | |
| }, | |
| { | |
| "epoch": 1.2963342207659225, | |
| "grad_norm": 3.828125, | |
| "learning_rate": 2.9112764451238074e-06, | |
| "loss": 0.5341, | |
| "step": 6330 | |
| }, | |
| { | |
| "epoch": 1.2983821421257424, | |
| "grad_norm": 3.609375, | |
| "learning_rate": 2.8962223018021116e-06, | |
| "loss": 0.4633, | |
| "step": 6340 | |
| }, | |
| { | |
| "epoch": 1.3004300634855621, | |
| "grad_norm": 3.828125, | |
| "learning_rate": 2.8811912963900013e-06, | |
| "loss": 0.4744, | |
| "step": 6350 | |
| }, | |
| { | |
| "epoch": 1.302477984845382, | |
| "grad_norm": 3.9375, | |
| "learning_rate": 2.8661835942024915e-06, | |
| "loss": 0.5308, | |
| "step": 6360 | |
| }, | |
| { | |
| "epoch": 1.3045259062052017, | |
| "grad_norm": 3.65625, | |
| "learning_rate": 2.8511993602983023e-06, | |
| "loss": 0.5155, | |
| "step": 6370 | |
| }, | |
| { | |
| "epoch": 1.3065738275650216, | |
| "grad_norm": 3.578125, | |
| "learning_rate": 2.836238759478045e-06, | |
| "loss": 0.4424, | |
| "step": 6380 | |
| }, | |
| { | |
| "epoch": 1.3086217489248413, | |
| "grad_norm": 3.984375, | |
| "learning_rate": 2.821301956282408e-06, | |
| "loss": 0.5507, | |
| "step": 6390 | |
| }, | |
| { | |
| "epoch": 1.310669670284661, | |
| "grad_norm": 3.234375, | |
| "learning_rate": 2.806389114990345e-06, | |
| "loss": 0.5209, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 1.3127175916444809, | |
| "grad_norm": 4.0625, | |
| "learning_rate": 2.7915003996172724e-06, | |
| "loss": 0.5002, | |
| "step": 6410 | |
| }, | |
| { | |
| "epoch": 1.3147655130043006, | |
| "grad_norm": 3.46875, | |
| "learning_rate": 2.776635973913262e-06, | |
| "loss": 0.4994, | |
| "step": 6420 | |
| }, | |
| { | |
| "epoch": 1.3168134343641205, | |
| "grad_norm": 3.109375, | |
| "learning_rate": 2.761796001361241e-06, | |
| "loss": 0.4868, | |
| "step": 6430 | |
| }, | |
| { | |
| "epoch": 1.3188613557239401, | |
| "grad_norm": 3.703125, | |
| "learning_rate": 2.746980645175199e-06, | |
| "loss": 0.4633, | |
| "step": 6440 | |
| }, | |
| { | |
| "epoch": 1.32090927708376, | |
| "grad_norm": 3.9375, | |
| "learning_rate": 2.732190068298378e-06, | |
| "loss": 0.5256, | |
| "step": 6450 | |
| }, | |
| { | |
| "epoch": 1.3229571984435797, | |
| "grad_norm": 3.75, | |
| "learning_rate": 2.7174244334015025e-06, | |
| "loss": 0.5183, | |
| "step": 6460 | |
| }, | |
| { | |
| "epoch": 1.3250051198033996, | |
| "grad_norm": 4.15625, | |
| "learning_rate": 2.7026839028809704e-06, | |
| "loss": 0.5399, | |
| "step": 6470 | |
| }, | |
| { | |
| "epoch": 1.3270530411632193, | |
| "grad_norm": 4.125, | |
| "learning_rate": 2.6879686388570782e-06, | |
| "loss": 0.5054, | |
| "step": 6480 | |
| }, | |
| { | |
| "epoch": 1.329100962523039, | |
| "grad_norm": 3.984375, | |
| "learning_rate": 2.6732788031722325e-06, | |
| "loss": 0.4731, | |
| "step": 6490 | |
| }, | |
| { | |
| "epoch": 1.331148883882859, | |
| "grad_norm": 4.59375, | |
| "learning_rate": 2.658614557389174e-06, | |
| "loss": 0.4896, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 1.3331968052426788, | |
| "grad_norm": 3.796875, | |
| "learning_rate": 2.6439760627892e-06, | |
| "loss": 0.5294, | |
| "step": 6510 | |
| }, | |
| { | |
| "epoch": 1.3352447266024985, | |
| "grad_norm": 3.421875, | |
| "learning_rate": 2.6293634803703853e-06, | |
| "loss": 0.5028, | |
| "step": 6520 | |
| }, | |
| { | |
| "epoch": 1.3372926479623182, | |
| "grad_norm": 3.96875, | |
| "learning_rate": 2.61477697084582e-06, | |
| "loss": 0.4852, | |
| "step": 6530 | |
| }, | |
| { | |
| "epoch": 1.339340569322138, | |
| "grad_norm": 4.34375, | |
| "learning_rate": 2.600216694641835e-06, | |
| "loss": 0.4643, | |
| "step": 6540 | |
| }, | |
| { | |
| "epoch": 1.3413884906819578, | |
| "grad_norm": 4.15625, | |
| "learning_rate": 2.5856828118962385e-06, | |
| "loss": 0.4863, | |
| "step": 6550 | |
| }, | |
| { | |
| "epoch": 1.3434364120417777, | |
| "grad_norm": 3.453125, | |
| "learning_rate": 2.571175482456565e-06, | |
| "loss": 0.5265, | |
| "step": 6560 | |
| }, | |
| { | |
| "epoch": 1.3454843334015973, | |
| "grad_norm": 3.453125, | |
| "learning_rate": 2.5566948658782947e-06, | |
| "loss": 0.512, | |
| "step": 6570 | |
| }, | |
| { | |
| "epoch": 1.347532254761417, | |
| "grad_norm": 3.65625, | |
| "learning_rate": 2.5422411214231272e-06, | |
| "loss": 0.453, | |
| "step": 6580 | |
| }, | |
| { | |
| "epoch": 1.349580176121237, | |
| "grad_norm": 4.25, | |
| "learning_rate": 2.5278144080572013e-06, | |
| "loss": 0.4624, | |
| "step": 6590 | |
| }, | |
| { | |
| "epoch": 1.3516280974810568, | |
| "grad_norm": 3.84375, | |
| "learning_rate": 2.513414884449373e-06, | |
| "loss": 0.5362, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 1.3536760188408765, | |
| "grad_norm": 3.6875, | |
| "learning_rate": 2.4990427089694517e-06, | |
| "loss": 0.466, | |
| "step": 6610 | |
| }, | |
| { | |
| "epoch": 1.3557239402006962, | |
| "grad_norm": 3.828125, | |
| "learning_rate": 2.484698039686465e-06, | |
| "loss": 0.513, | |
| "step": 6620 | |
| }, | |
| { | |
| "epoch": 1.357771861560516, | |
| "grad_norm": 4.0625, | |
| "learning_rate": 2.4703810343669204e-06, | |
| "loss": 0.4675, | |
| "step": 6630 | |
| }, | |
| { | |
| "epoch": 1.3598197829203358, | |
| "grad_norm": 4.28125, | |
| "learning_rate": 2.4560918504730712e-06, | |
| "loss": 0.461, | |
| "step": 6640 | |
| }, | |
| { | |
| "epoch": 1.3618677042801557, | |
| "grad_norm": 3.6875, | |
| "learning_rate": 2.4418306451611816e-06, | |
| "loss": 0.4595, | |
| "step": 6650 | |
| }, | |
| { | |
| "epoch": 1.3639156256399754, | |
| "grad_norm": 3.875, | |
| "learning_rate": 2.427597575279801e-06, | |
| "loss": 0.5385, | |
| "step": 6660 | |
| }, | |
| { | |
| "epoch": 1.3659635469997953, | |
| "grad_norm": 4.03125, | |
| "learning_rate": 2.413392797368034e-06, | |
| "loss": 0.4778, | |
| "step": 6670 | |
| }, | |
| { | |
| "epoch": 1.368011468359615, | |
| "grad_norm": 4.9375, | |
| "learning_rate": 2.3992164676538336e-06, | |
| "loss": 0.5586, | |
| "step": 6680 | |
| }, | |
| { | |
| "epoch": 1.3700593897194349, | |
| "grad_norm": 3.671875, | |
| "learning_rate": 2.385068742052255e-06, | |
| "loss": 0.4678, | |
| "step": 6690 | |
| }, | |
| { | |
| "epoch": 1.3721073110792545, | |
| "grad_norm": 3.921875, | |
| "learning_rate": 2.3709497761637744e-06, | |
| "loss": 0.5287, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 1.3741552324390742, | |
| "grad_norm": 4.59375, | |
| "learning_rate": 2.3568597252725466e-06, | |
| "loss": 0.5866, | |
| "step": 6710 | |
| }, | |
| { | |
| "epoch": 1.3762031537988941, | |
| "grad_norm": 3.75, | |
| "learning_rate": 2.3427987443447237e-06, | |
| "loss": 0.5044, | |
| "step": 6720 | |
| }, | |
| { | |
| "epoch": 1.378251075158714, | |
| "grad_norm": 4.15625, | |
| "learning_rate": 2.3287669880267317e-06, | |
| "loss": 0.5611, | |
| "step": 6730 | |
| }, | |
| { | |
| "epoch": 1.3802989965185337, | |
| "grad_norm": 3.65625, | |
| "learning_rate": 2.3147646106435773e-06, | |
| "loss": 0.4826, | |
| "step": 6740 | |
| }, | |
| { | |
| "epoch": 1.3823469178783534, | |
| "grad_norm": 3.5, | |
| "learning_rate": 2.300791766197151e-06, | |
| "loss": 0.4959, | |
| "step": 6750 | |
| }, | |
| { | |
| "epoch": 1.3843948392381733, | |
| "grad_norm": 3.84375, | |
| "learning_rate": 2.2868486083645325e-06, | |
| "loss": 0.4908, | |
| "step": 6760 | |
| }, | |
| { | |
| "epoch": 1.386442760597993, | |
| "grad_norm": 3.75, | |
| "learning_rate": 2.272935290496297e-06, | |
| "loss": 0.5306, | |
| "step": 6770 | |
| }, | |
| { | |
| "epoch": 1.3884906819578129, | |
| "grad_norm": 4.0625, | |
| "learning_rate": 2.259051965614839e-06, | |
| "loss": 0.5183, | |
| "step": 6780 | |
| }, | |
| { | |
| "epoch": 1.3905386033176326, | |
| "grad_norm": 4.25, | |
| "learning_rate": 2.2451987864126712e-06, | |
| "loss": 0.5608, | |
| "step": 6790 | |
| }, | |
| { | |
| "epoch": 1.3925865246774523, | |
| "grad_norm": 4.40625, | |
| "learning_rate": 2.231375905250769e-06, | |
| "loss": 0.5404, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 1.3946344460372722, | |
| "grad_norm": 4.375, | |
| "learning_rate": 2.2175834741568677e-06, | |
| "loss": 0.5083, | |
| "step": 6810 | |
| }, | |
| { | |
| "epoch": 1.396682367397092, | |
| "grad_norm": 3.265625, | |
| "learning_rate": 2.2038216448238185e-06, | |
| "loss": 0.4536, | |
| "step": 6820 | |
| }, | |
| { | |
| "epoch": 1.3987302887569117, | |
| "grad_norm": 3.328125, | |
| "learning_rate": 2.1900905686078945e-06, | |
| "loss": 0.4988, | |
| "step": 6830 | |
| }, | |
| { | |
| "epoch": 1.4007782101167314, | |
| "grad_norm": 3.421875, | |
| "learning_rate": 2.1763903965271466e-06, | |
| "loss": 0.5205, | |
| "step": 6840 | |
| }, | |
| { | |
| "epoch": 1.4028261314765513, | |
| "grad_norm": 4.3125, | |
| "learning_rate": 2.162721279259729e-06, | |
| "loss": 0.5292, | |
| "step": 6850 | |
| }, | |
| { | |
| "epoch": 1.404874052836371, | |
| "grad_norm": 3.765625, | |
| "learning_rate": 2.1490833671422484e-06, | |
| "loss": 0.484, | |
| "step": 6860 | |
| }, | |
| { | |
| "epoch": 1.406921974196191, | |
| "grad_norm": 3.8125, | |
| "learning_rate": 2.1354768101681077e-06, | |
| "loss": 0.462, | |
| "step": 6870 | |
| }, | |
| { | |
| "epoch": 1.4089698955560106, | |
| "grad_norm": 3.75, | |
| "learning_rate": 2.1219017579858586e-06, | |
| "loss": 0.5473, | |
| "step": 6880 | |
| }, | |
| { | |
| "epoch": 1.4110178169158305, | |
| "grad_norm": 4.25, | |
| "learning_rate": 2.108358359897553e-06, | |
| "loss": 0.4682, | |
| "step": 6890 | |
| }, | |
| { | |
| "epoch": 1.4130657382756502, | |
| "grad_norm": 3.765625, | |
| "learning_rate": 2.0948467648571085e-06, | |
| "loss": 0.516, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 1.41511365963547, | |
| "grad_norm": 4.375, | |
| "learning_rate": 2.0813671214686533e-06, | |
| "loss": 0.5238, | |
| "step": 6910 | |
| }, | |
| { | |
| "epoch": 1.4171615809952898, | |
| "grad_norm": 4.15625, | |
| "learning_rate": 2.067919577984916e-06, | |
| "loss": 0.4904, | |
| "step": 6920 | |
| }, | |
| { | |
| "epoch": 1.4192095023551095, | |
| "grad_norm": 4.375, | |
| "learning_rate": 2.054504282305569e-06, | |
| "loss": 0.543, | |
| "step": 6930 | |
| }, | |
| { | |
| "epoch": 1.4212574237149294, | |
| "grad_norm": 3.78125, | |
| "learning_rate": 2.041121381975624e-06, | |
| "loss": 0.4894, | |
| "step": 6940 | |
| }, | |
| { | |
| "epoch": 1.4233053450747493, | |
| "grad_norm": 3.25, | |
| "learning_rate": 2.027771024183798e-06, | |
| "loss": 0.4765, | |
| "step": 6950 | |
| }, | |
| { | |
| "epoch": 1.425353266434569, | |
| "grad_norm": 3.28125, | |
| "learning_rate": 2.0144533557608925e-06, | |
| "loss": 0.4883, | |
| "step": 6960 | |
| }, | |
| { | |
| "epoch": 1.4274011877943886, | |
| "grad_norm": 3.609375, | |
| "learning_rate": 2.0011685231781876e-06, | |
| "loss": 0.5183, | |
| "step": 6970 | |
| }, | |
| { | |
| "epoch": 1.4294491091542085, | |
| "grad_norm": 3.9375, | |
| "learning_rate": 1.987916672545822e-06, | |
| "loss": 0.5087, | |
| "step": 6980 | |
| }, | |
| { | |
| "epoch": 1.4314970305140282, | |
| "grad_norm": 3.984375, | |
| "learning_rate": 1.9746979496111936e-06, | |
| "loss": 0.4799, | |
| "step": 6990 | |
| }, | |
| { | |
| "epoch": 1.4335449518738481, | |
| "grad_norm": 4.78125, | |
| "learning_rate": 1.96151249975735e-06, | |
| "loss": 0.4663, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 1.4335449518738481, | |
| "eval_loss": 0.5521626472473145, | |
| "eval_runtime": 50.1391, | |
| "eval_samples_per_second": 82.012, | |
| "eval_steps_per_second": 41.006, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 1.4355928732336678, | |
| "grad_norm": 4.03125, | |
| "learning_rate": 1.948360468001393e-06, | |
| "loss": 0.5628, | |
| "step": 7010 | |
| }, | |
| { | |
| "epoch": 1.4376407945934875, | |
| "grad_norm": 4.375, | |
| "learning_rate": 1.935241998992889e-06, | |
| "loss": 0.4884, | |
| "step": 7020 | |
| }, | |
| { | |
| "epoch": 1.4396887159533074, | |
| "grad_norm": 4.1875, | |
| "learning_rate": 1.9221572370122606e-06, | |
| "loss": 0.5088, | |
| "step": 7030 | |
| }, | |
| { | |
| "epoch": 1.4417366373131273, | |
| "grad_norm": 3.90625, | |
| "learning_rate": 1.9091063259692255e-06, | |
| "loss": 0.4645, | |
| "step": 7040 | |
| }, | |
| { | |
| "epoch": 1.443784558672947, | |
| "grad_norm": 4.0625, | |
| "learning_rate": 1.896089409401185e-06, | |
| "loss": 0.5207, | |
| "step": 7050 | |
| }, | |
| { | |
| "epoch": 1.4458324800327667, | |
| "grad_norm": 5.15625, | |
| "learning_rate": 1.8831066304716738e-06, | |
| "loss": 0.5466, | |
| "step": 7060 | |
| }, | |
| { | |
| "epoch": 1.4478804013925866, | |
| "grad_norm": 4.125, | |
| "learning_rate": 1.8701581319687634e-06, | |
| "loss": 0.5313, | |
| "step": 7070 | |
| }, | |
| { | |
| "epoch": 1.4499283227524062, | |
| "grad_norm": 4.1875, | |
| "learning_rate": 1.8572440563035016e-06, | |
| "loss": 0.523, | |
| "step": 7080 | |
| }, | |
| { | |
| "epoch": 1.4519762441122261, | |
| "grad_norm": 3.9375, | |
| "learning_rate": 1.8443645455083465e-06, | |
| "loss": 0.4837, | |
| "step": 7090 | |
| }, | |
| { | |
| "epoch": 1.4540241654720458, | |
| "grad_norm": 3.765625, | |
| "learning_rate": 1.8315197412356006e-06, | |
| "loss": 0.4952, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 1.4560720868318657, | |
| "grad_norm": 3.515625, | |
| "learning_rate": 1.8187097847558532e-06, | |
| "loss": 0.4919, | |
| "step": 7110 | |
| }, | |
| { | |
| "epoch": 1.4581200081916854, | |
| "grad_norm": 4.09375, | |
| "learning_rate": 1.8059348169564366e-06, | |
| "loss": 0.4318, | |
| "step": 7120 | |
| }, | |
| { | |
| "epoch": 1.4601679295515053, | |
| "grad_norm": 3.828125, | |
| "learning_rate": 1.793194978339855e-06, | |
| "loss": 0.4651, | |
| "step": 7130 | |
| }, | |
| { | |
| "epoch": 1.462215850911325, | |
| "grad_norm": 3.5625, | |
| "learning_rate": 1.7804904090222664e-06, | |
| "loss": 0.4786, | |
| "step": 7140 | |
| }, | |
| { | |
| "epoch": 1.4642637722711447, | |
| "grad_norm": 3.890625, | |
| "learning_rate": 1.767821248731913e-06, | |
| "loss": 0.4668, | |
| "step": 7150 | |
| }, | |
| { | |
| "epoch": 1.4663116936309646, | |
| "grad_norm": 4.125, | |
| "learning_rate": 1.7551876368076154e-06, | |
| "loss": 0.5105, | |
| "step": 7160 | |
| }, | |
| { | |
| "epoch": 1.4683596149907845, | |
| "grad_norm": 4.40625, | |
| "learning_rate": 1.7425897121972068e-06, | |
| "loss": 0.501, | |
| "step": 7170 | |
| }, | |
| { | |
| "epoch": 1.4704075363506042, | |
| "grad_norm": 3.09375, | |
| "learning_rate": 1.7300276134560367e-06, | |
| "loss": 0.5251, | |
| "step": 7180 | |
| }, | |
| { | |
| "epoch": 1.4724554577104239, | |
| "grad_norm": 4.15625, | |
| "learning_rate": 1.717501478745423e-06, | |
| "loss": 0.5332, | |
| "step": 7190 | |
| }, | |
| { | |
| "epoch": 1.4745033790702438, | |
| "grad_norm": 4.28125, | |
| "learning_rate": 1.7050114458311446e-06, | |
| "loss": 0.4896, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 1.4765513004300634, | |
| "grad_norm": 4.5625, | |
| "learning_rate": 1.6925576520819225e-06, | |
| "loss": 0.5556, | |
| "step": 7210 | |
| }, | |
| { | |
| "epoch": 1.4785992217898833, | |
| "grad_norm": 4.625, | |
| "learning_rate": 1.6801402344679102e-06, | |
| "loss": 0.5056, | |
| "step": 7220 | |
| }, | |
| { | |
| "epoch": 1.480647143149703, | |
| "grad_norm": 3.296875, | |
| "learning_rate": 1.6677593295591848e-06, | |
| "loss": 0.5639, | |
| "step": 7230 | |
| }, | |
| { | |
| "epoch": 1.4826950645095227, | |
| "grad_norm": 4.28125, | |
| "learning_rate": 1.6554150735242535e-06, | |
| "loss": 0.5681, | |
| "step": 7240 | |
| }, | |
| { | |
| "epoch": 1.4847429858693426, | |
| "grad_norm": 4.4375, | |
| "learning_rate": 1.6431076021285381e-06, | |
| "loss": 0.4984, | |
| "step": 7250 | |
| }, | |
| { | |
| "epoch": 1.4867909072291625, | |
| "grad_norm": 3.46875, | |
| "learning_rate": 1.6308370507329057e-06, | |
| "loss": 0.453, | |
| "step": 7260 | |
| }, | |
| { | |
| "epoch": 1.4888388285889822, | |
| "grad_norm": 4.40625, | |
| "learning_rate": 1.6186035542921546e-06, | |
| "loss": 0.5102, | |
| "step": 7270 | |
| }, | |
| { | |
| "epoch": 1.4908867499488019, | |
| "grad_norm": 4.09375, | |
| "learning_rate": 1.6064072473535546e-06, | |
| "loss": 0.499, | |
| "step": 7280 | |
| }, | |
| { | |
| "epoch": 1.4929346713086218, | |
| "grad_norm": 4.65625, | |
| "learning_rate": 1.5942482640553486e-06, | |
| "loss": 0.5314, | |
| "step": 7290 | |
| }, | |
| { | |
| "epoch": 1.4949825926684415, | |
| "grad_norm": 3.84375, | |
| "learning_rate": 1.5821267381252858e-06, | |
| "loss": 0.4847, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 1.4970305140282614, | |
| "grad_norm": 3.40625, | |
| "learning_rate": 1.570042802879148e-06, | |
| "loss": 0.5141, | |
| "step": 7310 | |
| }, | |
| { | |
| "epoch": 1.499078435388081, | |
| "grad_norm": 4.46875, | |
| "learning_rate": 1.5579965912192873e-06, | |
| "loss": 0.4914, | |
| "step": 7320 | |
| }, | |
| { | |
| "epoch": 1.5011263567479007, | |
| "grad_norm": 3.453125, | |
| "learning_rate": 1.5459882356331596e-06, | |
| "loss": 0.5131, | |
| "step": 7330 | |
| }, | |
| { | |
| "epoch": 1.5031742781077206, | |
| "grad_norm": 3.21875, | |
| "learning_rate": 1.53401786819187e-06, | |
| "loss": 0.4868, | |
| "step": 7340 | |
| }, | |
| { | |
| "epoch": 1.5052221994675405, | |
| "grad_norm": 3.6875, | |
| "learning_rate": 1.5220856205487183e-06, | |
| "loss": 0.485, | |
| "step": 7350 | |
| }, | |
| { | |
| "epoch": 1.5072701208273602, | |
| "grad_norm": 3.890625, | |
| "learning_rate": 1.510191623937759e-06, | |
| "loss": 0.4816, | |
| "step": 7360 | |
| }, | |
| { | |
| "epoch": 1.50931804218718, | |
| "grad_norm": 3.546875, | |
| "learning_rate": 1.498336009172341e-06, | |
| "loss": 0.5506, | |
| "step": 7370 | |
| }, | |
| { | |
| "epoch": 1.5113659635469998, | |
| "grad_norm": 4.1875, | |
| "learning_rate": 1.4865189066436909e-06, | |
| "loss": 0.4692, | |
| "step": 7380 | |
| }, | |
| { | |
| "epoch": 1.5134138849068197, | |
| "grad_norm": 4.5, | |
| "learning_rate": 1.4747404463194553e-06, | |
| "loss": 0.5319, | |
| "step": 7390 | |
| }, | |
| { | |
| "epoch": 1.5154618062666394, | |
| "grad_norm": 3.546875, | |
| "learning_rate": 1.4630007577422949e-06, | |
| "loss": 0.4945, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 1.517509727626459, | |
| "grad_norm": 3.375, | |
| "learning_rate": 1.451299970028442e-06, | |
| "loss": 0.5442, | |
| "step": 7410 | |
| }, | |
| { | |
| "epoch": 1.5195576489862788, | |
| "grad_norm": 4.3125, | |
| "learning_rate": 1.4396382118662877e-06, | |
| "loss": 0.4956, | |
| "step": 7420 | |
| }, | |
| { | |
| "epoch": 1.5216055703460987, | |
| "grad_norm": 4.3125, | |
| "learning_rate": 1.4280156115149667e-06, | |
| "loss": 0.531, | |
| "step": 7430 | |
| }, | |
| { | |
| "epoch": 1.5236534917059186, | |
| "grad_norm": 3.671875, | |
| "learning_rate": 1.4164322968029442e-06, | |
| "loss": 0.4683, | |
| "step": 7440 | |
| }, | |
| { | |
| "epoch": 1.5257014130657383, | |
| "grad_norm": 3.5, | |
| "learning_rate": 1.4048883951266135e-06, | |
| "loss": 0.5077, | |
| "step": 7450 | |
| }, | |
| { | |
| "epoch": 1.527749334425558, | |
| "grad_norm": 3.765625, | |
| "learning_rate": 1.3933840334488903e-06, | |
| "loss": 0.4806, | |
| "step": 7460 | |
| }, | |
| { | |
| "epoch": 1.5297972557853778, | |
| "grad_norm": 4.71875, | |
| "learning_rate": 1.3819193382978207e-06, | |
| "loss": 0.452, | |
| "step": 7470 | |
| }, | |
| { | |
| "epoch": 1.5318451771451977, | |
| "grad_norm": 3.703125, | |
| "learning_rate": 1.3704944357651872e-06, | |
| "loss": 0.4668, | |
| "step": 7480 | |
| }, | |
| { | |
| "epoch": 1.5338930985050174, | |
| "grad_norm": 3.78125, | |
| "learning_rate": 1.3591094515051223e-06, | |
| "loss": 0.4809, | |
| "step": 7490 | |
| }, | |
| { | |
| "epoch": 1.535941019864837, | |
| "grad_norm": 3.6875, | |
| "learning_rate": 1.347764510732727e-06, | |
| "loss": 0.5493, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 1.537988941224657, | |
| "grad_norm": 3.375, | |
| "learning_rate": 1.3364597382226908e-06, | |
| "loss": 0.5105, | |
| "step": 7510 | |
| }, | |
| { | |
| "epoch": 1.540036862584477, | |
| "grad_norm": 3.90625, | |
| "learning_rate": 1.325195258307928e-06, | |
| "loss": 0.5163, | |
| "step": 7520 | |
| }, | |
| { | |
| "epoch": 1.5420847839442966, | |
| "grad_norm": 3.375, | |
| "learning_rate": 1.313971194878198e-06, | |
| "loss": 0.495, | |
| "step": 7530 | |
| }, | |
| { | |
| "epoch": 1.5441327053041163, | |
| "grad_norm": 3.890625, | |
| "learning_rate": 1.3027876713787524e-06, | |
| "loss": 0.4762, | |
| "step": 7540 | |
| }, | |
| { | |
| "epoch": 1.546180626663936, | |
| "grad_norm": 3.78125, | |
| "learning_rate": 1.2916448108089713e-06, | |
| "loss": 0.4488, | |
| "step": 7550 | |
| }, | |
| { | |
| "epoch": 1.5482285480237559, | |
| "grad_norm": 3.59375, | |
| "learning_rate": 1.2805427357210154e-06, | |
| "loss": 0.4543, | |
| "step": 7560 | |
| }, | |
| { | |
| "epoch": 1.5502764693835758, | |
| "grad_norm": 3.890625, | |
| "learning_rate": 1.2694815682184747e-06, | |
| "loss": 0.5251, | |
| "step": 7570 | |
| }, | |
| { | |
| "epoch": 1.5523243907433955, | |
| "grad_norm": 4.34375, | |
| "learning_rate": 1.2584614299550257e-06, | |
| "loss": 0.5251, | |
| "step": 7580 | |
| }, | |
| { | |
| "epoch": 1.5543723121032151, | |
| "grad_norm": 3.359375, | |
| "learning_rate": 1.2474824421330962e-06, | |
| "loss": 0.4684, | |
| "step": 7590 | |
| }, | |
| { | |
| "epoch": 1.556420233463035, | |
| "grad_norm": 4.15625, | |
| "learning_rate": 1.2365447255025293e-06, | |
| "loss": 0.4835, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 1.558468154822855, | |
| "grad_norm": 4.40625, | |
| "learning_rate": 1.2256484003592572e-06, | |
| "loss": 0.4605, | |
| "step": 7610 | |
| }, | |
| { | |
| "epoch": 1.5605160761826746, | |
| "grad_norm": 4.3125, | |
| "learning_rate": 1.2147935865439747e-06, | |
| "loss": 0.4669, | |
| "step": 7620 | |
| }, | |
| { | |
| "epoch": 1.5625639975424943, | |
| "grad_norm": 3.703125, | |
| "learning_rate": 1.2039804034408315e-06, | |
| "loss": 0.4901, | |
| "step": 7630 | |
| }, | |
| { | |
| "epoch": 1.564611918902314, | |
| "grad_norm": 3.484375, | |
| "learning_rate": 1.1932089699760995e-06, | |
| "loss": 0.5078, | |
| "step": 7640 | |
| }, | |
| { | |
| "epoch": 1.566659840262134, | |
| "grad_norm": 4.28125, | |
| "learning_rate": 1.1824794046168892e-06, | |
| "loss": 0.4499, | |
| "step": 7650 | |
| }, | |
| { | |
| "epoch": 1.5687077616219538, | |
| "grad_norm": 4.71875, | |
| "learning_rate": 1.1717918253698273e-06, | |
| "loss": 0.5427, | |
| "step": 7660 | |
| }, | |
| { | |
| "epoch": 1.5707556829817735, | |
| "grad_norm": 3.296875, | |
| "learning_rate": 1.1611463497797676e-06, | |
| "loss": 0.4713, | |
| "step": 7670 | |
| }, | |
| { | |
| "epoch": 1.5728036043415932, | |
| "grad_norm": 4.15625, | |
| "learning_rate": 1.1505430949284962e-06, | |
| "loss": 0.511, | |
| "step": 7680 | |
| }, | |
| { | |
| "epoch": 1.574851525701413, | |
| "grad_norm": 4.03125, | |
| "learning_rate": 1.1399821774334457e-06, | |
| "loss": 0.5032, | |
| "step": 7690 | |
| }, | |
| { | |
| "epoch": 1.576899447061233, | |
| "grad_norm": 3.890625, | |
| "learning_rate": 1.1294637134464104e-06, | |
| "loss": 0.4683, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 1.5789473684210527, | |
| "grad_norm": 4.21875, | |
| "learning_rate": 1.1189878186522684e-06, | |
| "loss": 0.4678, | |
| "step": 7710 | |
| }, | |
| { | |
| "epoch": 1.5809952897808723, | |
| "grad_norm": 4.40625, | |
| "learning_rate": 1.1085546082677123e-06, | |
| "loss": 0.4456, | |
| "step": 7720 | |
| }, | |
| { | |
| "epoch": 1.5830432111406922, | |
| "grad_norm": 4.0625, | |
| "learning_rate": 1.0981641970399786e-06, | |
| "loss": 0.4764, | |
| "step": 7730 | |
| }, | |
| { | |
| "epoch": 1.5850911325005121, | |
| "grad_norm": 4.03125, | |
| "learning_rate": 1.0878166992455874e-06, | |
| "loss": 0.4918, | |
| "step": 7740 | |
| }, | |
| { | |
| "epoch": 1.5871390538603318, | |
| "grad_norm": 3.65625, | |
| "learning_rate": 1.0775122286890894e-06, | |
| "loss": 0.4753, | |
| "step": 7750 | |
| }, | |
| { | |
| "epoch": 1.5891869752201515, | |
| "grad_norm": 3.96875, | |
| "learning_rate": 1.0672508987018016e-06, | |
| "loss": 0.4841, | |
| "step": 7760 | |
| }, | |
| { | |
| "epoch": 1.5912348965799712, | |
| "grad_norm": 4.21875, | |
| "learning_rate": 1.0570328221405796e-06, | |
| "loss": 0.5167, | |
| "step": 7770 | |
| }, | |
| { | |
| "epoch": 1.593282817939791, | |
| "grad_norm": 3.578125, | |
| "learning_rate": 1.046858111386556e-06, | |
| "loss": 0.5181, | |
| "step": 7780 | |
| }, | |
| { | |
| "epoch": 1.595330739299611, | |
| "grad_norm": 3.828125, | |
| "learning_rate": 1.0367268783439249e-06, | |
| "loss": 0.4854, | |
| "step": 7790 | |
| }, | |
| { | |
| "epoch": 1.5973786606594307, | |
| "grad_norm": 3.46875, | |
| "learning_rate": 1.0266392344386939e-06, | |
| "loss": 0.4486, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 1.5994265820192504, | |
| "grad_norm": 4.5, | |
| "learning_rate": 1.0165952906174675e-06, | |
| "loss": 0.4954, | |
| "step": 7810 | |
| }, | |
| { | |
| "epoch": 1.6014745033790703, | |
| "grad_norm": 4.03125, | |
| "learning_rate": 1.006595157346225e-06, | |
| "loss": 0.4548, | |
| "step": 7820 | |
| }, | |
| { | |
| "epoch": 1.6035224247388902, | |
| "grad_norm": 3.875, | |
| "learning_rate": 9.966389446091068e-07, | |
| "loss": 0.5134, | |
| "step": 7830 | |
| }, | |
| { | |
| "epoch": 1.6055703460987099, | |
| "grad_norm": 3.703125, | |
| "learning_rate": 9.867267619072013e-07, | |
| "loss": 0.4977, | |
| "step": 7840 | |
| }, | |
| { | |
| "epoch": 1.6076182674585295, | |
| "grad_norm": 4.09375, | |
| "learning_rate": 9.768587182573442e-07, | |
| "loss": 0.4621, | |
| "step": 7850 | |
| }, | |
| { | |
| "epoch": 1.6096661888183492, | |
| "grad_norm": 3.859375, | |
| "learning_rate": 9.67034922190917e-07, | |
| "loss": 0.5124, | |
| "step": 7860 | |
| }, | |
| { | |
| "epoch": 1.6117141101781691, | |
| "grad_norm": 4.65625, | |
| "learning_rate": 9.572554817526592e-07, | |
| "loss": 0.5, | |
| "step": 7870 | |
| }, | |
| { | |
| "epoch": 1.613762031537989, | |
| "grad_norm": 4.15625, | |
| "learning_rate": 9.475205044994651e-07, | |
| "loss": 0.44, | |
| "step": 7880 | |
| }, | |
| { | |
| "epoch": 1.6158099528978087, | |
| "grad_norm": 3.453125, | |
| "learning_rate": 9.378300974992238e-07, | |
| "loss": 0.5015, | |
| "step": 7890 | |
| }, | |
| { | |
| "epoch": 1.6178578742576284, | |
| "grad_norm": 3.8125, | |
| "learning_rate": 9.281843673296165e-07, | |
| "loss": 0.486, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 1.6199057956174483, | |
| "grad_norm": 3.5625, | |
| "learning_rate": 9.185834200769662e-07, | |
| "loss": 0.4988, | |
| "step": 7910 | |
| }, | |
| { | |
| "epoch": 1.6219537169772682, | |
| "grad_norm": 3.765625, | |
| "learning_rate": 9.090273613350564e-07, | |
| "loss": 0.5192, | |
| "step": 7920 | |
| }, | |
| { | |
| "epoch": 1.6240016383370879, | |
| "grad_norm": 4.0, | |
| "learning_rate": 8.995162962039761e-07, | |
| "loss": 0.512, | |
| "step": 7930 | |
| }, | |
| { | |
| "epoch": 1.6260495596969076, | |
| "grad_norm": 4.1875, | |
| "learning_rate": 8.900503292889628e-07, | |
| "loss": 0.5336, | |
| "step": 7940 | |
| }, | |
| { | |
| "epoch": 1.6280974810567275, | |
| "grad_norm": 3.828125, | |
| "learning_rate": 8.806295646992508e-07, | |
| "loss": 0.4872, | |
| "step": 7950 | |
| }, | |
| { | |
| "epoch": 1.6301454024165472, | |
| "grad_norm": 3.28125, | |
| "learning_rate": 8.712541060469271e-07, | |
| "loss": 0.5295, | |
| "step": 7960 | |
| }, | |
| { | |
| "epoch": 1.632193323776367, | |
| "grad_norm": 4.03125, | |
| "learning_rate": 8.61924056445796e-07, | |
| "loss": 0.4899, | |
| "step": 7970 | |
| }, | |
| { | |
| "epoch": 1.6342412451361867, | |
| "grad_norm": 3.5625, | |
| "learning_rate": 8.526395185102321e-07, | |
| "loss": 0.4615, | |
| "step": 7980 | |
| }, | |
| { | |
| "epoch": 1.6362891664960064, | |
| "grad_norm": 3.734375, | |
| "learning_rate": 8.434005943540718e-07, | |
| "loss": 0.4868, | |
| "step": 7990 | |
| }, | |
| { | |
| "epoch": 1.6383370878558263, | |
| "grad_norm": 4.3125, | |
| "learning_rate": 8.342073855894673e-07, | |
| "loss": 0.5378, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 1.6383370878558263, | |
| "eval_loss": 0.5521250367164612, | |
| "eval_runtime": 50.5238, | |
| "eval_samples_per_second": 81.387, | |
| "eval_steps_per_second": 40.694, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 1.6403850092156462, | |
| "grad_norm": 4.03125, | |
| "learning_rate": 8.250599933257919e-07, | |
| "loss": 0.4721, | |
| "step": 8010 | |
| }, | |
| { | |
| "epoch": 1.642432930575466, | |
| "grad_norm": 3.71875, | |
| "learning_rate": 8.15958518168506e-07, | |
| "loss": 0.4715, | |
| "step": 8020 | |
| }, | |
| { | |
| "epoch": 1.6444808519352856, | |
| "grad_norm": 3.859375, | |
| "learning_rate": 8.069030602180705e-07, | |
| "loss": 0.5022, | |
| "step": 8030 | |
| }, | |
| { | |
| "epoch": 1.6465287732951055, | |
| "grad_norm": 4.34375, | |
| "learning_rate": 7.978937190688318e-07, | |
| "loss": 0.5166, | |
| "step": 8040 | |
| }, | |
| { | |
| "epoch": 1.6485766946549254, | |
| "grad_norm": 3.796875, | |
| "learning_rate": 7.889305938079328e-07, | |
| "loss": 0.5324, | |
| "step": 8050 | |
| }, | |
| { | |
| "epoch": 1.650624616014745, | |
| "grad_norm": 3.078125, | |
| "learning_rate": 7.800137830142212e-07, | |
| "loss": 0.4688, | |
| "step": 8060 | |
| }, | |
| { | |
| "epoch": 1.6526725373745648, | |
| "grad_norm": 4.03125, | |
| "learning_rate": 7.711433847571664e-07, | |
| "loss": 0.5302, | |
| "step": 8070 | |
| }, | |
| { | |
| "epoch": 1.6547204587343844, | |
| "grad_norm": 3.8125, | |
| "learning_rate": 7.623194965957786e-07, | |
| "loss": 0.4856, | |
| "step": 8080 | |
| }, | |
| { | |
| "epoch": 1.6567683800942044, | |
| "grad_norm": 3.984375, | |
| "learning_rate": 7.535422155775423e-07, | |
| "loss": 0.5046, | |
| "step": 8090 | |
| }, | |
| { | |
| "epoch": 1.6588163014540243, | |
| "grad_norm": 3.921875, | |
| "learning_rate": 7.44811638237336e-07, | |
| "loss": 0.5231, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 1.660864222813844, | |
| "grad_norm": 3.421875, | |
| "learning_rate": 7.361278605963884e-07, | |
| "loss": 0.5012, | |
| "step": 8110 | |
| }, | |
| { | |
| "epoch": 1.6629121441736636, | |
| "grad_norm": 3.703125, | |
| "learning_rate": 7.274909781612033e-07, | |
| "loss": 0.4458, | |
| "step": 8120 | |
| }, | |
| { | |
| "epoch": 1.6649600655334835, | |
| "grad_norm": 4.15625, | |
| "learning_rate": 7.18901085922526e-07, | |
| "loss": 0.504, | |
| "step": 8130 | |
| }, | |
| { | |
| "epoch": 1.6670079868933034, | |
| "grad_norm": 3.53125, | |
| "learning_rate": 7.103582783542867e-07, | |
| "loss": 0.4965, | |
| "step": 8140 | |
| }, | |
| { | |
| "epoch": 1.669055908253123, | |
| "grad_norm": 3.96875, | |
| "learning_rate": 7.018626494125674e-07, | |
| "loss": 0.5263, | |
| "step": 8150 | |
| }, | |
| { | |
| "epoch": 1.6711038296129428, | |
| "grad_norm": 4.53125, | |
| "learning_rate": 6.934142925345661e-07, | |
| "loss": 0.5062, | |
| "step": 8160 | |
| }, | |
| { | |
| "epoch": 1.6731517509727627, | |
| "grad_norm": 3.078125, | |
| "learning_rate": 6.850133006375704e-07, | |
| "loss": 0.4966, | |
| "step": 8170 | |
| }, | |
| { | |
| "epoch": 1.6751996723325824, | |
| "grad_norm": 4.21875, | |
| "learning_rate": 6.766597661179352e-07, | |
| "loss": 0.4877, | |
| "step": 8180 | |
| }, | |
| { | |
| "epoch": 1.6772475936924023, | |
| "grad_norm": 4.1875, | |
| "learning_rate": 6.683537808500673e-07, | |
| "loss": 0.5252, | |
| "step": 8190 | |
| }, | |
| { | |
| "epoch": 1.679295515052222, | |
| "grad_norm": 3.984375, | |
| "learning_rate": 6.600954361854117e-07, | |
| "loss": 0.5023, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 1.6813434364120416, | |
| "grad_norm": 4.65625, | |
| "learning_rate": 6.518848229514541e-07, | |
| "loss": 0.5164, | |
| "step": 8210 | |
| }, | |
| { | |
| "epoch": 1.6833913577718616, | |
| "grad_norm": 3.921875, | |
| "learning_rate": 6.437220314507098e-07, | |
| "loss": 0.4991, | |
| "step": 8220 | |
| }, | |
| { | |
| "epoch": 1.6854392791316815, | |
| "grad_norm": 3.828125, | |
| "learning_rate": 6.356071514597467e-07, | |
| "loss": 0.4688, | |
| "step": 8230 | |
| }, | |
| { | |
| "epoch": 1.6874872004915011, | |
| "grad_norm": 4.3125, | |
| "learning_rate": 6.275402722281798e-07, | |
| "loss": 0.5088, | |
| "step": 8240 | |
| }, | |
| { | |
| "epoch": 1.6895351218513208, | |
| "grad_norm": 3.703125, | |
| "learning_rate": 6.195214824777068e-07, | |
| "loss": 0.496, | |
| "step": 8250 | |
| }, | |
| { | |
| "epoch": 1.6915830432111407, | |
| "grad_norm": 3.875, | |
| "learning_rate": 6.115508704011208e-07, | |
| "loss": 0.5199, | |
| "step": 8260 | |
| }, | |
| { | |
| "epoch": 1.6936309645709606, | |
| "grad_norm": 3.296875, | |
| "learning_rate": 6.036285236613437e-07, | |
| "loss": 0.4849, | |
| "step": 8270 | |
| }, | |
| { | |
| "epoch": 1.6956788859307803, | |
| "grad_norm": 3.453125, | |
| "learning_rate": 5.957545293904632e-07, | |
| "loss": 0.489, | |
| "step": 8280 | |
| }, | |
| { | |
| "epoch": 1.6977268072906, | |
| "grad_norm": 4.25, | |
| "learning_rate": 5.879289741887739e-07, | |
| "loss": 0.4736, | |
| "step": 8290 | |
| }, | |
| { | |
| "epoch": 1.6997747286504197, | |
| "grad_norm": 4.65625, | |
| "learning_rate": 5.801519441238229e-07, | |
| "loss": 0.4806, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 1.7018226500102396, | |
| "grad_norm": 3.359375, | |
| "learning_rate": 5.72423524729469e-07, | |
| "loss": 0.4315, | |
| "step": 8310 | |
| }, | |
| { | |
| "epoch": 1.7038705713700595, | |
| "grad_norm": 4.03125, | |
| "learning_rate": 5.647438010049305e-07, | |
| "loss": 0.5183, | |
| "step": 8320 | |
| }, | |
| { | |
| "epoch": 1.7059184927298792, | |
| "grad_norm": 3.328125, | |
| "learning_rate": 5.571128574138667e-07, | |
| "loss": 0.4632, | |
| "step": 8330 | |
| }, | |
| { | |
| "epoch": 1.7079664140896988, | |
| "grad_norm": 4.0, | |
| "learning_rate": 5.495307778834319e-07, | |
| "loss": 0.5044, | |
| "step": 8340 | |
| }, | |
| { | |
| "epoch": 1.7100143354495188, | |
| "grad_norm": 3.375, | |
| "learning_rate": 5.419976458033666e-07, | |
| "loss": 0.5054, | |
| "step": 8350 | |
| }, | |
| { | |
| "epoch": 1.7120622568093387, | |
| "grad_norm": 3.90625, | |
| "learning_rate": 5.345135440250687e-07, | |
| "loss": 0.5582, | |
| "step": 8360 | |
| }, | |
| { | |
| "epoch": 1.7141101781691583, | |
| "grad_norm": 3.9375, | |
| "learning_rate": 5.270785548606927e-07, | |
| "loss": 0.5177, | |
| "step": 8370 | |
| }, | |
| { | |
| "epoch": 1.716158099528978, | |
| "grad_norm": 3.921875, | |
| "learning_rate": 5.196927600822366e-07, | |
| "loss": 0.5106, | |
| "step": 8380 | |
| }, | |
| { | |
| "epoch": 1.718206020888798, | |
| "grad_norm": 3.40625, | |
| "learning_rate": 5.123562409206457e-07, | |
| "loss": 0.4848, | |
| "step": 8390 | |
| }, | |
| { | |
| "epoch": 1.7202539422486176, | |
| "grad_norm": 3.796875, | |
| "learning_rate": 5.050690780649193e-07, | |
| "loss": 0.5181, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 1.7223018636084375, | |
| "grad_norm": 4.3125, | |
| "learning_rate": 4.978313516612232e-07, | |
| "loss": 0.4963, | |
| "step": 8410 | |
| }, | |
| { | |
| "epoch": 1.7243497849682572, | |
| "grad_norm": 3.1875, | |
| "learning_rate": 4.906431413120056e-07, | |
| "loss": 0.5136, | |
| "step": 8420 | |
| }, | |
| { | |
| "epoch": 1.7263977063280769, | |
| "grad_norm": 4.0, | |
| "learning_rate": 4.835045260751297e-07, | |
| "loss": 0.5049, | |
| "step": 8430 | |
| }, | |
| { | |
| "epoch": 1.7284456276878968, | |
| "grad_norm": 3.546875, | |
| "learning_rate": 4.7641558446298974e-07, | |
| "loss": 0.5492, | |
| "step": 8440 | |
| }, | |
| { | |
| "epoch": 1.7304935490477167, | |
| "grad_norm": 3.921875, | |
| "learning_rate": 4.6937639444166514e-07, | |
| "loss": 0.5012, | |
| "step": 8450 | |
| }, | |
| { | |
| "epoch": 1.7325414704075364, | |
| "grad_norm": 3.359375, | |
| "learning_rate": 4.6238703343004565e-07, | |
| "loss": 0.4954, | |
| "step": 8460 | |
| }, | |
| { | |
| "epoch": 1.734589391767356, | |
| "grad_norm": 4.1875, | |
| "learning_rate": 4.5544757829899535e-07, | |
| "loss": 0.4714, | |
| "step": 8470 | |
| }, | |
| { | |
| "epoch": 1.736637313127176, | |
| "grad_norm": 3.890625, | |
| "learning_rate": 4.485581053704957e-07, | |
| "loss": 0.485, | |
| "step": 8480 | |
| }, | |
| { | |
| "epoch": 1.7386852344869959, | |
| "grad_norm": 3.390625, | |
| "learning_rate": 4.417186904168125e-07, | |
| "loss": 0.4994, | |
| "step": 8490 | |
| }, | |
| { | |
| "epoch": 1.7407331558468155, | |
| "grad_norm": 3.453125, | |
| "learning_rate": 4.3492940865966006e-07, | |
| "loss": 0.4771, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 1.7427810772066352, | |
| "grad_norm": 3.578125, | |
| "learning_rate": 4.2819033476937386e-07, | |
| "loss": 0.4589, | |
| "step": 8510 | |
| }, | |
| { | |
| "epoch": 1.744828998566455, | |
| "grad_norm": 4.15625, | |
| "learning_rate": 4.2150154286409006e-07, | |
| "loss": 0.5249, | |
| "step": 8520 | |
| }, | |
| { | |
| "epoch": 1.7468769199262748, | |
| "grad_norm": 4.0, | |
| "learning_rate": 4.148631065089315e-07, | |
| "loss": 0.5176, | |
| "step": 8530 | |
| }, | |
| { | |
| "epoch": 1.7489248412860947, | |
| "grad_norm": 3.84375, | |
| "learning_rate": 4.0827509871519455e-07, | |
| "loss": 0.5037, | |
| "step": 8540 | |
| }, | |
| { | |
| "epoch": 1.7509727626459144, | |
| "grad_norm": 3.953125, | |
| "learning_rate": 4.0173759193955107e-07, | |
| "loss": 0.5298, | |
| "step": 8550 | |
| }, | |
| { | |
| "epoch": 1.753020684005734, | |
| "grad_norm": 3.703125, | |
| "learning_rate": 3.952506580832477e-07, | |
| "loss": 0.5212, | |
| "step": 8560 | |
| }, | |
| { | |
| "epoch": 1.755068605365554, | |
| "grad_norm": 4.15625, | |
| "learning_rate": 3.8881436849131925e-07, | |
| "loss": 0.5017, | |
| "step": 8570 | |
| }, | |
| { | |
| "epoch": 1.7571165267253739, | |
| "grad_norm": 3.265625, | |
| "learning_rate": 3.8242879395179677e-07, | |
| "loss": 0.5061, | |
| "step": 8580 | |
| }, | |
| { | |
| "epoch": 1.7591644480851936, | |
| "grad_norm": 3.921875, | |
| "learning_rate": 3.760940046949385e-07, | |
| "loss": 0.4965, | |
| "step": 8590 | |
| }, | |
| { | |
| "epoch": 1.7612123694450132, | |
| "grad_norm": 3.96875, | |
| "learning_rate": 3.69810070392449e-07, | |
| "loss": 0.5191, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 1.763260290804833, | |
| "grad_norm": 4.1875, | |
| "learning_rate": 3.63577060156719e-07, | |
| "loss": 0.5086, | |
| "step": 8610 | |
| }, | |
| { | |
| "epoch": 1.7653082121646528, | |
| "grad_norm": 4.40625, | |
| "learning_rate": 3.573950425400602e-07, | |
| "loss": 0.5109, | |
| "step": 8620 | |
| }, | |
| { | |
| "epoch": 1.7673561335244727, | |
| "grad_norm": 4.34375, | |
| "learning_rate": 3.512640855339561e-07, | |
| "loss": 0.4675, | |
| "step": 8630 | |
| }, | |
| { | |
| "epoch": 1.7694040548842924, | |
| "grad_norm": 3.828125, | |
| "learning_rate": 3.4518425656831e-07, | |
| "loss": 0.5042, | |
| "step": 8640 | |
| }, | |
| { | |
| "epoch": 1.771451976244112, | |
| "grad_norm": 4.09375, | |
| "learning_rate": 3.391556225107073e-07, | |
| "loss": 0.5347, | |
| "step": 8650 | |
| }, | |
| { | |
| "epoch": 1.773499897603932, | |
| "grad_norm": 4.15625, | |
| "learning_rate": 3.3317824966567713e-07, | |
| "loss": 0.4946, | |
| "step": 8660 | |
| }, | |
| { | |
| "epoch": 1.775547818963752, | |
| "grad_norm": 3.96875, | |
| "learning_rate": 3.272522037739634e-07, | |
| "loss": 0.4695, | |
| "step": 8670 | |
| }, | |
| { | |
| "epoch": 1.7775957403235716, | |
| "grad_norm": 4.625, | |
| "learning_rate": 3.213775500118038e-07, | |
| "loss": 0.5263, | |
| "step": 8680 | |
| }, | |
| { | |
| "epoch": 1.7796436616833913, | |
| "grad_norm": 4.1875, | |
| "learning_rate": 3.155543529902111e-07, | |
| "loss": 0.4979, | |
| "step": 8690 | |
| }, | |
| { | |
| "epoch": 1.7816915830432112, | |
| "grad_norm": 3.390625, | |
| "learning_rate": 3.0978267675426164e-07, | |
| "loss": 0.4523, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 1.783739504403031, | |
| "grad_norm": 4.0625, | |
| "learning_rate": 3.040625847823958e-07, | |
| "loss": 0.5158, | |
| "step": 8710 | |
| }, | |
| { | |
| "epoch": 1.7857874257628508, | |
| "grad_norm": 4.0625, | |
| "learning_rate": 2.983941399857138e-07, | |
| "loss": 0.475, | |
| "step": 8720 | |
| }, | |
| { | |
| "epoch": 1.7878353471226704, | |
| "grad_norm": 3.515625, | |
| "learning_rate": 2.9277740470728623e-07, | |
| "loss": 0.5373, | |
| "step": 8730 | |
| }, | |
| { | |
| "epoch": 1.7898832684824901, | |
| "grad_norm": 3.90625, | |
| "learning_rate": 2.8721244072146995e-07, | |
| "loss": 0.4803, | |
| "step": 8740 | |
| }, | |
| { | |
| "epoch": 1.79193118984231, | |
| "grad_norm": 3.90625, | |
| "learning_rate": 2.816993092332265e-07, | |
| "loss": 0.5106, | |
| "step": 8750 | |
| }, | |
| { | |
| "epoch": 1.79397911120213, | |
| "grad_norm": 3.828125, | |
| "learning_rate": 2.7623807087745035e-07, | |
| "loss": 0.5295, | |
| "step": 8760 | |
| }, | |
| { | |
| "epoch": 1.7960270325619496, | |
| "grad_norm": 3.609375, | |
| "learning_rate": 2.708287857183006e-07, | |
| "loss": 0.5699, | |
| "step": 8770 | |
| }, | |
| { | |
| "epoch": 1.7980749539217693, | |
| "grad_norm": 4.0, | |
| "learning_rate": 2.654715132485414e-07, | |
| "loss": 0.4657, | |
| "step": 8780 | |
| }, | |
| { | |
| "epoch": 1.8001228752815892, | |
| "grad_norm": 3.796875, | |
| "learning_rate": 2.6016631238888865e-07, | |
| "loss": 0.4737, | |
| "step": 8790 | |
| }, | |
| { | |
| "epoch": 1.8021707966414091, | |
| "grad_norm": 3.796875, | |
| "learning_rate": 2.549132414873584e-07, | |
| "loss": 0.5056, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 1.8042187180012288, | |
| "grad_norm": 3.015625, | |
| "learning_rate": 2.497123583186295e-07, | |
| "loss": 0.5065, | |
| "step": 8810 | |
| }, | |
| { | |
| "epoch": 1.8062666393610485, | |
| "grad_norm": 3.6875, | |
| "learning_rate": 2.4456372008340724e-07, | |
| "loss": 0.4631, | |
| "step": 8820 | |
| }, | |
| { | |
| "epoch": 1.8083145607208682, | |
| "grad_norm": 3.859375, | |
| "learning_rate": 2.394673834077882e-07, | |
| "loss": 0.5392, | |
| "step": 8830 | |
| }, | |
| { | |
| "epoch": 1.810362482080688, | |
| "grad_norm": 3.78125, | |
| "learning_rate": 2.3442340434264798e-07, | |
| "loss": 0.4679, | |
| "step": 8840 | |
| }, | |
| { | |
| "epoch": 1.812410403440508, | |
| "grad_norm": 4.5625, | |
| "learning_rate": 2.2943183836301676e-07, | |
| "loss": 0.518, | |
| "step": 8850 | |
| }, | |
| { | |
| "epoch": 1.8144583248003276, | |
| "grad_norm": 4.0625, | |
| "learning_rate": 2.2449274036747072e-07, | |
| "loss": 0.5188, | |
| "step": 8860 | |
| }, | |
| { | |
| "epoch": 1.8165062461601473, | |
| "grad_norm": 4.0, | |
| "learning_rate": 2.1960616467753104e-07, | |
| "loss": 0.4933, | |
| "step": 8870 | |
| }, | |
| { | |
| "epoch": 1.8185541675199672, | |
| "grad_norm": 3.84375, | |
| "learning_rate": 2.1477216503706267e-07, | |
| "loss": 0.5042, | |
| "step": 8880 | |
| }, | |
| { | |
| "epoch": 1.8206020888797871, | |
| "grad_norm": 3.8125, | |
| "learning_rate": 2.0999079461168692e-07, | |
| "loss": 0.4993, | |
| "step": 8890 | |
| }, | |
| { | |
| "epoch": 1.8226500102396068, | |
| "grad_norm": 4.15625, | |
| "learning_rate": 2.0526210598819373e-07, | |
| "loss": 0.5079, | |
| "step": 8900 | |
| }, | |
| { | |
| "epoch": 1.8246979315994265, | |
| "grad_norm": 3.578125, | |
| "learning_rate": 2.0058615117396486e-07, | |
| "loss": 0.49, | |
| "step": 8910 | |
| }, | |
| { | |
| "epoch": 1.8267458529592464, | |
| "grad_norm": 4.65625, | |
| "learning_rate": 1.9596298159640149e-07, | |
| "loss": 0.5116, | |
| "step": 8920 | |
| }, | |
| { | |
| "epoch": 1.8287937743190663, | |
| "grad_norm": 3.796875, | |
| "learning_rate": 1.9139264810235757e-07, | |
| "loss": 0.5092, | |
| "step": 8930 | |
| }, | |
| { | |
| "epoch": 1.830841695678886, | |
| "grad_norm": 3.859375, | |
| "learning_rate": 1.8687520095758583e-07, | |
| "loss": 0.5138, | |
| "step": 8940 | |
| }, | |
| { | |
| "epoch": 1.8328896170387057, | |
| "grad_norm": 3.828125, | |
| "learning_rate": 1.8241068984617415e-07, | |
| "loss": 0.5163, | |
| "step": 8950 | |
| }, | |
| { | |
| "epoch": 1.8349375383985254, | |
| "grad_norm": 3.921875, | |
| "learning_rate": 1.7799916387001183e-07, | |
| "loss": 0.4741, | |
| "step": 8960 | |
| }, | |
| { | |
| "epoch": 1.8369854597583453, | |
| "grad_norm": 3.484375, | |
| "learning_rate": 1.736406715482397e-07, | |
| "loss": 0.4613, | |
| "step": 8970 | |
| }, | |
| { | |
| "epoch": 1.8390333811181652, | |
| "grad_norm": 4.09375, | |
| "learning_rate": 1.6933526081672202e-07, | |
| "loss": 0.5445, | |
| "step": 8980 | |
| }, | |
| { | |
| "epoch": 1.8410813024779848, | |
| "grad_norm": 4.28125, | |
| "learning_rate": 1.650829790275177e-07, | |
| "loss": 0.4681, | |
| "step": 8990 | |
| }, | |
| { | |
| "epoch": 1.8431292238378045, | |
| "grad_norm": 4.65625, | |
| "learning_rate": 1.608838729483575e-07, | |
| "loss": 0.5008, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 1.8431292238378045, | |
| "eval_loss": 0.5520676374435425, | |
| "eval_runtime": 50.2055, | |
| "eval_samples_per_second": 81.903, | |
| "eval_steps_per_second": 40.952, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 1.8451771451976244, | |
| "grad_norm": 4.375, | |
| "learning_rate": 1.5673798876213398e-07, | |
| "loss": 0.4875, | |
| "step": 9010 | |
| }, | |
| { | |
| "epoch": 1.8472250665574443, | |
| "grad_norm": 4.25, | |
| "learning_rate": 1.5264537206638896e-07, | |
| "loss": 0.4904, | |
| "step": 9020 | |
| }, | |
| { | |
| "epoch": 1.849272987917264, | |
| "grad_norm": 3.234375, | |
| "learning_rate": 1.4860606787281518e-07, | |
| "loss": 0.5164, | |
| "step": 9030 | |
| }, | |
| { | |
| "epoch": 1.8513209092770837, | |
| "grad_norm": 4.34375, | |
| "learning_rate": 1.4462012060676045e-07, | |
| "loss": 0.4678, | |
| "step": 9040 | |
| }, | |
| { | |
| "epoch": 1.8533688306369034, | |
| "grad_norm": 3.796875, | |
| "learning_rate": 1.4068757410673762e-07, | |
| "loss": 0.4999, | |
| "step": 9050 | |
| }, | |
| { | |
| "epoch": 1.8554167519967233, | |
| "grad_norm": 3.375, | |
| "learning_rate": 1.368084716239465e-07, | |
| "loss": 0.4937, | |
| "step": 9060 | |
| }, | |
| { | |
| "epoch": 1.8574646733565432, | |
| "grad_norm": 4.0625, | |
| "learning_rate": 1.3298285582179104e-07, | |
| "loss": 0.5074, | |
| "step": 9070 | |
| }, | |
| { | |
| "epoch": 1.8595125947163629, | |
| "grad_norm": 3.921875, | |
| "learning_rate": 1.2921076877541794e-07, | |
| "loss": 0.4735, | |
| "step": 9080 | |
| }, | |
| { | |
| "epoch": 1.8615605160761826, | |
| "grad_norm": 4.75, | |
| "learning_rate": 1.2549225197124814e-07, | |
| "loss": 0.4848, | |
| "step": 9090 | |
| }, | |
| { | |
| "epoch": 1.8636084374360025, | |
| "grad_norm": 3.71875, | |
| "learning_rate": 1.2182734630652448e-07, | |
| "loss": 0.4821, | |
| "step": 9100 | |
| }, | |
| { | |
| "epoch": 1.8656563587958224, | |
| "grad_norm": 3.484375, | |
| "learning_rate": 1.1821609208885865e-07, | |
| "loss": 0.4889, | |
| "step": 9110 | |
| }, | |
| { | |
| "epoch": 1.867704280155642, | |
| "grad_norm": 4.4375, | |
| "learning_rate": 1.14658529035791e-07, | |
| "loss": 0.4905, | |
| "step": 9120 | |
| }, | |
| { | |
| "epoch": 1.8697522015154617, | |
| "grad_norm": 3.53125, | |
| "learning_rate": 1.1115469627434983e-07, | |
| "loss": 0.4923, | |
| "step": 9130 | |
| }, | |
| { | |
| "epoch": 1.8718001228752816, | |
| "grad_norm": 3.1875, | |
| "learning_rate": 1.0770463234062556e-07, | |
| "loss": 0.4689, | |
| "step": 9140 | |
| }, | |
| { | |
| "epoch": 1.8738480442351013, | |
| "grad_norm": 4.03125, | |
| "learning_rate": 1.0430837517934278e-07, | |
| "loss": 0.4526, | |
| "step": 9150 | |
| }, | |
| { | |
| "epoch": 1.8758959655949212, | |
| "grad_norm": 3.8125, | |
| "learning_rate": 1.0096596214344723e-07, | |
| "loss": 0.5039, | |
| "step": 9160 | |
| }, | |
| { | |
| "epoch": 1.877943886954741, | |
| "grad_norm": 3.796875, | |
| "learning_rate": 9.767742999368945e-08, | |
| "loss": 0.5123, | |
| "step": 9170 | |
| }, | |
| { | |
| "epoch": 1.8799918083145606, | |
| "grad_norm": 3.9375, | |
| "learning_rate": 9.444281489822737e-08, | |
| "loss": 0.5069, | |
| "step": 9180 | |
| }, | |
| { | |
| "epoch": 1.8820397296743805, | |
| "grad_norm": 3.375, | |
| "learning_rate": 9.126215243221992e-08, | |
| "loss": 0.5134, | |
| "step": 9190 | |
| }, | |
| { | |
| "epoch": 1.8840876510342004, | |
| "grad_norm": 3.96875, | |
| "learning_rate": 8.813547757744568e-08, | |
| "loss": 0.5314, | |
| "step": 9200 | |
| }, | |
| { | |
| "epoch": 1.88613557239402, | |
| "grad_norm": 3.71875, | |
| "learning_rate": 8.506282472190819e-08, | |
| "loss": 0.5024, | |
| "step": 9210 | |
| }, | |
| { | |
| "epoch": 1.8881834937538398, | |
| "grad_norm": 3.90625, | |
| "learning_rate": 8.204422765946741e-08, | |
| "loss": 0.4582, | |
| "step": 9220 | |
| }, | |
| { | |
| "epoch": 1.8902314151136597, | |
| "grad_norm": 3.578125, | |
| "learning_rate": 7.907971958945992e-08, | |
| "loss": 0.5259, | |
| "step": 9230 | |
| }, | |
| { | |
| "epoch": 1.8922793364734796, | |
| "grad_norm": 4.5625, | |
| "learning_rate": 7.616933311633823e-08, | |
| "loss": 0.4899, | |
| "step": 9240 | |
| }, | |
| { | |
| "epoch": 1.8943272578332992, | |
| "grad_norm": 4.40625, | |
| "learning_rate": 7.331310024931038e-08, | |
| "loss": 0.4689, | |
| "step": 9250 | |
| }, | |
| { | |
| "epoch": 1.896375179193119, | |
| "grad_norm": 4.625, | |
| "learning_rate": 7.051105240198975e-08, | |
| "loss": 0.5188, | |
| "step": 9260 | |
| }, | |
| { | |
| "epoch": 1.8984231005529386, | |
| "grad_norm": 3.796875, | |
| "learning_rate": 6.776322039204642e-08, | |
| "loss": 0.4636, | |
| "step": 9270 | |
| }, | |
| { | |
| "epoch": 1.9004710219127585, | |
| "grad_norm": 3.984375, | |
| "learning_rate": 6.506963444087355e-08, | |
| "loss": 0.4749, | |
| "step": 9280 | |
| }, | |
| { | |
| "epoch": 1.9025189432725784, | |
| "grad_norm": 4.34375, | |
| "learning_rate": 6.243032417324768e-08, | |
| "loss": 0.549, | |
| "step": 9290 | |
| }, | |
| { | |
| "epoch": 1.904566864632398, | |
| "grad_norm": 3.953125, | |
| "learning_rate": 5.984531861701004e-08, | |
| "loss": 0.4987, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 1.9066147859922178, | |
| "grad_norm": 4.0, | |
| "learning_rate": 5.7314646202742405e-08, | |
| "loss": 0.5011, | |
| "step": 9310 | |
| }, | |
| { | |
| "epoch": 1.9086627073520377, | |
| "grad_norm": 3.671875, | |
| "learning_rate": 5.483833476345624e-08, | |
| "loss": 0.5187, | |
| "step": 9320 | |
| }, | |
| { | |
| "epoch": 1.9107106287118576, | |
| "grad_norm": 4.0625, | |
| "learning_rate": 5.241641153428734e-08, | |
| "loss": 0.5182, | |
| "step": 9330 | |
| }, | |
| { | |
| "epoch": 1.9127585500716773, | |
| "grad_norm": 4.09375, | |
| "learning_rate": 5.004890315219446e-08, | |
| "loss": 0.5202, | |
| "step": 9340 | |
| }, | |
| { | |
| "epoch": 1.914806471431497, | |
| "grad_norm": 3.53125, | |
| "learning_rate": 4.7735835655667864e-08, | |
| "loss": 0.5006, | |
| "step": 9350 | |
| }, | |
| { | |
| "epoch": 1.9168543927913169, | |
| "grad_norm": 4.5625, | |
| "learning_rate": 4.547723448444286e-08, | |
| "loss": 0.4896, | |
| "step": 9360 | |
| }, | |
| { | |
| "epoch": 1.9189023141511365, | |
| "grad_norm": 4.5, | |
| "learning_rate": 4.3273124479218386e-08, | |
| "loss": 0.5277, | |
| "step": 9370 | |
| }, | |
| { | |
| "epoch": 1.9209502355109564, | |
| "grad_norm": 4.4375, | |
| "learning_rate": 4.112352988138557e-08, | |
| "loss": 0.4983, | |
| "step": 9380 | |
| }, | |
| { | |
| "epoch": 1.9229981568707761, | |
| "grad_norm": 4.40625, | |
| "learning_rate": 3.902847433276014e-08, | |
| "loss": 0.5211, | |
| "step": 9390 | |
| }, | |
| { | |
| "epoch": 1.9250460782305958, | |
| "grad_norm": 3.6875, | |
| "learning_rate": 3.698798087532485e-08, | |
| "loss": 0.4465, | |
| "step": 9400 | |
| }, | |
| { | |
| "epoch": 1.9270939995904157, | |
| "grad_norm": 3.09375, | |
| "learning_rate": 3.500207195096972e-08, | |
| "loss": 0.4819, | |
| "step": 9410 | |
| }, | |
| { | |
| "epoch": 1.9291419209502356, | |
| "grad_norm": 3.875, | |
| "learning_rate": 3.3070769401254424e-08, | |
| "loss": 0.5247, | |
| "step": 9420 | |
| }, | |
| { | |
| "epoch": 1.9311898423100553, | |
| "grad_norm": 4.125, | |
| "learning_rate": 3.1194094467159044e-08, | |
| "loss": 0.5271, | |
| "step": 9430 | |
| }, | |
| { | |
| "epoch": 1.933237763669875, | |
| "grad_norm": 3.546875, | |
| "learning_rate": 2.9372067788857594e-08, | |
| "loss": 0.4789, | |
| "step": 9440 | |
| }, | |
| { | |
| "epoch": 1.935285685029695, | |
| "grad_norm": 4.4375, | |
| "learning_rate": 2.760470940548543e-08, | |
| "loss": 0.4851, | |
| "step": 9450 | |
| }, | |
| { | |
| "epoch": 1.9373336063895148, | |
| "grad_norm": 4.8125, | |
| "learning_rate": 2.58920387549233e-08, | |
| "loss": 0.4704, | |
| "step": 9460 | |
| }, | |
| { | |
| "epoch": 1.9393815277493345, | |
| "grad_norm": 3.328125, | |
| "learning_rate": 2.4234074673580853e-08, | |
| "loss": 0.5505, | |
| "step": 9470 | |
| }, | |
| { | |
| "epoch": 1.9414294491091542, | |
| "grad_norm": 4.125, | |
| "learning_rate": 2.2630835396190155e-08, | |
| "loss": 0.5198, | |
| "step": 9480 | |
| }, | |
| { | |
| "epoch": 1.9434773704689738, | |
| "grad_norm": 3.203125, | |
| "learning_rate": 2.1082338555605265e-08, | |
| "loss": 0.4826, | |
| "step": 9490 | |
| }, | |
| { | |
| "epoch": 1.9455252918287937, | |
| "grad_norm": 3.578125, | |
| "learning_rate": 1.9588601182608524e-08, | |
| "loss": 0.5109, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 1.9475732131886137, | |
| "grad_norm": 3.6875, | |
| "learning_rate": 1.814963970572292e-08, | |
| "loss": 0.5035, | |
| "step": 9510 | |
| }, | |
| { | |
| "epoch": 1.9496211345484333, | |
| "grad_norm": 3.859375, | |
| "learning_rate": 1.676546995103223e-08, | |
| "loss": 0.513, | |
| "step": 9520 | |
| }, | |
| { | |
| "epoch": 1.951669055908253, | |
| "grad_norm": 3.734375, | |
| "learning_rate": 1.5436107142003943e-08, | |
| "loss": 0.5197, | |
| "step": 9530 | |
| }, | |
| { | |
| "epoch": 1.953716977268073, | |
| "grad_norm": 3.203125, | |
| "learning_rate": 1.4161565899327157e-08, | |
| "loss": 0.4731, | |
| "step": 9540 | |
| }, | |
| { | |
| "epoch": 1.9557648986278928, | |
| "grad_norm": 3.953125, | |
| "learning_rate": 1.2941860240746617e-08, | |
| "loss": 0.4463, | |
| "step": 9550 | |
| }, | |
| { | |
| "epoch": 1.9578128199877125, | |
| "grad_norm": 3.640625, | |
| "learning_rate": 1.1777003580911161e-08, | |
| "loss": 0.5008, | |
| "step": 9560 | |
| }, | |
| { | |
| "epoch": 1.9598607413475322, | |
| "grad_norm": 3.71875, | |
| "learning_rate": 1.0667008731225504e-08, | |
| "loss": 0.4966, | |
| "step": 9570 | |
| }, | |
| { | |
| "epoch": 1.961908662707352, | |
| "grad_norm": 3.953125, | |
| "learning_rate": 9.611887899710349e-09, | |
| "loss": 0.5175, | |
| "step": 9580 | |
| }, | |
| { | |
| "epoch": 1.9639565840671718, | |
| "grad_norm": 3.71875, | |
| "learning_rate": 8.61165269086639e-09, | |
| "loss": 0.479, | |
| "step": 9590 | |
| }, | |
| { | |
| "epoch": 1.9660045054269917, | |
| "grad_norm": 3.765625, | |
| "learning_rate": 7.666314105547745e-09, | |
| "loss": 0.5428, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 1.9680524267868114, | |
| "grad_norm": 4.3125, | |
| "learning_rate": 6.775882540841494e-09, | |
| "loss": 0.4741, | |
| "step": 9610 | |
| }, | |
| { | |
| "epoch": 1.970100348146631, | |
| "grad_norm": 4.03125, | |
| "learning_rate": 5.940367789951107e-09, | |
| "loss": 0.4811, | |
| "step": 9620 | |
| }, | |
| { | |
| "epoch": 1.972148269506451, | |
| "grad_norm": 3.609375, | |
| "learning_rate": 5.159779042092083e-09, | |
| "loss": 0.54, | |
| "step": 9630 | |
| }, | |
| { | |
| "epoch": 1.9741961908662709, | |
| "grad_norm": 4.0, | |
| "learning_rate": 4.434124882388146e-09, | |
| "loss": 0.4908, | |
| "step": 9640 | |
| }, | |
| { | |
| "epoch": 1.9762441122260905, | |
| "grad_norm": 3.546875, | |
| "learning_rate": 3.763413291776874e-09, | |
| "loss": 0.5134, | |
| "step": 9650 | |
| }, | |
| { | |
| "epoch": 1.9782920335859102, | |
| "grad_norm": 4.09375, | |
| "learning_rate": 3.1476516469247655e-09, | |
| "loss": 0.4972, | |
| "step": 9660 | |
| }, | |
| { | |
| "epoch": 1.9803399549457301, | |
| "grad_norm": 3.828125, | |
| "learning_rate": 2.586846720141756e-09, | |
| "loss": 0.5337, | |
| "step": 9670 | |
| }, | |
| { | |
| "epoch": 1.98238787630555, | |
| "grad_norm": 3.625, | |
| "learning_rate": 2.0810046793118266e-09, | |
| "loss": 0.5048, | |
| "step": 9680 | |
| }, | |
| { | |
| "epoch": 1.9844357976653697, | |
| "grad_norm": 3.515625, | |
| "learning_rate": 1.6301310878197307e-09, | |
| "loss": 0.472, | |
| "step": 9690 | |
| }, | |
| { | |
| "epoch": 1.9864837190251894, | |
| "grad_norm": 3.921875, | |
| "learning_rate": 1.2342309044943712e-09, | |
| "loss": 0.4877, | |
| "step": 9700 | |
| }, | |
| { | |
| "epoch": 1.988531640385009, | |
| "grad_norm": 3.46875, | |
| "learning_rate": 8.933084835521799e-10, | |
| "loss": 0.4863, | |
| "step": 9710 | |
| }, | |
| { | |
| "epoch": 1.990579561744829, | |
| "grad_norm": 4.03125, | |
| "learning_rate": 6.073675745482677e-10, | |
| "loss": 0.4992, | |
| "step": 9720 | |
| }, | |
| { | |
| "epoch": 1.9926274831046489, | |
| "grad_norm": 3.53125, | |
| "learning_rate": 3.764113223375665e-10, | |
| "loss": 0.5157, | |
| "step": 9730 | |
| }, | |
| { | |
| "epoch": 1.9946754044644686, | |
| "grad_norm": 4.3125, | |
| "learning_rate": 2.0044226703819225e-10, | |
| "loss": 0.4896, | |
| "step": 9740 | |
| }, | |
| { | |
| "epoch": 1.9967233258242882, | |
| "grad_norm": 3.234375, | |
| "learning_rate": 7.946234400590947e-11, | |
| "loss": 0.4628, | |
| "step": 9750 | |
| }, | |
| { | |
| "epoch": 1.9987712471841081, | |
| "grad_norm": 4.125, | |
| "learning_rate": 1.3472883809706461e-11, | |
| "loss": 0.4633, | |
| "step": 9760 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 9766, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 2, | |
| "save_steps": 1000, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 6.580184799992873e+17, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |