| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 3.0, | |
| "eval_steps": 500, | |
| "global_step": 1875, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.008, | |
| "grad_norm": 0.8485889434814453, | |
| "learning_rate": 0.0001999964908278481, | |
| "loss": 1.2049, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.016, | |
| "grad_norm": 0.47789862751960754, | |
| "learning_rate": 0.00019998596355767805, | |
| "loss": 0.9333, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.024, | |
| "grad_norm": 1.017558217048645, | |
| "learning_rate": 0.00019996841892833, | |
| "loss": 0.8671, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.032, | |
| "grad_norm": 0.6610977053642273, | |
| "learning_rate": 0.00019994385817114646, | |
| "loss": 0.7979, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 0.6075429320335388, | |
| "learning_rate": 0.00019991228300988585, | |
| "loss": 0.7662, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.048, | |
| "grad_norm": 0.6595763564109802, | |
| "learning_rate": 0.00019987369566060176, | |
| "loss": 0.7929, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.056, | |
| "grad_norm": 0.6968618035316467, | |
| "learning_rate": 0.00019982809883148722, | |
| "loss": 0.7683, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.064, | |
| "grad_norm": 0.4889592230319977, | |
| "learning_rate": 0.00019977549572268468, | |
| "loss": 0.8667, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.072, | |
| "grad_norm": 0.6651108264923096, | |
| "learning_rate": 0.0001997158900260614, | |
| "loss": 0.8446, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 0.5898510217666626, | |
| "learning_rate": 0.00019964928592495045, | |
| "loss": 0.9051, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.088, | |
| "grad_norm": 0.4398016035556793, | |
| "learning_rate": 0.00019957568809385694, | |
| "loss": 0.7235, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.096, | |
| "grad_norm": 0.6901968121528625, | |
| "learning_rate": 0.00019949510169813003, | |
| "loss": 0.8169, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.104, | |
| "grad_norm": 0.6267213225364685, | |
| "learning_rate": 0.00019940753239360047, | |
| "loss": 0.8266, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.112, | |
| "grad_norm": 0.48524895310401917, | |
| "learning_rate": 0.00019931298632618356, | |
| "loss": 0.758, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 0.5294132232666016, | |
| "learning_rate": 0.0001992114701314478, | |
| "loss": 0.7759, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.128, | |
| "grad_norm": 0.48957982659339905, | |
| "learning_rate": 0.0001991029909341493, | |
| "loss": 0.7797, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.136, | |
| "grad_norm": 0.645412802696228, | |
| "learning_rate": 0.00019898755634773158, | |
| "loss": 0.7437, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.144, | |
| "grad_norm": 0.43297675251960754, | |
| "learning_rate": 0.0001988651744737914, | |
| "loss": 0.8043, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.152, | |
| "grad_norm": 0.5513920783996582, | |
| "learning_rate": 0.00019873585390151003, | |
| "loss": 0.7701, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 0.8462435007095337, | |
| "learning_rate": 0.0001985996037070505, | |
| "loss": 0.709, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.168, | |
| "grad_norm": 0.6892585158348083, | |
| "learning_rate": 0.00019845643345292054, | |
| "loss": 0.7377, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.176, | |
| "grad_norm": 0.4617864191532135, | |
| "learning_rate": 0.00019830635318730154, | |
| "loss": 0.8352, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.184, | |
| "grad_norm": 0.6300354599952698, | |
| "learning_rate": 0.0001981493734433433, | |
| "loss": 0.7738, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.192, | |
| "grad_norm": 0.8086859583854675, | |
| "learning_rate": 0.0001979855052384247, | |
| "loss": 0.8067, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 0.6272985935211182, | |
| "learning_rate": 0.00019781476007338058, | |
| "loss": 0.7456, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.208, | |
| "grad_norm": 0.44750839471817017, | |
| "learning_rate": 0.00019763714993169452, | |
| "loss": 0.758, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.216, | |
| "grad_norm": 0.5053977370262146, | |
| "learning_rate": 0.00019745268727865774, | |
| "loss": 0.7895, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.224, | |
| "grad_norm": 0.41920769214630127, | |
| "learning_rate": 0.00019726138506049438, | |
| "loss": 0.7302, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.232, | |
| "grad_norm": 0.38280290365219116, | |
| "learning_rate": 0.00019706325670345275, | |
| "loss": 0.8152, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 0.554710865020752, | |
| "learning_rate": 0.0001968583161128631, | |
| "loss": 0.8461, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.248, | |
| "grad_norm": 0.5612509250640869, | |
| "learning_rate": 0.00019664657767216176, | |
| "loss": 0.7787, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.256, | |
| "grad_norm": 0.610614538192749, | |
| "learning_rate": 0.00019642805624188147, | |
| "loss": 0.7574, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.264, | |
| "grad_norm": 0.679517924785614, | |
| "learning_rate": 0.0001962027671586086, | |
| "loss": 0.8487, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.272, | |
| "grad_norm": 0.6685434579849243, | |
| "learning_rate": 0.00019597072623390668, | |
| "loss": 0.6611, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 0.480293869972229, | |
| "learning_rate": 0.00019573194975320673, | |
| "loss": 0.7802, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.288, | |
| "grad_norm": 0.7727369070053101, | |
| "learning_rate": 0.00019548645447466431, | |
| "loss": 0.6727, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.296, | |
| "grad_norm": 0.6371043920516968, | |
| "learning_rate": 0.00019523425762798329, | |
| "loss": 0.7502, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.304, | |
| "grad_norm": 0.6399966478347778, | |
| "learning_rate": 0.00019497537691320668, | |
| "loss": 0.8401, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.312, | |
| "grad_norm": 0.7263137698173523, | |
| "learning_rate": 0.00019470983049947444, | |
| "loss": 0.7494, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 0.402416467666626, | |
| "learning_rate": 0.00019443763702374812, | |
| "loss": 0.7842, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.328, | |
| "grad_norm": 0.6639626026153564, | |
| "learning_rate": 0.00019415881558950302, | |
| "loss": 0.8082, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.336, | |
| "grad_norm": 0.5801042914390564, | |
| "learning_rate": 0.00019387338576538744, | |
| "loss": 0.7883, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.344, | |
| "grad_norm": 0.5533607006072998, | |
| "learning_rate": 0.00019358136758384912, | |
| "loss": 0.7356, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.352, | |
| "grad_norm": 0.6019654273986816, | |
| "learning_rate": 0.00019328278153972947, | |
| "loss": 0.7891, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 0.5344104170799255, | |
| "learning_rate": 0.00019297764858882514, | |
| "loss": 0.7671, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.368, | |
| "grad_norm": 0.5494843125343323, | |
| "learning_rate": 0.0001926659901464172, | |
| "loss": 0.6608, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.376, | |
| "grad_norm": 0.465420126914978, | |
| "learning_rate": 0.00019234782808576824, | |
| "loss": 0.647, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.384, | |
| "grad_norm": 0.5202775001525879, | |
| "learning_rate": 0.00019202318473658705, | |
| "loss": 0.729, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.392, | |
| "grad_norm": 0.5757818222045898, | |
| "learning_rate": 0.00019169208288346166, | |
| "loss": 0.6713, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 0.46555572748184204, | |
| "learning_rate": 0.0001913545457642601, | |
| "loss": 0.7049, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.408, | |
| "grad_norm": 0.5101790428161621, | |
| "learning_rate": 0.00019101059706849957, | |
| "loss": 0.7419, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.416, | |
| "grad_norm": 0.6083744764328003, | |
| "learning_rate": 0.00019066026093568378, | |
| "loss": 0.7148, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.424, | |
| "grad_norm": 0.4719640612602234, | |
| "learning_rate": 0.00019030356195360874, | |
| "loss": 0.7493, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.432, | |
| "grad_norm": 0.7365225553512573, | |
| "learning_rate": 0.0001899405251566371, | |
| "loss": 0.7652, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 0.4452705383300781, | |
| "learning_rate": 0.0001895711760239413, | |
| "loss": 0.7438, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.448, | |
| "grad_norm": 0.6071786284446716, | |
| "learning_rate": 0.0001891955404777151, | |
| "loss": 0.7683, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.456, | |
| "grad_norm": 0.5774498581886292, | |
| "learning_rate": 0.00018881364488135448, | |
| "loss": 0.8115, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.464, | |
| "grad_norm": 0.6134682893753052, | |
| "learning_rate": 0.00018842551603760724, | |
| "loss": 0.8335, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.472, | |
| "grad_norm": 0.4869893193244934, | |
| "learning_rate": 0.00018803118118669202, | |
| "loss": 0.6933, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 0.6457111239433289, | |
| "learning_rate": 0.00018763066800438636, | |
| "loss": 0.7515, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.488, | |
| "grad_norm": 0.59674471616745, | |
| "learning_rate": 0.0001872240046000844, | |
| "loss": 0.6931, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.496, | |
| "grad_norm": 0.44608160853385925, | |
| "learning_rate": 0.00018681121951482393, | |
| "loss": 0.782, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.504, | |
| "grad_norm": 0.5934664607048035, | |
| "learning_rate": 0.00018639234171928353, | |
| "loss": 0.7361, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.512, | |
| "grad_norm": 0.49716323614120483, | |
| "learning_rate": 0.0001859674006117491, | |
| "loss": 0.7443, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 0.47995495796203613, | |
| "learning_rate": 0.00018553642601605068, | |
| "loss": 0.7221, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.528, | |
| "grad_norm": 0.5177399516105652, | |
| "learning_rate": 0.00018509944817946922, | |
| "loss": 0.7622, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.536, | |
| "grad_norm": 0.6638798713684082, | |
| "learning_rate": 0.0001846564977706138, | |
| "loss": 0.8556, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.544, | |
| "grad_norm": 0.5056771636009216, | |
| "learning_rate": 0.00018420760587726923, | |
| "loss": 0.7814, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.552, | |
| "grad_norm": 0.44543707370758057, | |
| "learning_rate": 0.0001837528040042142, | |
| "loss": 0.722, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 0.6765120625495911, | |
| "learning_rate": 0.00018329212407100994, | |
| "loss": 0.7903, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.568, | |
| "grad_norm": 0.49232372641563416, | |
| "learning_rate": 0.00018282559840976042, | |
| "loss": 0.6996, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.576, | |
| "grad_norm": 0.47392791509628296, | |
| "learning_rate": 0.00018235325976284275, | |
| "loss": 0.773, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.584, | |
| "grad_norm": 0.5056615471839905, | |
| "learning_rate": 0.00018187514128060946, | |
| "loss": 0.728, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.592, | |
| "grad_norm": 0.5857616662979126, | |
| "learning_rate": 0.00018139127651906184, | |
| "loss": 0.7659, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 0.5966864228248596, | |
| "learning_rate": 0.00018090169943749476, | |
| "loss": 0.7039, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.608, | |
| "grad_norm": 0.4524347484111786, | |
| "learning_rate": 0.00018040644439611348, | |
| "loss": 0.7125, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.616, | |
| "grad_norm": 0.5570976138114929, | |
| "learning_rate": 0.00017990554615362198, | |
| "loss": 0.698, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.624, | |
| "grad_norm": 0.6045777201652527, | |
| "learning_rate": 0.00017939903986478355, | |
| "loss": 0.8255, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.632, | |
| "grad_norm": 0.6149687767028809, | |
| "learning_rate": 0.00017888696107795342, | |
| "loss": 0.6616, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 0.4873579144477844, | |
| "learning_rate": 0.000178369345732584, | |
| "loss": 0.7452, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.648, | |
| "grad_norm": 0.5569061636924744, | |
| "learning_rate": 0.00017784623015670238, | |
| "loss": 0.7652, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.656, | |
| "grad_norm": 0.5825181603431702, | |
| "learning_rate": 0.00017731765106436073, | |
| "loss": 0.7793, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.664, | |
| "grad_norm": 0.4047383666038513, | |
| "learning_rate": 0.00017678364555305978, | |
| "loss": 0.6875, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.672, | |
| "grad_norm": 0.5080836415290833, | |
| "learning_rate": 0.0001762442511011448, | |
| "loss": 0.7465, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 0.5825940370559692, | |
| "learning_rate": 0.00017569950556517566, | |
| "loss": 0.7205, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.688, | |
| "grad_norm": 0.476992666721344, | |
| "learning_rate": 0.00017514944717726962, | |
| "loss": 0.6589, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.696, | |
| "grad_norm": 0.7424727082252502, | |
| "learning_rate": 0.00017459411454241822, | |
| "loss": 0.7035, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.704, | |
| "grad_norm": 0.6544787287712097, | |
| "learning_rate": 0.00017403354663577783, | |
| "loss": 0.787, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.712, | |
| "grad_norm": 0.49425187706947327, | |
| "learning_rate": 0.00017346778279993415, | |
| "loss": 0.7515, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 0.5473236441612244, | |
| "learning_rate": 0.00017289686274214118, | |
| "loss": 0.7199, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.728, | |
| "grad_norm": 0.6773544549942017, | |
| "learning_rate": 0.00017232082653153422, | |
| "loss": 0.8037, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.736, | |
| "grad_norm": 0.6355096101760864, | |
| "learning_rate": 0.00017173971459631787, | |
| "loss": 0.7502, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.744, | |
| "grad_norm": 0.47867000102996826, | |
| "learning_rate": 0.00017115356772092857, | |
| "loss": 0.7446, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.752, | |
| "grad_norm": 0.5135357975959778, | |
| "learning_rate": 0.0001705624270431721, | |
| "loss": 0.6507, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 0.48866042494773865, | |
| "learning_rate": 0.00016996633405133655, | |
| "loss": 0.7164, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.768, | |
| "grad_norm": 0.5892354249954224, | |
| "learning_rate": 0.0001693653305812805, | |
| "loss": 0.7621, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.776, | |
| "grad_norm": 0.6633970141410828, | |
| "learning_rate": 0.00016875945881349676, | |
| "loss": 0.7623, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.784, | |
| "grad_norm": 0.6444060802459717, | |
| "learning_rate": 0.000168148761270152, | |
| "loss": 0.6606, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.792, | |
| "grad_norm": 0.7012648582458496, | |
| "learning_rate": 0.00016753328081210245, | |
| "loss": 0.6941, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 0.7064160704612732, | |
| "learning_rate": 0.00016691306063588583, | |
| "loss": 0.6841, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.808, | |
| "grad_norm": 0.7241398096084595, | |
| "learning_rate": 0.00016628814427068953, | |
| "loss": 0.6996, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 0.816, | |
| "grad_norm": 0.7807374596595764, | |
| "learning_rate": 0.00016565857557529566, | |
| "loss": 0.7542, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.824, | |
| "grad_norm": 0.763768196105957, | |
| "learning_rate": 0.00016502439873500289, | |
| "loss": 0.7175, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 0.832, | |
| "grad_norm": 0.6105090379714966, | |
| "learning_rate": 0.0001643856582585254, | |
| "loss": 0.7565, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 0.5686540603637695, | |
| "learning_rate": 0.000163742398974869, | |
| "loss": 0.7339, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.848, | |
| "grad_norm": 0.5341500043869019, | |
| "learning_rate": 0.00016309466603018496, | |
| "loss": 0.569, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.856, | |
| "grad_norm": 0.7274748682975769, | |
| "learning_rate": 0.00016244250488460158, | |
| "loss": 0.7556, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 0.864, | |
| "grad_norm": 0.7321165204048157, | |
| "learning_rate": 0.00016178596130903344, | |
| "loss": 0.7084, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.872, | |
| "grad_norm": 0.5086159110069275, | |
| "learning_rate": 0.00016112508138196917, | |
| "loss": 0.6935, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 0.4714389443397522, | |
| "learning_rate": 0.0001604599114862375, | |
| "loss": 0.7076, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.888, | |
| "grad_norm": 0.5031452178955078, | |
| "learning_rate": 0.0001597904983057519, | |
| "loss": 0.7151, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 0.896, | |
| "grad_norm": 0.7745943665504456, | |
| "learning_rate": 0.0001591168888222342, | |
| "loss": 0.7001, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.904, | |
| "grad_norm": 0.6076303124427795, | |
| "learning_rate": 0.00015843913031191723, | |
| "loss": 0.7285, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 0.912, | |
| "grad_norm": 0.7456529140472412, | |
| "learning_rate": 0.00015775727034222675, | |
| "loss": 0.8041, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 0.5760998725891113, | |
| "learning_rate": 0.0001570713567684432, | |
| "loss": 0.7353, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 0.928, | |
| "grad_norm": 0.7057327032089233, | |
| "learning_rate": 0.00015638143773034267, | |
| "loss": 0.7792, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.936, | |
| "grad_norm": 0.7615967392921448, | |
| "learning_rate": 0.00015568756164881882, | |
| "loss": 1.0121, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 0.944, | |
| "grad_norm": 0.6304950714111328, | |
| "learning_rate": 0.000154989777222484, | |
| "loss": 0.7727, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.952, | |
| "grad_norm": 0.6852543950080872, | |
| "learning_rate": 0.00015428813342425177, | |
| "loss": 0.741, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 0.6379660964012146, | |
| "learning_rate": 0.00015358267949789966, | |
| "loss": 0.6919, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.968, | |
| "grad_norm": 0.5846463441848755, | |
| "learning_rate": 0.00015287346495461315, | |
| "loss": 0.7163, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 0.976, | |
| "grad_norm": 0.5999557971954346, | |
| "learning_rate": 0.0001521605395695108, | |
| "loss": 0.8152, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.984, | |
| "grad_norm": 0.5806307196617126, | |
| "learning_rate": 0.00015144395337815064, | |
| "loss": 0.6709, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 0.992, | |
| "grad_norm": 0.6559942960739136, | |
| "learning_rate": 0.00015072375667301893, | |
| "loss": 0.6527, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 0.6287715435028076, | |
| "learning_rate": 0.00015000000000000001, | |
| "loss": 0.8194, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 1.008, | |
| "grad_norm": 0.616222620010376, | |
| "learning_rate": 0.00014927273415482915, | |
| "loss": 0.6627, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 1.016, | |
| "grad_norm": 0.4750412106513977, | |
| "learning_rate": 0.0001485420101795274, | |
| "loss": 0.6366, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 1.024, | |
| "grad_norm": 0.5122964978218079, | |
| "learning_rate": 0.00014780787935881923, | |
| "loss": 0.6717, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 1.032, | |
| "grad_norm": 0.7382633090019226, | |
| "learning_rate": 0.0001470703932165333, | |
| "loss": 0.6483, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "grad_norm": 0.6540554761886597, | |
| "learning_rate": 0.00014632960351198618, | |
| "loss": 0.6151, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 1.048, | |
| "grad_norm": 0.4776591956615448, | |
| "learning_rate": 0.00014558556223635003, | |
| "loss": 0.6707, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 1.056, | |
| "grad_norm": 0.8012662529945374, | |
| "learning_rate": 0.00014483832160900326, | |
| "loss": 0.6125, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 1.064, | |
| "grad_norm": 0.6735953092575073, | |
| "learning_rate": 0.00014408793407386588, | |
| "loss": 0.6206, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 1.072, | |
| "grad_norm": 0.5640230774879456, | |
| "learning_rate": 0.00014333445229571873, | |
| "loss": 0.6161, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "grad_norm": 0.5928654074668884, | |
| "learning_rate": 0.00014257792915650728, | |
| "loss": 0.6583, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 1.088, | |
| "grad_norm": 0.7347397208213806, | |
| "learning_rate": 0.00014181841775163013, | |
| "loss": 0.6222, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 1.096, | |
| "grad_norm": 0.593773365020752, | |
| "learning_rate": 0.0001410559713862128, | |
| "loss": 0.716, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 1.104, | |
| "grad_norm": 0.6244611144065857, | |
| "learning_rate": 0.00014029064357136628, | |
| "loss": 0.6198, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 1.112, | |
| "grad_norm": 0.5083370804786682, | |
| "learning_rate": 0.00013952248802043165, | |
| "loss": 0.6389, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "grad_norm": 0.5241413116455078, | |
| "learning_rate": 0.0001387515586452103, | |
| "loss": 0.6842, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 1.1280000000000001, | |
| "grad_norm": 0.524029016494751, | |
| "learning_rate": 0.00013797790955218014, | |
| "loss": 0.6071, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 1.1360000000000001, | |
| "grad_norm": 0.5097878575325012, | |
| "learning_rate": 0.00013720159503869815, | |
| "loss": 0.5915, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 1.144, | |
| "grad_norm": 0.5782963037490845, | |
| "learning_rate": 0.00013642266958918984, | |
| "loss": 0.6794, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 1.152, | |
| "grad_norm": 0.6088266372680664, | |
| "learning_rate": 0.00013564118787132506, | |
| "loss": 0.6773, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "grad_norm": 0.7768995761871338, | |
| "learning_rate": 0.00013485720473218154, | |
| "loss": 0.668, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 1.168, | |
| "grad_norm": 0.6645551919937134, | |
| "learning_rate": 0.0001340707751943952, | |
| "loss": 0.6997, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 1.176, | |
| "grad_norm": 0.9228842258453369, | |
| "learning_rate": 0.00013328195445229868, | |
| "loss": 0.831, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 1.184, | |
| "grad_norm": 0.7556049823760986, | |
| "learning_rate": 0.00013249079786804765, | |
| "loss": 0.6378, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 1.192, | |
| "grad_norm": 0.832775354385376, | |
| "learning_rate": 0.0001316973609677352, | |
| "loss": 0.6547, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "grad_norm": 0.7329304814338684, | |
| "learning_rate": 0.00013090169943749476, | |
| "loss": 0.5808, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 1.208, | |
| "grad_norm": 0.7193475961685181, | |
| "learning_rate": 0.00013010386911959206, | |
| "loss": 0.5582, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 1.216, | |
| "grad_norm": 0.6274734735488892, | |
| "learning_rate": 0.00012930392600850573, | |
| "loss": 0.5801, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 1.224, | |
| "grad_norm": 0.6485865712165833, | |
| "learning_rate": 0.0001285019262469976, | |
| "loss": 0.65, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 1.232, | |
| "grad_norm": 0.7164427042007446, | |
| "learning_rate": 0.00012769792612217224, | |
| "loss": 0.6627, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "grad_norm": 0.600775957107544, | |
| "learning_rate": 0.00012689198206152657, | |
| "loss": 0.5603, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 1.248, | |
| "grad_norm": 0.8377975225448608, | |
| "learning_rate": 0.00012608415062898972, | |
| "loss": 0.6525, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 1.256, | |
| "grad_norm": 0.8069924116134644, | |
| "learning_rate": 0.00012527448852095295, | |
| "loss": 0.6731, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 1.264, | |
| "grad_norm": 0.6501213908195496, | |
| "learning_rate": 0.00012446305256229073, | |
| "loss": 0.6255, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 1.272, | |
| "grad_norm": 0.62812340259552, | |
| "learning_rate": 0.00012364989970237248, | |
| "loss": 0.6585, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "grad_norm": 0.5702307820320129, | |
| "learning_rate": 0.00012283508701106557, | |
| "loss": 0.5996, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 1.288, | |
| "grad_norm": 0.6311281323432922, | |
| "learning_rate": 0.00012201867167473015, | |
| "loss": 0.6355, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 1.296, | |
| "grad_norm": 0.5885419249534607, | |
| "learning_rate": 0.00012120071099220549, | |
| "loss": 0.6615, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 1.304, | |
| "grad_norm": 0.5239307284355164, | |
| "learning_rate": 0.0001203812623707885, | |
| "loss": 0.6096, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 1.312, | |
| "grad_norm": 0.6101869940757751, | |
| "learning_rate": 0.00011956038332220483, | |
| "loss": 0.5984, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "grad_norm": 0.4395413100719452, | |
| "learning_rate": 0.00011873813145857249, | |
| "loss": 0.5569, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 1.328, | |
| "grad_norm": 0.8984820246696472, | |
| "learning_rate": 0.00011791456448835825, | |
| "loss": 0.7088, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 1.336, | |
| "grad_norm": 0.7709664106369019, | |
| "learning_rate": 0.00011708974021232769, | |
| "loss": 0.6731, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 1.3439999999999999, | |
| "grad_norm": 0.6782217025756836, | |
| "learning_rate": 0.00011626371651948838, | |
| "loss": 0.6188, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 1.3519999999999999, | |
| "grad_norm": 0.6427358984947205, | |
| "learning_rate": 0.00011543655138302714, | |
| "loss": 0.7004, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 1.3599999999999999, | |
| "grad_norm": 0.5902594923973083, | |
| "learning_rate": 0.00011460830285624118, | |
| "loss": 0.5884, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 1.3679999999999999, | |
| "grad_norm": 0.5935835838317871, | |
| "learning_rate": 0.0001137790290684638, | |
| "loss": 0.5739, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 1.376, | |
| "grad_norm": 0.6752728223800659, | |
| "learning_rate": 0.00011294878822098469, | |
| "loss": 0.6435, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 1.384, | |
| "grad_norm": 0.7927135825157166, | |
| "learning_rate": 0.00011211763858296507, | |
| "loss": 0.6897, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 1.392, | |
| "grad_norm": 0.714499294757843, | |
| "learning_rate": 0.00011128563848734816, | |
| "loss": 0.6641, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "grad_norm": 0.7086356282234192, | |
| "learning_rate": 0.00011045284632676536, | |
| "loss": 0.6273, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 1.408, | |
| "grad_norm": 0.6125518679618835, | |
| "learning_rate": 0.00010961932054943778, | |
| "loss": 0.6437, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 1.416, | |
| "grad_norm": 0.5635287165641785, | |
| "learning_rate": 0.00010878511965507434, | |
| "loss": 0.6345, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 1.424, | |
| "grad_norm": 0.47936007380485535, | |
| "learning_rate": 0.00010795030219076599, | |
| "loss": 0.5913, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 1.432, | |
| "grad_norm": 0.7142558097839355, | |
| "learning_rate": 0.00010711492674687671, | |
| "loss": 0.6482, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "grad_norm": 0.5252729058265686, | |
| "learning_rate": 0.00010627905195293135, | |
| "loss": 0.6165, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 1.448, | |
| "grad_norm": 0.896318793296814, | |
| "learning_rate": 0.00010544273647350092, | |
| "loss": 0.634, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 1.456, | |
| "grad_norm": 0.6029036045074463, | |
| "learning_rate": 0.00010460603900408523, | |
| "loss": 0.6509, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 1.464, | |
| "grad_norm": 0.6835671663284302, | |
| "learning_rate": 0.00010376901826699348, | |
| "loss": 0.6212, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 1.472, | |
| "grad_norm": 0.7098750472068787, | |
| "learning_rate": 0.00010293173300722285, | |
| "loss": 0.7305, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "grad_norm": 0.675316333770752, | |
| "learning_rate": 0.0001020942419883357, | |
| "loss": 0.6685, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 1.488, | |
| "grad_norm": 0.6675406098365784, | |
| "learning_rate": 0.00010125660398833528, | |
| "loss": 0.6214, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 1.496, | |
| "grad_norm": 0.6629154682159424, | |
| "learning_rate": 0.0001004188777955404, | |
| "loss": 0.6035, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 1.504, | |
| "grad_norm": 0.7732692360877991, | |
| "learning_rate": 9.958112220445963e-05, | |
| "loss": 0.5868, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 1.512, | |
| "grad_norm": 0.6238484978675842, | |
| "learning_rate": 9.874339601166473e-05, | |
| "loss": 0.6003, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "grad_norm": 0.6622412800788879, | |
| "learning_rate": 9.790575801166432e-05, | |
| "loss": 0.5854, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 1.528, | |
| "grad_norm": 0.7598085999488831, | |
| "learning_rate": 9.706826699277718e-05, | |
| "loss": 0.5882, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 1.536, | |
| "grad_norm": 0.8870390057563782, | |
| "learning_rate": 9.623098173300654e-05, | |
| "loss": 0.7187, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 1.544, | |
| "grad_norm": 0.702232837677002, | |
| "learning_rate": 9.539396099591476e-05, | |
| "loss": 0.6156, | |
| "step": 965 | |
| }, | |
| { | |
| "epoch": 1.552, | |
| "grad_norm": 0.694139301776886, | |
| "learning_rate": 9.455726352649911e-05, | |
| "loss": 0.6488, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "grad_norm": 0.5684956312179565, | |
| "learning_rate": 9.372094804706867e-05, | |
| "loss": 0.6601, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 1.568, | |
| "grad_norm": 0.6327118277549744, | |
| "learning_rate": 9.288507325312335e-05, | |
| "loss": 0.5968, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 1.576, | |
| "grad_norm": 0.5732144117355347, | |
| "learning_rate": 9.204969780923403e-05, | |
| "loss": 0.7034, | |
| "step": 985 | |
| }, | |
| { | |
| "epoch": 1.584, | |
| "grad_norm": 0.7025273442268372, | |
| "learning_rate": 9.121488034492569e-05, | |
| "loss": 0.5973, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 1.592, | |
| "grad_norm": 0.9270740151405334, | |
| "learning_rate": 9.038067945056227e-05, | |
| "loss": 0.7877, | |
| "step": 995 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "grad_norm": 0.6676818132400513, | |
| "learning_rate": 8.954715367323468e-05, | |
| "loss": 0.644, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 1.608, | |
| "grad_norm": 0.6687547564506531, | |
| "learning_rate": 8.871436151265184e-05, | |
| "loss": 0.6678, | |
| "step": 1005 | |
| }, | |
| { | |
| "epoch": 1.616, | |
| "grad_norm": 0.7400322556495667, | |
| "learning_rate": 8.788236141703498e-05, | |
| "loss": 0.6088, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 1.624, | |
| "grad_norm": 0.5504963994026184, | |
| "learning_rate": 8.705121177901532e-05, | |
| "loss": 0.6219, | |
| "step": 1015 | |
| }, | |
| { | |
| "epoch": 1.6320000000000001, | |
| "grad_norm": 0.8088738322257996, | |
| "learning_rate": 8.62209709315362e-05, | |
| "loss": 0.6698, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 1.6400000000000001, | |
| "grad_norm": 0.7380816340446472, | |
| "learning_rate": 8.539169714375885e-05, | |
| "loss": 0.6207, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 1.6480000000000001, | |
| "grad_norm": 0.6346850395202637, | |
| "learning_rate": 8.456344861697289e-05, | |
| "loss": 0.626, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 1.6560000000000001, | |
| "grad_norm": 0.49918702244758606, | |
| "learning_rate": 8.373628348051165e-05, | |
| "loss": 0.6972, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 1.6640000000000001, | |
| "grad_norm": 0.7200607657432556, | |
| "learning_rate": 8.291025978767235e-05, | |
| "loss": 0.6282, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 1.6720000000000002, | |
| "grad_norm": 0.8350688815116882, | |
| "learning_rate": 8.208543551164178e-05, | |
| "loss": 0.6219, | |
| "step": 1045 | |
| }, | |
| { | |
| "epoch": 1.6800000000000002, | |
| "grad_norm": 0.5957468748092651, | |
| "learning_rate": 8.126186854142752e-05, | |
| "loss": 0.622, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 1.688, | |
| "grad_norm": 0.7494757175445557, | |
| "learning_rate": 8.04396166777952e-05, | |
| "loss": 0.5801, | |
| "step": 1055 | |
| }, | |
| { | |
| "epoch": 1.696, | |
| "grad_norm": 0.7109113931655884, | |
| "learning_rate": 7.961873762921153e-05, | |
| "loss": 0.598, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 1.704, | |
| "grad_norm": 0.7941219210624695, | |
| "learning_rate": 7.879928900779456e-05, | |
| "loss": 0.699, | |
| "step": 1065 | |
| }, | |
| { | |
| "epoch": 1.712, | |
| "grad_norm": 0.7282152771949768, | |
| "learning_rate": 7.798132832526986e-05, | |
| "loss": 0.5882, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "grad_norm": 0.5923735499382019, | |
| "learning_rate": 7.716491298893442e-05, | |
| "loss": 0.5321, | |
| "step": 1075 | |
| }, | |
| { | |
| "epoch": 1.728, | |
| "grad_norm": 0.5571523308753967, | |
| "learning_rate": 7.635010029762756e-05, | |
| "loss": 0.6647, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 1.736, | |
| "grad_norm": 0.6043043732643127, | |
| "learning_rate": 7.553694743770928e-05, | |
| "loss": 0.628, | |
| "step": 1085 | |
| }, | |
| { | |
| "epoch": 1.744, | |
| "grad_norm": 0.8733720183372498, | |
| "learning_rate": 7.472551147904708e-05, | |
| "loss": 0.6262, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 1.752, | |
| "grad_norm": 0.5561848878860474, | |
| "learning_rate": 7.391584937101033e-05, | |
| "loss": 0.6131, | |
| "step": 1095 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "grad_norm": 0.6611356735229492, | |
| "learning_rate": 7.310801793847344e-05, | |
| "loss": 0.6494, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 1.768, | |
| "grad_norm": 0.5700333714485168, | |
| "learning_rate": 7.230207387782776e-05, | |
| "loss": 0.5514, | |
| "step": 1105 | |
| }, | |
| { | |
| "epoch": 1.776, | |
| "grad_norm": 0.705662727355957, | |
| "learning_rate": 7.149807375300239e-05, | |
| "loss": 0.5823, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 1.784, | |
| "grad_norm": 0.9632449150085449, | |
| "learning_rate": 7.069607399149428e-05, | |
| "loss": 0.7207, | |
| "step": 1115 | |
| }, | |
| { | |
| "epoch": 1.792, | |
| "grad_norm": 0.6412242650985718, | |
| "learning_rate": 6.989613088040796e-05, | |
| "loss": 0.7006, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "grad_norm": 0.538044273853302, | |
| "learning_rate": 6.909830056250527e-05, | |
| "loss": 0.6222, | |
| "step": 1125 | |
| }, | |
| { | |
| "epoch": 1.808, | |
| "grad_norm": 0.6124762296676636, | |
| "learning_rate": 6.830263903226483e-05, | |
| "loss": 0.6569, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 1.8159999999999998, | |
| "grad_norm": 0.7250857949256897, | |
| "learning_rate": 6.750920213195238e-05, | |
| "loss": 0.543, | |
| "step": 1135 | |
| }, | |
| { | |
| "epoch": 1.8239999999999998, | |
| "grad_norm": 0.6840488910675049, | |
| "learning_rate": 6.671804554770135e-05, | |
| "loss": 0.6334, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 1.8319999999999999, | |
| "grad_norm": 0.8932304382324219, | |
| "learning_rate": 6.592922480560483e-05, | |
| "loss": 0.6701, | |
| "step": 1145 | |
| }, | |
| { | |
| "epoch": 1.8399999999999999, | |
| "grad_norm": 0.9220761656761169, | |
| "learning_rate": 6.51427952678185e-05, | |
| "loss": 0.6216, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 1.8479999999999999, | |
| "grad_norm": 0.688731849193573, | |
| "learning_rate": 6.435881212867493e-05, | |
| "loss": 0.5877, | |
| "step": 1155 | |
| }, | |
| { | |
| "epoch": 1.8559999999999999, | |
| "grad_norm": 0.6370134949684143, | |
| "learning_rate": 6.357733041081018e-05, | |
| "loss": 0.6256, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 1.8639999999999999, | |
| "grad_norm": 0.7421153783798218, | |
| "learning_rate": 6.27984049613019e-05, | |
| "loss": 0.7062, | |
| "step": 1165 | |
| }, | |
| { | |
| "epoch": 1.8719999999999999, | |
| "grad_norm": 0.8129108548164368, | |
| "learning_rate": 6.20220904478199e-05, | |
| "loss": 0.6304, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "grad_norm": 0.7866687774658203, | |
| "learning_rate": 6.12484413547897e-05, | |
| "loss": 0.7695, | |
| "step": 1175 | |
| }, | |
| { | |
| "epoch": 1.888, | |
| "grad_norm": 0.8307198882102966, | |
| "learning_rate": 6.047751197956838e-05, | |
| "loss": 0.5723, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 1.896, | |
| "grad_norm": 0.5541536211967468, | |
| "learning_rate": 5.9709356428633746e-05, | |
| "loss": 0.6847, | |
| "step": 1185 | |
| }, | |
| { | |
| "epoch": 1.904, | |
| "grad_norm": 0.731555163860321, | |
| "learning_rate": 5.8944028613787206e-05, | |
| "loss": 0.6618, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 1.912, | |
| "grad_norm": 1.020991325378418, | |
| "learning_rate": 5.818158224836987e-05, | |
| "loss": 0.6275, | |
| "step": 1195 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "grad_norm": 0.7172287702560425, | |
| "learning_rate": 5.7422070843492734e-05, | |
| "loss": 0.5617, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 1.928, | |
| "grad_norm": 0.7232096791267395, | |
| "learning_rate": 5.666554770428129e-05, | |
| "loss": 0.6278, | |
| "step": 1205 | |
| }, | |
| { | |
| "epoch": 1.936, | |
| "grad_norm": 0.7234194278717041, | |
| "learning_rate": 5.591206592613416e-05, | |
| "loss": 0.6713, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 1.944, | |
| "grad_norm": 0.6906010508537292, | |
| "learning_rate": 5.5161678390996796e-05, | |
| "loss": 0.6113, | |
| "step": 1215 | |
| }, | |
| { | |
| "epoch": 1.952, | |
| "grad_norm": 0.9057906270027161, | |
| "learning_rate": 5.441443776365003e-05, | |
| "loss": 0.5587, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "grad_norm": 0.7144932150840759, | |
| "learning_rate": 5.3670396488013854e-05, | |
| "loss": 0.5601, | |
| "step": 1225 | |
| }, | |
| { | |
| "epoch": 1.968, | |
| "grad_norm": 1.0071252584457397, | |
| "learning_rate": 5.292960678346675e-05, | |
| "loss": 0.5941, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 1.976, | |
| "grad_norm": 0.7061260342597961, | |
| "learning_rate": 5.2192120641180786e-05, | |
| "loss": 0.6285, | |
| "step": 1235 | |
| }, | |
| { | |
| "epoch": 1.984, | |
| "grad_norm": 0.6937069892883301, | |
| "learning_rate": 5.145798982047261e-05, | |
| "loss": 0.6516, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 1.992, | |
| "grad_norm": 0.5176392793655396, | |
| "learning_rate": 5.072726584517086e-05, | |
| "loss": 0.5904, | |
| "step": 1245 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "grad_norm": 0.8501409292221069, | |
| "learning_rate": 5.000000000000002e-05, | |
| "loss": 0.619, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 2.008, | |
| "grad_norm": 0.5451085567474365, | |
| "learning_rate": 4.927624332698109e-05, | |
| "loss": 0.6058, | |
| "step": 1255 | |
| }, | |
| { | |
| "epoch": 2.016, | |
| "grad_norm": 0.8410437107086182, | |
| "learning_rate": 4.8556046621849346e-05, | |
| "loss": 0.6248, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 2.024, | |
| "grad_norm": 0.6089378595352173, | |
| "learning_rate": 4.783946043048923e-05, | |
| "loss": 0.5247, | |
| "step": 1265 | |
| }, | |
| { | |
| "epoch": 2.032, | |
| "grad_norm": 0.47437986731529236, | |
| "learning_rate": 4.712653504538683e-05, | |
| "loss": 0.5439, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "grad_norm": 0.6925654411315918, | |
| "learning_rate": 4.6417320502100316e-05, | |
| "loss": 0.4491, | |
| "step": 1275 | |
| }, | |
| { | |
| "epoch": 2.048, | |
| "grad_norm": 0.8538162708282471, | |
| "learning_rate": 4.5711866575748276e-05, | |
| "loss": 0.52, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 2.056, | |
| "grad_norm": 0.8048768639564514, | |
| "learning_rate": 4.501022277751602e-05, | |
| "loss": 0.5259, | |
| "step": 1285 | |
| }, | |
| { | |
| "epoch": 2.064, | |
| "grad_norm": 0.7191641330718994, | |
| "learning_rate": 4.431243835118124e-05, | |
| "loss": 0.5025, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 2.072, | |
| "grad_norm": 1.0918892621994019, | |
| "learning_rate": 4.361856226965733e-05, | |
| "loss": 0.4772, | |
| "step": 1295 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "grad_norm": 0.8152824640274048, | |
| "learning_rate": 4.2928643231556844e-05, | |
| "loss": 0.5945, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 2.088, | |
| "grad_norm": 0.640073299407959, | |
| "learning_rate": 4.224272965777326e-05, | |
| "loss": 0.4813, | |
| "step": 1305 | |
| }, | |
| { | |
| "epoch": 2.096, | |
| "grad_norm": 0.7172432541847229, | |
| "learning_rate": 4.15608696880828e-05, | |
| "loss": 0.5315, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 2.104, | |
| "grad_norm": 1.1283674240112305, | |
| "learning_rate": 4.08831111777658e-05, | |
| "loss": 0.5591, | |
| "step": 1315 | |
| }, | |
| { | |
| "epoch": 2.112, | |
| "grad_norm": 0.8184736967086792, | |
| "learning_rate": 4.020950169424815e-05, | |
| "loss": 0.605, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "grad_norm": 0.6823618412017822, | |
| "learning_rate": 3.954008851376252e-05, | |
| "loss": 0.4955, | |
| "step": 1325 | |
| }, | |
| { | |
| "epoch": 2.128, | |
| "grad_norm": 0.8576385378837585, | |
| "learning_rate": 3.887491861803085e-05, | |
| "loss": 0.5757, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 2.136, | |
| "grad_norm": 0.967835009098053, | |
| "learning_rate": 3.821403869096658e-05, | |
| "loss": 0.5313, | |
| "step": 1335 | |
| }, | |
| { | |
| "epoch": 2.144, | |
| "grad_norm": 0.7330173254013062, | |
| "learning_rate": 3.755749511539845e-05, | |
| "loss": 0.5904, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 2.152, | |
| "grad_norm": 0.6664792895317078, | |
| "learning_rate": 3.690533396981504e-05, | |
| "loss": 0.4679, | |
| "step": 1345 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "grad_norm": 0.7639065980911255, | |
| "learning_rate": 3.6257601025131026e-05, | |
| "loss": 0.5235, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 2.168, | |
| "grad_norm": 0.6960520148277283, | |
| "learning_rate": 3.561434174147463e-05, | |
| "loss": 0.5797, | |
| "step": 1355 | |
| }, | |
| { | |
| "epoch": 2.176, | |
| "grad_norm": 0.8231356143951416, | |
| "learning_rate": 3.497560126499709e-05, | |
| "loss": 0.5772, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 2.184, | |
| "grad_norm": 0.8968437910079956, | |
| "learning_rate": 3.4341424424704375e-05, | |
| "loss": 0.5316, | |
| "step": 1365 | |
| }, | |
| { | |
| "epoch": 2.192, | |
| "grad_norm": 0.8037480711936951, | |
| "learning_rate": 3.371185572931048e-05, | |
| "loss": 0.5646, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "grad_norm": 0.9330148696899414, | |
| "learning_rate": 3.308693936411421e-05, | |
| "loss": 0.5431, | |
| "step": 1375 | |
| }, | |
| { | |
| "epoch": 2.208, | |
| "grad_norm": 0.6958775520324707, | |
| "learning_rate": 3.246671918789755e-05, | |
| "loss": 0.5403, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 2.216, | |
| "grad_norm": 0.9870476126670837, | |
| "learning_rate": 3.1851238729848034e-05, | |
| "loss": 0.5329, | |
| "step": 1385 | |
| }, | |
| { | |
| "epoch": 2.224, | |
| "grad_norm": 0.5036590099334717, | |
| "learning_rate": 3.124054118650327e-05, | |
| "loss": 0.5696, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 2.232, | |
| "grad_norm": 0.8640053868293762, | |
| "learning_rate": 3.063466941871952e-05, | |
| "loss": 0.59, | |
| "step": 1395 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "grad_norm": 0.6065173149108887, | |
| "learning_rate": 3.0033665948663448e-05, | |
| "loss": 0.5116, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 2.248, | |
| "grad_norm": 1.083775520324707, | |
| "learning_rate": 2.9437572956827964e-05, | |
| "loss": 0.5783, | |
| "step": 1405 | |
| }, | |
| { | |
| "epoch": 2.2560000000000002, | |
| "grad_norm": 0.7090497016906738, | |
| "learning_rate": 2.8846432279071467e-05, | |
| "loss": 0.6259, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 2.2640000000000002, | |
| "grad_norm": 0.742468535900116, | |
| "learning_rate": 2.826028540368215e-05, | |
| "loss": 0.5759, | |
| "step": 1415 | |
| }, | |
| { | |
| "epoch": 2.2720000000000002, | |
| "grad_norm": 0.9219839572906494, | |
| "learning_rate": 2.7679173468465812e-05, | |
| "loss": 0.497, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 2.2800000000000002, | |
| "grad_norm": 0.7159206867218018, | |
| "learning_rate": 2.7103137257858868e-05, | |
| "loss": 0.619, | |
| "step": 1425 | |
| }, | |
| { | |
| "epoch": 2.288, | |
| "grad_norm": 0.6997727751731873, | |
| "learning_rate": 2.6532217200065858e-05, | |
| "loss": 0.5858, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 2.296, | |
| "grad_norm": 0.7493643164634705, | |
| "learning_rate": 2.5966453364222186e-05, | |
| "loss": 0.6291, | |
| "step": 1435 | |
| }, | |
| { | |
| "epoch": 2.304, | |
| "grad_norm": 0.8311699032783508, | |
| "learning_rate": 2.540588545758179e-05, | |
| "loss": 0.6418, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 2.312, | |
| "grad_norm": 0.7084354758262634, | |
| "learning_rate": 2.48505528227304e-05, | |
| "loss": 0.5483, | |
| "step": 1445 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "grad_norm": 0.734438955783844, | |
| "learning_rate": 2.4300494434824373e-05, | |
| "loss": 0.6071, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 2.328, | |
| "grad_norm": 0.8913635015487671, | |
| "learning_rate": 2.37557488988552e-05, | |
| "loss": 0.5099, | |
| "step": 1455 | |
| }, | |
| { | |
| "epoch": 2.336, | |
| "grad_norm": 0.8349048495292664, | |
| "learning_rate": 2.321635444694028e-05, | |
| "loss": 0.5186, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 2.344, | |
| "grad_norm": 0.6164011359214783, | |
| "learning_rate": 2.2682348935639274e-05, | |
| "loss": 0.5043, | |
| "step": 1465 | |
| }, | |
| { | |
| "epoch": 2.352, | |
| "grad_norm": 1.044892430305481, | |
| "learning_rate": 2.2153769843297667e-05, | |
| "loss": 0.61, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "grad_norm": 0.9142879247665405, | |
| "learning_rate": 2.163065426741603e-05, | |
| "loss": 0.5987, | |
| "step": 1475 | |
| }, | |
| { | |
| "epoch": 2.368, | |
| "grad_norm": 0.6232836842536926, | |
| "learning_rate": 2.1113038922046602e-05, | |
| "loss": 0.5212, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 2.376, | |
| "grad_norm": 0.49558231234550476, | |
| "learning_rate": 2.0600960135216462e-05, | |
| "loss": 0.4796, | |
| "step": 1485 | |
| }, | |
| { | |
| "epoch": 2.384, | |
| "grad_norm": 0.7887687683105469, | |
| "learning_rate": 2.009445384637805e-05, | |
| "loss": 0.4844, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 2.392, | |
| "grad_norm": 0.8086990714073181, | |
| "learning_rate": 1.9593555603886538e-05, | |
| "loss": 0.5085, | |
| "step": 1495 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "grad_norm": 0.6713303327560425, | |
| "learning_rate": 1.9098300562505266e-05, | |
| "loss": 0.4839, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 2.408, | |
| "grad_norm": 0.6262741684913635, | |
| "learning_rate": 1.8608723480938206e-05, | |
| "loss": 0.5715, | |
| "step": 1505 | |
| }, | |
| { | |
| "epoch": 2.416, | |
| "grad_norm": 0.8025808334350586, | |
| "learning_rate": 1.812485871939056e-05, | |
| "loss": 0.5266, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 2.424, | |
| "grad_norm": 0.8753231167793274, | |
| "learning_rate": 1.7646740237157256e-05, | |
| "loss": 0.5422, | |
| "step": 1515 | |
| }, | |
| { | |
| "epoch": 2.432, | |
| "grad_norm": 0.6459301710128784, | |
| "learning_rate": 1.7174401590239587e-05, | |
| "loss": 0.5553, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "grad_norm": 0.6917416453361511, | |
| "learning_rate": 1.6707875928990058e-05, | |
| "loss": 0.5765, | |
| "step": 1525 | |
| }, | |
| { | |
| "epoch": 2.448, | |
| "grad_norm": 0.7890029549598694, | |
| "learning_rate": 1.6247195995785837e-05, | |
| "loss": 0.549, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 2.456, | |
| "grad_norm": 0.9913660883903503, | |
| "learning_rate": 1.579239412273078e-05, | |
| "loss": 0.4876, | |
| "step": 1535 | |
| }, | |
| { | |
| "epoch": 2.464, | |
| "grad_norm": 0.9030985832214355, | |
| "learning_rate": 1.5343502229386207e-05, | |
| "loss": 0.5546, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 2.472, | |
| "grad_norm": 0.9133403301239014, | |
| "learning_rate": 1.4900551820530828e-05, | |
| "loss": 0.5356, | |
| "step": 1545 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "grad_norm": 0.7083793878555298, | |
| "learning_rate": 1.4463573983949341e-05, | |
| "loss": 0.5142, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 2.488, | |
| "grad_norm": 1.095435619354248, | |
| "learning_rate": 1.40325993882509e-05, | |
| "loss": 0.6054, | |
| "step": 1555 | |
| }, | |
| { | |
| "epoch": 2.496, | |
| "grad_norm": 0.8825190663337708, | |
| "learning_rate": 1.3607658280716473e-05, | |
| "loss": 0.5294, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 2.504, | |
| "grad_norm": 0.9436343908309937, | |
| "learning_rate": 1.3188780485176088e-05, | |
| "loss": 0.5294, | |
| "step": 1565 | |
| }, | |
| { | |
| "epoch": 2.512, | |
| "grad_norm": 1.0125439167022705, | |
| "learning_rate": 1.2775995399915631e-05, | |
| "loss": 0.4905, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "grad_norm": 0.8476350903511047, | |
| "learning_rate": 1.2369331995613665e-05, | |
| "loss": 0.5186, | |
| "step": 1575 | |
| }, | |
| { | |
| "epoch": 2.528, | |
| "grad_norm": 0.9092681407928467, | |
| "learning_rate": 1.196881881330798e-05, | |
| "loss": 0.4909, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 2.536, | |
| "grad_norm": 0.7970360517501831, | |
| "learning_rate": 1.1574483962392767e-05, | |
| "loss": 0.5303, | |
| "step": 1585 | |
| }, | |
| { | |
| "epoch": 2.544, | |
| "grad_norm": 0.8575041890144348, | |
| "learning_rate": 1.1186355118645554e-05, | |
| "loss": 0.5169, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 2.552, | |
| "grad_norm": 0.7397408485412598, | |
| "learning_rate": 1.0804459522284926e-05, | |
| "loss": 0.5339, | |
| "step": 1595 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "grad_norm": 0.7415968179702759, | |
| "learning_rate": 1.042882397605871e-05, | |
| "loss": 0.5283, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 2.568, | |
| "grad_norm": 0.7035180926322937, | |
| "learning_rate": 1.0059474843362892e-05, | |
| "loss": 0.5576, | |
| "step": 1605 | |
| }, | |
| { | |
| "epoch": 2.576, | |
| "grad_norm": 0.9805112481117249, | |
| "learning_rate": 9.696438046391288e-06, | |
| "loss": 0.5136, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 2.584, | |
| "grad_norm": 0.6661838889122009, | |
| "learning_rate": 9.339739064316233e-06, | |
| "loss": 0.5885, | |
| "step": 1615 | |
| }, | |
| { | |
| "epoch": 2.592, | |
| "grad_norm": 0.8581559062004089, | |
| "learning_rate": 8.989402931500434e-06, | |
| "loss": 0.5, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "grad_norm": 0.7146279811859131, | |
| "learning_rate": 8.645454235739903e-06, | |
| "loss": 0.5325, | |
| "step": 1625 | |
| }, | |
| { | |
| "epoch": 2.608, | |
| "grad_norm": 0.9474234580993652, | |
| "learning_rate": 8.307917116538378e-06, | |
| "loss": 0.5772, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 2.616, | |
| "grad_norm": 0.9583209753036499, | |
| "learning_rate": 7.976815263412963e-06, | |
| "loss": 0.5736, | |
| "step": 1635 | |
| }, | |
| { | |
| "epoch": 2.624, | |
| "grad_norm": 0.7156705260276794, | |
| "learning_rate": 7.652171914231776e-06, | |
| "loss": 0.5199, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 2.632, | |
| "grad_norm": 0.8224849700927734, | |
| "learning_rate": 7.3340098535827905e-06, | |
| "loss": 0.5753, | |
| "step": 1645 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "grad_norm": 0.8689257502555847, | |
| "learning_rate": 7.022351411174866e-06, | |
| "loss": 0.5424, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 2.648, | |
| "grad_norm": 0.6636053323745728, | |
| "learning_rate": 6.717218460270536e-06, | |
| "loss": 0.5555, | |
| "step": 1655 | |
| }, | |
| { | |
| "epoch": 2.656, | |
| "grad_norm": 0.8688860535621643, | |
| "learning_rate": 6.418632416150927e-06, | |
| "loss": 0.4936, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 2.664, | |
| "grad_norm": 0.6272854208946228, | |
| "learning_rate": 6.126614234612593e-06, | |
| "loss": 0.6291, | |
| "step": 1665 | |
| }, | |
| { | |
| "epoch": 2.672, | |
| "grad_norm": 1.2240337133407593, | |
| "learning_rate": 5.8411844104969916e-06, | |
| "loss": 0.5197, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "grad_norm": 0.9820936918258667, | |
| "learning_rate": 5.562362976251901e-06, | |
| "loss": 0.5398, | |
| "step": 1675 | |
| }, | |
| { | |
| "epoch": 2.6879999999999997, | |
| "grad_norm": 1.1582359075546265, | |
| "learning_rate": 5.290169500525577e-06, | |
| "loss": 0.6059, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 2.6959999999999997, | |
| "grad_norm": 0.5501114726066589, | |
| "learning_rate": 5.024623086793323e-06, | |
| "loss": 0.531, | |
| "step": 1685 | |
| }, | |
| { | |
| "epoch": 2.7039999999999997, | |
| "grad_norm": 0.8848717212677002, | |
| "learning_rate": 4.765742372016735e-06, | |
| "loss": 0.6054, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 2.7119999999999997, | |
| "grad_norm": 0.7358693480491638, | |
| "learning_rate": 4.513545525335705e-06, | |
| "loss": 0.5173, | |
| "step": 1695 | |
| }, | |
| { | |
| "epoch": 2.7199999999999998, | |
| "grad_norm": 0.9218215942382812, | |
| "learning_rate": 4.268050246793276e-06, | |
| "loss": 0.4944, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 2.7279999999999998, | |
| "grad_norm": 0.6374716758728027, | |
| "learning_rate": 4.029273766093333e-06, | |
| "loss": 0.5183, | |
| "step": 1705 | |
| }, | |
| { | |
| "epoch": 2.7359999999999998, | |
| "grad_norm": 0.583243191242218, | |
| "learning_rate": 3.797232841391407e-06, | |
| "loss": 0.668, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 2.7439999999999998, | |
| "grad_norm": 0.8384690284729004, | |
| "learning_rate": 3.5719437581185454e-06, | |
| "loss": 0.5068, | |
| "step": 1715 | |
| }, | |
| { | |
| "epoch": 2.752, | |
| "grad_norm": 0.8034130334854126, | |
| "learning_rate": 3.3534223278382405e-06, | |
| "loss": 0.5823, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "grad_norm": 0.8146041631698608, | |
| "learning_rate": 3.1416838871368924e-06, | |
| "loss": 0.6111, | |
| "step": 1725 | |
| }, | |
| { | |
| "epoch": 2.768, | |
| "grad_norm": 0.8122982382774353, | |
| "learning_rate": 2.936743296547273e-06, | |
| "loss": 0.5231, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 2.776, | |
| "grad_norm": 0.7326982021331787, | |
| "learning_rate": 2.738614939505646e-06, | |
| "loss": 0.5236, | |
| "step": 1735 | |
| }, | |
| { | |
| "epoch": 2.784, | |
| "grad_norm": 0.7472147345542908, | |
| "learning_rate": 2.5473127213422763e-06, | |
| "loss": 0.5657, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 2.792, | |
| "grad_norm": 0.8197700381278992, | |
| "learning_rate": 2.3628500683055222e-06, | |
| "loss": 0.5518, | |
| "step": 1745 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "grad_norm": 0.8733732104301453, | |
| "learning_rate": 2.1852399266194314e-06, | |
| "loss": 0.4908, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 2.808, | |
| "grad_norm": 0.8913092017173767, | |
| "learning_rate": 2.014494761575314e-06, | |
| "loss": 0.5459, | |
| "step": 1755 | |
| }, | |
| { | |
| "epoch": 2.816, | |
| "grad_norm": 1.1259772777557373, | |
| "learning_rate": 1.8506265566567094e-06, | |
| "loss": 0.5208, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 2.824, | |
| "grad_norm": 0.7692184448242188, | |
| "learning_rate": 1.6936468126984572e-06, | |
| "loss": 0.5824, | |
| "step": 1765 | |
| }, | |
| { | |
| "epoch": 2.832, | |
| "grad_norm": 0.588602602481842, | |
| "learning_rate": 1.543566547079467e-06, | |
| "loss": 0.5512, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "grad_norm": 0.6324055790901184, | |
| "learning_rate": 1.400396292949513e-06, | |
| "loss": 0.6327, | |
| "step": 1775 | |
| }, | |
| { | |
| "epoch": 2.848, | |
| "grad_norm": 0.7608378529548645, | |
| "learning_rate": 1.26414609848996e-06, | |
| "loss": 0.5292, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 2.856, | |
| "grad_norm": 0.7972851395606995, | |
| "learning_rate": 1.134825526208605e-06, | |
| "loss": 0.5692, | |
| "step": 1785 | |
| }, | |
| { | |
| "epoch": 2.864, | |
| "grad_norm": 0.9705446362495422, | |
| "learning_rate": 1.0124436522684243e-06, | |
| "loss": 0.5532, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 2.872, | |
| "grad_norm": 0.6317399144172668, | |
| "learning_rate": 8.970090658507291e-07, | |
| "loss": 0.5314, | |
| "step": 1795 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "grad_norm": 0.6457757949829102, | |
| "learning_rate": 7.885298685522235e-07, | |
| "loss": 0.524, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 2.888, | |
| "grad_norm": 0.8593656420707703, | |
| "learning_rate": 6.870136738164612e-07, | |
| "loss": 0.5227, | |
| "step": 1805 | |
| }, | |
| { | |
| "epoch": 2.896, | |
| "grad_norm": 1.0187020301818848, | |
| "learning_rate": 5.924676063995382e-07, | |
| "loss": 0.5993, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 2.904, | |
| "grad_norm": 0.7082214951515198, | |
| "learning_rate": 5.048983018699827e-07, | |
| "loss": 0.5618, | |
| "step": 1815 | |
| }, | |
| { | |
| "epoch": 2.912, | |
| "grad_norm": 0.6521438956260681, | |
| "learning_rate": 4.2431190614309335e-07, | |
| "loss": 0.5504, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "grad_norm": 0.8906036615371704, | |
| "learning_rate": 3.50714075049563e-07, | |
| "loss": 0.5147, | |
| "step": 1825 | |
| }, | |
| { | |
| "epoch": 2.928, | |
| "grad_norm": 1.0908008813858032, | |
| "learning_rate": 2.841099739386066e-07, | |
| "loss": 0.5564, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 2.936, | |
| "grad_norm": 0.6374122500419617, | |
| "learning_rate": 2.2450427731534053e-07, | |
| "loss": 0.5188, | |
| "step": 1835 | |
| }, | |
| { | |
| "epoch": 2.944, | |
| "grad_norm": 0.9616740345954895, | |
| "learning_rate": 1.7190116851280026e-07, | |
| "loss": 0.5438, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 2.952, | |
| "grad_norm": 1.0712924003601074, | |
| "learning_rate": 1.2630433939825327e-07, | |
| "loss": 0.4962, | |
| "step": 1845 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "grad_norm": 0.8226613998413086, | |
| "learning_rate": 8.771699011416168e-08, | |
| "loss": 0.5021, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 2.968, | |
| "grad_norm": 0.9519492983818054, | |
| "learning_rate": 5.6141828853573106e-08, | |
| "loss": 0.5277, | |
| "step": 1855 | |
| }, | |
| { | |
| "epoch": 2.976, | |
| "grad_norm": 0.9817518591880798, | |
| "learning_rate": 3.1581071670006015e-08, | |
| "loss": 0.5764, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 2.984, | |
| "grad_norm": 0.7039242386817932, | |
| "learning_rate": 1.4036442321962995e-08, | |
| "loss": 0.5408, | |
| "step": 1865 | |
| }, | |
| { | |
| "epoch": 2.992, | |
| "grad_norm": 0.591012179851532, | |
| "learning_rate": 3.509172151938689e-09, | |
| "loss": 0.5014, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "grad_norm": 0.7672661542892456, | |
| "learning_rate": 0.0, | |
| "loss": 0.6182, | |
| "step": 1875 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "step": 1875, | |
| "total_flos": 1.1764404625814323e+17, | |
| "train_loss": 0.6465437274932861, | |
| "train_runtime": 2365.2475, | |
| "train_samples_per_second": 12.684, | |
| "train_steps_per_second": 0.793 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 1875, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 100, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.1764404625814323e+17, | |
| "train_batch_size": 16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |