NLLB-DOLMA / trainer_state.json
SinaAhmadi's picture
best model config +
420e9a3
{
"best_metric": 14.6166,
"best_model_checkpoint": "./nllb_finetuned_base_II_improved/checkpoint-18612",
"epoch": 49.87428932406822,
"eval_steps": 500,
"global_step": 19750,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.002526847757422615,
"grad_norm": 0.03834182024002075,
"learning_rate": 5.063291139240507e-08,
"loss": 0.0009,
"step": 1
},
{
"epoch": 0.2526847757422615,
"grad_norm": 0.055799346417188644,
"learning_rate": 5.063291139240506e-06,
"loss": 0.002,
"step": 100
},
{
"epoch": 0.505369551484523,
"grad_norm": 0.5792537927627563,
"learning_rate": 1.0126582278481012e-05,
"loss": 0.0021,
"step": 200
},
{
"epoch": 0.7580543272267846,
"grad_norm": 0.008475471287965775,
"learning_rate": 1.5189873417721521e-05,
"loss": 0.0023,
"step": 300
},
{
"epoch": 1.0,
"eval_bleu": 14.2158,
"eval_gen_len": 14.4698,
"eval_loss": 5.007449150085449,
"eval_runtime": 537.7389,
"eval_samples_per_second": 10.937,
"eval_steps_per_second": 1.369,
"step": 396
},
{
"epoch": 1.0101073910296905,
"grad_norm": 0.264018177986145,
"learning_rate": 2.0253164556962025e-05,
"loss": 0.0032,
"step": 400
},
{
"epoch": 1.262792166771952,
"grad_norm": 0.011777933686971664,
"learning_rate": 2.5316455696202533e-05,
"loss": 0.0025,
"step": 500
},
{
"epoch": 1.5154769425142134,
"grad_norm": 0.027416467666625977,
"learning_rate": 3.0379746835443042e-05,
"loss": 0.0027,
"step": 600
},
{
"epoch": 1.768161718256475,
"grad_norm": 0.14406521618366241,
"learning_rate": 3.5443037974683544e-05,
"loss": 0.0024,
"step": 700
},
{
"epoch": 2.0,
"eval_bleu": 14.2734,
"eval_gen_len": 14.3467,
"eval_loss": 5.034389972686768,
"eval_runtime": 533.7562,
"eval_samples_per_second": 11.018,
"eval_steps_per_second": 1.379,
"step": 792
},
{
"epoch": 2.020214782059381,
"grad_norm": 0.0178065225481987,
"learning_rate": 4.050632911392405e-05,
"loss": 0.0024,
"step": 800
},
{
"epoch": 2.2728995578016424,
"grad_norm": 0.018611092120409012,
"learning_rate": 4.556962025316456e-05,
"loss": 0.002,
"step": 900
},
{
"epoch": 2.525584333543904,
"grad_norm": 0.01996198110282421,
"learning_rate": 5.0632911392405066e-05,
"loss": 0.004,
"step": 1000
},
{
"epoch": 2.7782691092861658,
"grad_norm": 0.01915626786649227,
"learning_rate": 5.569620253164557e-05,
"loss": 0.0027,
"step": 1100
},
{
"epoch": 3.0,
"eval_bleu": 14.2059,
"eval_gen_len": 14.3137,
"eval_loss": 5.052088737487793,
"eval_runtime": 529.7444,
"eval_samples_per_second": 11.102,
"eval_steps_per_second": 1.389,
"step": 1188
},
{
"epoch": 3.0303221730890715,
"grad_norm": 0.06470153480768204,
"learning_rate": 6.0759493670886084e-05,
"loss": 0.0032,
"step": 1200
},
{
"epoch": 3.283006948831333,
"grad_norm": 0.019087791442871094,
"learning_rate": 6.582278481012658e-05,
"loss": 0.0029,
"step": 1300
},
{
"epoch": 3.5356917245735944,
"grad_norm": 0.7321619391441345,
"learning_rate": 7.088607594936709e-05,
"loss": 0.004,
"step": 1400
},
{
"epoch": 3.788376500315856,
"grad_norm": 0.08558339625597,
"learning_rate": 7.59493670886076e-05,
"loss": 0.0039,
"step": 1500
},
{
"epoch": 4.0,
"eval_bleu": 13.7306,
"eval_gen_len": 14.7393,
"eval_loss": 5.003371715545654,
"eval_runtime": 542.3943,
"eval_samples_per_second": 10.843,
"eval_steps_per_second": 1.357,
"step": 1584
},
{
"epoch": 4.040429564118762,
"grad_norm": 0.026173993945121765,
"learning_rate": 8.10126582278481e-05,
"loss": 0.0044,
"step": 1600
},
{
"epoch": 4.293114339861023,
"grad_norm": 0.01453397236764431,
"learning_rate": 8.607594936708861e-05,
"loss": 0.0051,
"step": 1700
},
{
"epoch": 4.545799115603285,
"grad_norm": 0.01603219285607338,
"learning_rate": 9.113924050632912e-05,
"loss": 0.0061,
"step": 1800
},
{
"epoch": 4.798483891345547,
"grad_norm": 0.0410076268017292,
"learning_rate": 9.620253164556962e-05,
"loss": 0.0069,
"step": 1900
},
{
"epoch": 5.0,
"eval_bleu": 13.8802,
"eval_gen_len": 14.4926,
"eval_loss": 5.007750988006592,
"eval_runtime": 536.2346,
"eval_samples_per_second": 10.967,
"eval_steps_per_second": 1.373,
"step": 1980
},
{
"epoch": 5.050536955148452,
"grad_norm": 0.03955426439642906,
"learning_rate": 0.00010126582278481013,
"loss": 0.0062,
"step": 2000
},
{
"epoch": 5.303221730890714,
"grad_norm": 0.07262468338012695,
"learning_rate": 0.00010632911392405063,
"loss": 0.0082,
"step": 2100
},
{
"epoch": 5.555906506632976,
"grad_norm": 0.8466947078704834,
"learning_rate": 0.00011139240506329114,
"loss": 0.011,
"step": 2200
},
{
"epoch": 5.808591282375237,
"grad_norm": 0.4297804832458496,
"learning_rate": 0.00011645569620253166,
"loss": 0.013,
"step": 2300
},
{
"epoch": 6.0,
"eval_bleu": 13.5899,
"eval_gen_len": 14.494,
"eval_loss": 4.995738983154297,
"eval_runtime": 536.5639,
"eval_samples_per_second": 10.96,
"eval_steps_per_second": 1.372,
"step": 2376
},
{
"epoch": 6.060644346178143,
"grad_norm": 0.14445021748542786,
"learning_rate": 0.00012151898734177217,
"loss": 0.0133,
"step": 2400
},
{
"epoch": 6.313329121920404,
"grad_norm": 0.1337200254201889,
"learning_rate": 0.00012658227848101267,
"loss": 0.0115,
"step": 2500
},
{
"epoch": 6.566013897662666,
"grad_norm": 0.5902842879295349,
"learning_rate": 0.00013164556962025315,
"loss": 0.0124,
"step": 2600
},
{
"epoch": 6.818698673404928,
"grad_norm": 0.709742546081543,
"learning_rate": 0.00013670886075949366,
"loss": 0.0165,
"step": 2700
},
{
"epoch": 7.0,
"eval_bleu": 13.324,
"eval_gen_len": 14.9148,
"eval_loss": 4.997140407562256,
"eval_runtime": 541.5428,
"eval_samples_per_second": 10.86,
"eval_steps_per_second": 1.359,
"step": 2772
},
{
"epoch": 7.070751737207833,
"grad_norm": 0.3325135111808777,
"learning_rate": 0.00014177215189873418,
"loss": 0.0205,
"step": 2800
},
{
"epoch": 7.323436512950095,
"grad_norm": 0.8612397909164429,
"learning_rate": 0.0001468354430379747,
"loss": 0.0141,
"step": 2900
},
{
"epoch": 7.576121288692356,
"grad_norm": 0.5200220346450806,
"learning_rate": 0.0001518987341772152,
"loss": 0.0131,
"step": 3000
},
{
"epoch": 7.828806064434618,
"grad_norm": 0.18251368403434753,
"learning_rate": 0.00015696202531645568,
"loss": 0.0195,
"step": 3100
},
{
"epoch": 8.0,
"eval_bleu": 13.5516,
"eval_gen_len": 14.4363,
"eval_loss": 4.994857311248779,
"eval_runtime": 535.7057,
"eval_samples_per_second": 10.978,
"eval_steps_per_second": 1.374,
"step": 3168
},
{
"epoch": 8.080859128237524,
"grad_norm": 0.8557725548744202,
"learning_rate": 0.0001620253164556962,
"loss": 0.0207,
"step": 3200
},
{
"epoch": 8.333543903979786,
"grad_norm": 0.6467623114585876,
"learning_rate": 0.0001670886075949367,
"loss": 0.0223,
"step": 3300
},
{
"epoch": 8.586228679722046,
"grad_norm": 0.9199110269546509,
"learning_rate": 0.00017215189873417722,
"loss": 0.0197,
"step": 3400
},
{
"epoch": 8.838913455464308,
"grad_norm": 0.3197132647037506,
"learning_rate": 0.00017721518987341773,
"loss": 0.0218,
"step": 3500
},
{
"epoch": 9.0,
"eval_bleu": 13.6364,
"eval_gen_len": 14.1306,
"eval_loss": 4.960774898529053,
"eval_runtime": 526.2912,
"eval_samples_per_second": 11.174,
"eval_steps_per_second": 1.398,
"step": 3564
},
{
"epoch": 9.090966519267214,
"grad_norm": 0.1220143586397171,
"learning_rate": 0.00018227848101265824,
"loss": 0.0297,
"step": 3600
},
{
"epoch": 9.343651295009476,
"grad_norm": 0.31142914295196533,
"learning_rate": 0.00018734177215189873,
"loss": 0.0212,
"step": 3700
},
{
"epoch": 9.596336070751738,
"grad_norm": 0.7474793791770935,
"learning_rate": 0.00019240506329113924,
"loss": 0.0218,
"step": 3800
},
{
"epoch": 9.849020846494,
"grad_norm": 0.4890414774417877,
"learning_rate": 0.00019746835443037975,
"loss": 0.0249,
"step": 3900
},
{
"epoch": 10.0,
"eval_bleu": 13.1309,
"eval_gen_len": 14.3164,
"eval_loss": 4.990716934204102,
"eval_runtime": 536.4931,
"eval_samples_per_second": 10.962,
"eval_steps_per_second": 1.372,
"step": 3960
},
{
"epoch": 10.101073910296904,
"grad_norm": 0.650865375995636,
"learning_rate": 0.00019936740506329114,
"loss": 0.0282,
"step": 4000
},
{
"epoch": 10.353758686039166,
"grad_norm": 0.616794764995575,
"learning_rate": 0.00019810221518987343,
"loss": 0.0232,
"step": 4100
},
{
"epoch": 10.606443461781428,
"grad_norm": 0.1520662158727646,
"learning_rate": 0.00019683702531645572,
"loss": 0.0271,
"step": 4200
},
{
"epoch": 10.85912823752369,
"grad_norm": 0.6819707751274109,
"learning_rate": 0.00019557183544303798,
"loss": 0.0237,
"step": 4300
},
{
"epoch": 11.0,
"eval_bleu": 13.389,
"eval_gen_len": 14.4307,
"eval_loss": 4.9949188232421875,
"eval_runtime": 532.7791,
"eval_samples_per_second": 11.038,
"eval_steps_per_second": 1.381,
"step": 4356
},
{
"epoch": 11.111181301326596,
"grad_norm": 0.14406563341617584,
"learning_rate": 0.00019430664556962024,
"loss": 0.0236,
"step": 4400
},
{
"epoch": 11.363866077068856,
"grad_norm": 0.30580630898475647,
"learning_rate": 0.00019304145569620256,
"loss": 0.0199,
"step": 4500
},
{
"epoch": 11.616550852811118,
"grad_norm": 0.2938636541366577,
"learning_rate": 0.00019177626582278482,
"loss": 0.0212,
"step": 4600
},
{
"epoch": 11.86923562855338,
"grad_norm": 0.30495572090148926,
"learning_rate": 0.00019051107594936708,
"loss": 0.0183,
"step": 4700
},
{
"epoch": 12.0,
"eval_bleu": 13.4564,
"eval_gen_len": 14.6526,
"eval_loss": 5.026724338531494,
"eval_runtime": 539.3787,
"eval_samples_per_second": 10.903,
"eval_steps_per_second": 1.365,
"step": 4752
},
{
"epoch": 12.121288692356286,
"grad_norm": 0.4794355630874634,
"learning_rate": 0.0001892458860759494,
"loss": 0.0209,
"step": 4800
},
{
"epoch": 12.373973468098548,
"grad_norm": 0.08356456458568573,
"learning_rate": 0.00018798069620253166,
"loss": 0.0181,
"step": 4900
},
{
"epoch": 12.626658243840808,
"grad_norm": 0.12282651662826538,
"learning_rate": 0.00018671550632911392,
"loss": 0.0164,
"step": 5000
},
{
"epoch": 12.87934301958307,
"grad_norm": 0.44435593485832214,
"learning_rate": 0.0001854503164556962,
"loss": 0.0212,
"step": 5100
},
{
"epoch": 13.0,
"eval_bleu": 13.59,
"eval_gen_len": 14.2952,
"eval_loss": 5.072442054748535,
"eval_runtime": 529.5691,
"eval_samples_per_second": 11.105,
"eval_steps_per_second": 1.39,
"step": 5148
},
{
"epoch": 13.131396083385976,
"grad_norm": 0.24175015091896057,
"learning_rate": 0.0001841851265822785,
"loss": 0.0175,
"step": 5200
},
{
"epoch": 13.384080859128238,
"grad_norm": 0.18917737901210785,
"learning_rate": 0.00018291993670886076,
"loss": 0.0154,
"step": 5300
},
{
"epoch": 13.6367656348705,
"grad_norm": 0.10555186867713928,
"learning_rate": 0.00018165474683544305,
"loss": 0.0171,
"step": 5400
},
{
"epoch": 13.88945041061276,
"grad_norm": 0.5594084858894348,
"learning_rate": 0.00018038955696202534,
"loss": 0.0158,
"step": 5500
},
{
"epoch": 14.0,
"eval_bleu": 13.3564,
"eval_gen_len": 14.5018,
"eval_loss": 5.083232879638672,
"eval_runtime": 540.6508,
"eval_samples_per_second": 10.878,
"eval_steps_per_second": 1.361,
"step": 5544
},
{
"epoch": 14.141503474415666,
"grad_norm": 0.5325390100479126,
"learning_rate": 0.0001791243670886076,
"loss": 0.0148,
"step": 5600
},
{
"epoch": 14.394188250157928,
"grad_norm": 0.16868151724338531,
"learning_rate": 0.00017785917721518986,
"loss": 0.0136,
"step": 5700
},
{
"epoch": 14.64687302590019,
"grad_norm": 0.22546489536762238,
"learning_rate": 0.00017659398734177218,
"loss": 0.0129,
"step": 5800
},
{
"epoch": 14.899557801642452,
"grad_norm": 0.4606294631958008,
"learning_rate": 0.00017532879746835444,
"loss": 0.0149,
"step": 5900
},
{
"epoch": 15.0,
"eval_bleu": 13.71,
"eval_gen_len": 14.4261,
"eval_loss": 5.047977447509766,
"eval_runtime": 536.9368,
"eval_samples_per_second": 10.953,
"eval_steps_per_second": 1.371,
"step": 5940
},
{
"epoch": 15.151610865445356,
"grad_norm": 0.2645217776298523,
"learning_rate": 0.0001740636075949367,
"loss": 0.0149,
"step": 6000
},
{
"epoch": 15.404295641187618,
"grad_norm": 0.18864086270332336,
"learning_rate": 0.000172798417721519,
"loss": 0.0133,
"step": 6100
},
{
"epoch": 15.65698041692988,
"grad_norm": 0.3654481768608093,
"learning_rate": 0.00017153322784810128,
"loss": 0.0127,
"step": 6200
},
{
"epoch": 15.909665192672142,
"grad_norm": 0.47671639919281006,
"learning_rate": 0.00017026803797468354,
"loss": 0.0152,
"step": 6300
},
{
"epoch": 16.0,
"eval_bleu": 13.3368,
"eval_gen_len": 14.4033,
"eval_loss": 5.045422554016113,
"eval_runtime": 533.8922,
"eval_samples_per_second": 11.015,
"eval_steps_per_second": 1.379,
"step": 6336
},
{
"epoch": 16.161718256475048,
"grad_norm": 0.23764218389987946,
"learning_rate": 0.00016900284810126583,
"loss": 0.0174,
"step": 6400
},
{
"epoch": 16.414403032217308,
"grad_norm": 0.2533399164676666,
"learning_rate": 0.00016775031012658228,
"loss": 0.016,
"step": 6500
},
{
"epoch": 16.66708780795957,
"grad_norm": 0.1537361741065979,
"learning_rate": 0.00016648512025316457,
"loss": 0.015,
"step": 6600
},
{
"epoch": 16.919772583701832,
"grad_norm": 0.2747589349746704,
"learning_rate": 0.00016521993037974683,
"loss": 0.0179,
"step": 6700
},
{
"epoch": 17.0,
"eval_bleu": 13.2518,
"eval_gen_len": 14.4889,
"eval_loss": 5.028156280517578,
"eval_runtime": 538.1372,
"eval_samples_per_second": 10.928,
"eval_steps_per_second": 1.368,
"step": 6732
},
{
"epoch": 17.171825647504736,
"grad_norm": 0.216201052069664,
"learning_rate": 0.00016395474050632914,
"loss": 0.0151,
"step": 6800
},
{
"epoch": 17.424510423247,
"grad_norm": 0.08341749012470245,
"learning_rate": 0.00016270220253164556,
"loss": 0.0119,
"step": 6900
},
{
"epoch": 17.67719519898926,
"grad_norm": 0.6413178443908691,
"learning_rate": 0.00016143701265822785,
"loss": 0.0145,
"step": 7000
},
{
"epoch": 17.929879974731524,
"grad_norm": 0.05727590620517731,
"learning_rate": 0.00016018447468354433,
"loss": 0.0139,
"step": 7100
},
{
"epoch": 18.0,
"eval_bleu": 13.4478,
"eval_gen_len": 14.5729,
"eval_loss": 5.039746284484863,
"eval_runtime": 538.0034,
"eval_samples_per_second": 10.931,
"eval_steps_per_second": 1.368,
"step": 7128
},
{
"epoch": 18.181933038534428,
"grad_norm": 0.09456099569797516,
"learning_rate": 0.0001589192848101266,
"loss": 0.0107,
"step": 7200
},
{
"epoch": 18.434617814276688,
"grad_norm": 0.5516796112060547,
"learning_rate": 0.00015765409493670885,
"loss": 0.0082,
"step": 7300
},
{
"epoch": 18.68730259001895,
"grad_norm": 0.73766028881073,
"learning_rate": 0.00015638890506329114,
"loss": 0.0118,
"step": 7400
},
{
"epoch": 18.939987365761212,
"grad_norm": 0.25607988238334656,
"learning_rate": 0.00015512371518987343,
"loss": 0.0124,
"step": 7500
},
{
"epoch": 19.0,
"eval_bleu": 13.418,
"eval_gen_len": 14.4207,
"eval_loss": 5.1243767738342285,
"eval_runtime": 534.6138,
"eval_samples_per_second": 11.0,
"eval_steps_per_second": 1.377,
"step": 7524
},
{
"epoch": 19.19204042956412,
"grad_norm": 0.19441089034080505,
"learning_rate": 0.00015385852531645572,
"loss": 0.0113,
"step": 7600
},
{
"epoch": 19.44472520530638,
"grad_norm": 0.41388899087905884,
"learning_rate": 0.00015259333544303798,
"loss": 0.0118,
"step": 7700
},
{
"epoch": 19.697409981048644,
"grad_norm": 0.20537462830543518,
"learning_rate": 0.00015132814556962027,
"loss": 0.0081,
"step": 7800
},
{
"epoch": 19.950094756790904,
"grad_norm": 0.07861676812171936,
"learning_rate": 0.00015006295569620253,
"loss": 0.0107,
"step": 7900
},
{
"epoch": 20.0,
"eval_bleu": 13.4141,
"eval_gen_len": 14.5943,
"eval_loss": 5.130436897277832,
"eval_runtime": 536.0796,
"eval_samples_per_second": 10.97,
"eval_steps_per_second": 1.373,
"step": 7920
},
{
"epoch": 20.202147820593808,
"grad_norm": 0.4762495756149292,
"learning_rate": 0.0001487977658227848,
"loss": 0.009,
"step": 8000
},
{
"epoch": 20.45483259633607,
"grad_norm": 1.3390281200408936,
"learning_rate": 0.0001475325759493671,
"loss": 0.0095,
"step": 8100
},
{
"epoch": 20.707517372078332,
"grad_norm": 0.06131729856133461,
"learning_rate": 0.0001462673860759494,
"loss": 0.0088,
"step": 8200
},
{
"epoch": 20.960202147820596,
"grad_norm": 0.10782083868980408,
"learning_rate": 0.00014500219620253166,
"loss": 0.0104,
"step": 8300
},
{
"epoch": 21.0,
"eval_bleu": 13.6054,
"eval_gen_len": 14.0954,
"eval_loss": 5.0841474533081055,
"eval_runtime": 523.9577,
"eval_samples_per_second": 11.224,
"eval_steps_per_second": 1.405,
"step": 8316
},
{
"epoch": 21.2122552116235,
"grad_norm": 0.5227333903312683,
"learning_rate": 0.00014373700632911395,
"loss": 0.011,
"step": 8400
},
{
"epoch": 21.46493998736576,
"grad_norm": 0.3831775486469269,
"learning_rate": 0.0001424718164556962,
"loss": 0.0087,
"step": 8500
},
{
"epoch": 21.717624763108024,
"grad_norm": 0.18052732944488525,
"learning_rate": 0.00014120662658227847,
"loss": 0.0091,
"step": 8600
},
{
"epoch": 21.970309538850284,
"grad_norm": 0.17256955802440643,
"learning_rate": 0.00013994143670886076,
"loss": 0.0121,
"step": 8700
},
{
"epoch": 22.0,
"eval_bleu": 13.4688,
"eval_gen_len": 14.6354,
"eval_loss": 5.096054553985596,
"eval_runtime": 540.1153,
"eval_samples_per_second": 10.888,
"eval_steps_per_second": 1.363,
"step": 8712
},
{
"epoch": 22.22236260265319,
"grad_norm": 0.847822904586792,
"learning_rate": 0.00013867624683544305,
"loss": 0.008,
"step": 8800
},
{
"epoch": 22.475047378395452,
"grad_norm": 0.04867393895983696,
"learning_rate": 0.00013741105696202534,
"loss": 0.0115,
"step": 8900
},
{
"epoch": 22.727732154137712,
"grad_norm": 0.24624444544315338,
"learning_rate": 0.0001361458670886076,
"loss": 0.0098,
"step": 9000
},
{
"epoch": 22.980416929879976,
"grad_norm": 0.6290240287780762,
"learning_rate": 0.0001348806772151899,
"loss": 0.0086,
"step": 9100
},
{
"epoch": 23.0,
"eval_bleu": 13.5374,
"eval_gen_len": 14.4979,
"eval_loss": 5.132958889007568,
"eval_runtime": 536.865,
"eval_samples_per_second": 10.954,
"eval_steps_per_second": 1.371,
"step": 9108
},
{
"epoch": 23.23246999368288,
"grad_norm": 0.19094829261302948,
"learning_rate": 0.00013361548734177215,
"loss": 0.0069,
"step": 9200
},
{
"epoch": 23.485154769425144,
"grad_norm": 0.031919267028570175,
"learning_rate": 0.0001323502974683544,
"loss": 0.0105,
"step": 9300
},
{
"epoch": 23.737839545167404,
"grad_norm": 0.07248106598854065,
"learning_rate": 0.00013108510759493673,
"loss": 0.0089,
"step": 9400
},
{
"epoch": 23.990524320909664,
"grad_norm": 0.059244658797979355,
"learning_rate": 0.00012981991772151902,
"loss": 0.0097,
"step": 9500
},
{
"epoch": 24.0,
"eval_bleu": 13.4956,
"eval_gen_len": 14.4816,
"eval_loss": 5.115517616271973,
"eval_runtime": 538.8426,
"eval_samples_per_second": 10.914,
"eval_steps_per_second": 1.366,
"step": 9504
},
{
"epoch": 24.242577384712572,
"grad_norm": 0.6686831116676331,
"learning_rate": 0.00012855472784810128,
"loss": 0.0075,
"step": 9600
},
{
"epoch": 24.495262160454832,
"grad_norm": 0.15613651275634766,
"learning_rate": 0.00012728953797468357,
"loss": 0.0071,
"step": 9700
},
{
"epoch": 24.747946936197096,
"grad_norm": 0.03811512514948845,
"learning_rate": 0.00012602434810126583,
"loss": 0.0062,
"step": 9800
},
{
"epoch": 25.0,
"grad_norm": 0.0894945040345192,
"learning_rate": 0.0001247591582278481,
"loss": 0.0074,
"step": 9900
},
{
"epoch": 25.0,
"eval_bleu": 13.8177,
"eval_gen_len": 14.3275,
"eval_loss": 5.174163341522217,
"eval_runtime": 531.7973,
"eval_samples_per_second": 11.059,
"eval_steps_per_second": 1.384,
"step": 9900
},
{
"epoch": 25.25268477574226,
"grad_norm": 0.20836782455444336,
"learning_rate": 0.00012349396835443038,
"loss": 0.0057,
"step": 10000
},
{
"epoch": 25.505369551484524,
"grad_norm": 0.5370838642120361,
"learning_rate": 0.00012222877848101267,
"loss": 0.0065,
"step": 10100
},
{
"epoch": 25.758054327226784,
"grad_norm": 0.21669617295265198,
"learning_rate": 0.00012096358860759496,
"loss": 0.0058,
"step": 10200
},
{
"epoch": 26.0,
"eval_bleu": 13.6641,
"eval_gen_len": 14.219,
"eval_loss": 5.147903919219971,
"eval_runtime": 528.0869,
"eval_samples_per_second": 11.136,
"eval_steps_per_second": 1.394,
"step": 10296
},
{
"epoch": 26.010107391029692,
"grad_norm": 0.2630160450935364,
"learning_rate": 0.00011969839873417723,
"loss": 0.006,
"step": 10300
},
{
"epoch": 26.262792166771952,
"grad_norm": 0.048449043184518814,
"learning_rate": 0.00011843320886075951,
"loss": 0.0086,
"step": 10400
},
{
"epoch": 26.515476942514212,
"grad_norm": 0.34334516525268555,
"learning_rate": 0.00011716801898734178,
"loss": 0.0058,
"step": 10500
},
{
"epoch": 26.768161718256476,
"grad_norm": 0.12147314101457596,
"learning_rate": 0.00011590282911392404,
"loss": 0.0058,
"step": 10600
},
{
"epoch": 27.0,
"eval_bleu": 13.7447,
"eval_gen_len": 14.1751,
"eval_loss": 5.193230628967285,
"eval_runtime": 528.8414,
"eval_samples_per_second": 11.121,
"eval_steps_per_second": 1.392,
"step": 10692
},
{
"epoch": 27.02021478205938,
"grad_norm": 0.29292619228363037,
"learning_rate": 0.00011463763924050632,
"loss": 0.0078,
"step": 10700
},
{
"epoch": 27.272899557801644,
"grad_norm": 0.5977900624275208,
"learning_rate": 0.00011337244936708862,
"loss": 0.0066,
"step": 10800
},
{
"epoch": 27.525584333543904,
"grad_norm": 0.8689119815826416,
"learning_rate": 0.00011210725949367088,
"loss": 0.0063,
"step": 10900
},
{
"epoch": 27.778269109286164,
"grad_norm": 0.28833064436912537,
"learning_rate": 0.00011084206962025316,
"loss": 0.0044,
"step": 11000
},
{
"epoch": 28.0,
"eval_bleu": 13.488,
"eval_gen_len": 14.7169,
"eval_loss": 5.161106109619141,
"eval_runtime": 537.3924,
"eval_samples_per_second": 10.944,
"eval_steps_per_second": 1.37,
"step": 11088
},
{
"epoch": 28.030322173089072,
"grad_norm": 0.7058536410331726,
"learning_rate": 0.00010957687974683546,
"loss": 0.0063,
"step": 11100
},
{
"epoch": 28.283006948831332,
"grad_norm": 0.12089253216981888,
"learning_rate": 0.00010831168987341774,
"loss": 0.0082,
"step": 11200
},
{
"epoch": 28.535691724573596,
"grad_norm": 0.6483332514762878,
"learning_rate": 0.00010704649999999999,
"loss": 0.0058,
"step": 11300
},
{
"epoch": 28.788376500315856,
"grad_norm": 0.10533745586872101,
"learning_rate": 0.00010578131012658227,
"loss": 0.0083,
"step": 11400
},
{
"epoch": 29.0,
"eval_bleu": 13.8153,
"eval_gen_len": 14.3556,
"eval_loss": 5.157731533050537,
"eval_runtime": 531.5203,
"eval_samples_per_second": 11.064,
"eval_steps_per_second": 1.385,
"step": 11484
},
{
"epoch": 29.04042956411876,
"grad_norm": 0.058924734592437744,
"learning_rate": 0.00010451612025316458,
"loss": 0.0067,
"step": 11500
},
{
"epoch": 29.293114339861024,
"grad_norm": 0.2670024633407593,
"learning_rate": 0.00010325093037974685,
"loss": 0.0052,
"step": 11600
},
{
"epoch": 29.545799115603284,
"grad_norm": 0.3689946234226227,
"learning_rate": 0.00010198574050632911,
"loss": 0.0067,
"step": 11700
},
{
"epoch": 29.798483891345548,
"grad_norm": 0.1571495085954666,
"learning_rate": 0.00010072055063291139,
"loss": 0.0053,
"step": 11800
},
{
"epoch": 30.0,
"eval_bleu": 14.1224,
"eval_gen_len": 14.1012,
"eval_loss": 5.206099033355713,
"eval_runtime": 526.5861,
"eval_samples_per_second": 11.168,
"eval_steps_per_second": 1.398,
"step": 11880
},
{
"epoch": 30.050536955148452,
"grad_norm": 0.02085699699819088,
"learning_rate": 9.945536075949369e-05,
"loss": 0.0044,
"step": 11900
},
{
"epoch": 30.303221730890712,
"grad_norm": 0.041063107550144196,
"learning_rate": 9.819017088607595e-05,
"loss": 0.005,
"step": 12000
},
{
"epoch": 30.555906506632976,
"grad_norm": 0.03547859191894531,
"learning_rate": 9.692498101265823e-05,
"loss": 0.0065,
"step": 12100
},
{
"epoch": 30.808591282375236,
"grad_norm": 0.03922571241855621,
"learning_rate": 9.56597911392405e-05,
"loss": 0.0046,
"step": 12200
},
{
"epoch": 31.0,
"eval_bleu": 13.9126,
"eval_gen_len": 14.5045,
"eval_loss": 5.24795389175415,
"eval_runtime": 530.0839,
"eval_samples_per_second": 11.094,
"eval_steps_per_second": 1.388,
"step": 12276
},
{
"epoch": 31.060644346178144,
"grad_norm": 0.1546187549829483,
"learning_rate": 9.439460126582279e-05,
"loss": 0.0036,
"step": 12300
},
{
"epoch": 31.313329121920404,
"grad_norm": 0.18324150145053864,
"learning_rate": 9.312941139240507e-05,
"loss": 0.0037,
"step": 12400
},
{
"epoch": 31.566013897662664,
"grad_norm": 0.019374554976820946,
"learning_rate": 9.186422151898734e-05,
"loss": 0.0054,
"step": 12500
},
{
"epoch": 31.818698673404928,
"grad_norm": 0.04015549644827843,
"learning_rate": 9.059903164556963e-05,
"loss": 0.0054,
"step": 12600
},
{
"epoch": 32.0,
"eval_bleu": 14.019,
"eval_gen_len": 14.16,
"eval_loss": 5.196505069732666,
"eval_runtime": 526.0182,
"eval_samples_per_second": 11.18,
"eval_steps_per_second": 1.399,
"step": 12672
},
{
"epoch": 32.070751737207836,
"grad_norm": 0.7235105633735657,
"learning_rate": 8.933384177215191e-05,
"loss": 0.0046,
"step": 12700
},
{
"epoch": 32.323436512950096,
"grad_norm": 0.01608334481716156,
"learning_rate": 8.806865189873417e-05,
"loss": 0.0039,
"step": 12800
},
{
"epoch": 32.576121288692356,
"grad_norm": 0.027920836582779884,
"learning_rate": 8.680346202531647e-05,
"loss": 0.0051,
"step": 12900
},
{
"epoch": 32.828806064434616,
"grad_norm": 0.9806521534919739,
"learning_rate": 8.553827215189875e-05,
"loss": 0.0035,
"step": 13000
},
{
"epoch": 33.0,
"eval_bleu": 14.004,
"eval_gen_len": 14.4037,
"eval_loss": 5.184691905975342,
"eval_runtime": 529.8109,
"eval_samples_per_second": 11.1,
"eval_steps_per_second": 1.389,
"step": 13068
},
{
"epoch": 33.08085912823752,
"grad_norm": 0.03555789962410927,
"learning_rate": 8.427308227848101e-05,
"loss": 0.004,
"step": 13100
},
{
"epoch": 33.33354390397979,
"grad_norm": 0.0131059680134058,
"learning_rate": 8.300789240506328e-05,
"loss": 0.0027,
"step": 13200
},
{
"epoch": 33.58622867972205,
"grad_norm": 0.08038444817066193,
"learning_rate": 8.175535443037976e-05,
"loss": 0.0036,
"step": 13300
},
{
"epoch": 33.83891345546431,
"grad_norm": 0.102862149477005,
"learning_rate": 8.049016455696203e-05,
"loss": 0.0032,
"step": 13400
},
{
"epoch": 34.0,
"eval_bleu": 14.228,
"eval_gen_len": 14.2273,
"eval_loss": 5.212389945983887,
"eval_runtime": 524.326,
"eval_samples_per_second": 11.216,
"eval_steps_per_second": 1.404,
"step": 13464
},
{
"epoch": 34.09096651926721,
"grad_norm": 0.011331531219184399,
"learning_rate": 7.922497468354431e-05,
"loss": 0.0041,
"step": 13500
},
{
"epoch": 34.34365129500947,
"grad_norm": 0.2522341012954712,
"learning_rate": 7.79597848101266e-05,
"loss": 0.0038,
"step": 13600
},
{
"epoch": 34.59633607075174,
"grad_norm": 0.0073499069549143314,
"learning_rate": 7.669459493670887e-05,
"loss": 0.0028,
"step": 13700
},
{
"epoch": 34.849020846494,
"grad_norm": 0.006374124903231859,
"learning_rate": 7.542940506329114e-05,
"loss": 0.0024,
"step": 13800
},
{
"epoch": 35.0,
"eval_bleu": 14.2703,
"eval_gen_len": 14.0995,
"eval_loss": 5.208999156951904,
"eval_runtime": 526.3437,
"eval_samples_per_second": 11.173,
"eval_steps_per_second": 1.398,
"step": 13860
},
{
"epoch": 35.101073910296904,
"grad_norm": 0.10820985585451126,
"learning_rate": 7.416421518987341e-05,
"loss": 0.0031,
"step": 13900
},
{
"epoch": 35.353758686039164,
"grad_norm": 0.011806854978203773,
"learning_rate": 7.289902531645571e-05,
"loss": 0.0035,
"step": 14000
},
{
"epoch": 35.606443461781424,
"grad_norm": 0.039563409984111786,
"learning_rate": 7.163383544303798e-05,
"loss": 0.0025,
"step": 14100
},
{
"epoch": 35.85912823752369,
"grad_norm": 0.03173103556036949,
"learning_rate": 7.036864556962025e-05,
"loss": 0.0029,
"step": 14200
},
{
"epoch": 36.0,
"eval_bleu": 13.7593,
"eval_gen_len": 14.604,
"eval_loss": 5.232673168182373,
"eval_runtime": 537.4765,
"eval_samples_per_second": 10.942,
"eval_steps_per_second": 1.369,
"step": 14256
},
{
"epoch": 36.111181301326596,
"grad_norm": 0.08217954635620117,
"learning_rate": 6.910345569620254e-05,
"loss": 0.003,
"step": 14300
},
{
"epoch": 36.363866077068856,
"grad_norm": 0.010925635695457458,
"learning_rate": 6.783826582278481e-05,
"loss": 0.0036,
"step": 14400
},
{
"epoch": 36.616550852811116,
"grad_norm": 0.6118758916854858,
"learning_rate": 6.657307594936709e-05,
"loss": 0.0039,
"step": 14500
},
{
"epoch": 36.869235628553376,
"grad_norm": 0.01771758496761322,
"learning_rate": 6.530788607594938e-05,
"loss": 0.0043,
"step": 14600
},
{
"epoch": 37.0,
"eval_bleu": 14.3019,
"eval_gen_len": 14.0886,
"eval_loss": 5.2005295753479,
"eval_runtime": 524.3574,
"eval_samples_per_second": 11.216,
"eval_steps_per_second": 1.404,
"step": 14652
},
{
"epoch": 37.12128869235629,
"grad_norm": 0.015916310250759125,
"learning_rate": 6.404269620253165e-05,
"loss": 0.0029,
"step": 14700
},
{
"epoch": 37.37397346809855,
"grad_norm": 0.022140830755233765,
"learning_rate": 6.277750632911393e-05,
"loss": 0.0034,
"step": 14800
},
{
"epoch": 37.62665824384081,
"grad_norm": 0.01691320165991783,
"learning_rate": 6.151231645569619e-05,
"loss": 0.003,
"step": 14900
},
{
"epoch": 37.87934301958307,
"grad_norm": 0.02250687964260578,
"learning_rate": 6.024712658227849e-05,
"loss": 0.0022,
"step": 15000
},
{
"epoch": 38.0,
"eval_bleu": 14.2565,
"eval_gen_len": 14.1928,
"eval_loss": 5.221782684326172,
"eval_runtime": 528.2072,
"eval_samples_per_second": 11.134,
"eval_steps_per_second": 1.393,
"step": 15048
},
{
"epoch": 38.13139608338598,
"grad_norm": 0.008933404460549355,
"learning_rate": 5.898193670886076e-05,
"loss": 0.0025,
"step": 15100
},
{
"epoch": 38.38408085912824,
"grad_norm": 0.008681390434503555,
"learning_rate": 5.771674683544304e-05,
"loss": 0.0023,
"step": 15200
},
{
"epoch": 38.6367656348705,
"grad_norm": 0.032701168209314346,
"learning_rate": 5.6451556962025333e-05,
"loss": 0.0021,
"step": 15300
},
{
"epoch": 38.88945041061276,
"grad_norm": 0.01919535920023918,
"learning_rate": 5.51863670886076e-05,
"loss": 0.0031,
"step": 15400
},
{
"epoch": 39.0,
"eval_bleu": 14.1208,
"eval_gen_len": 14.438,
"eval_loss": 5.240268707275391,
"eval_runtime": 534.3637,
"eval_samples_per_second": 11.006,
"eval_steps_per_second": 1.377,
"step": 15444
},
{
"epoch": 39.141503474415664,
"grad_norm": 0.07539849728345871,
"learning_rate": 5.3921177215189864e-05,
"loss": 0.0022,
"step": 15500
},
{
"epoch": 39.39418825015793,
"grad_norm": 0.00894533097743988,
"learning_rate": 5.265598734177216e-05,
"loss": 0.0021,
"step": 15600
},
{
"epoch": 39.64687302590019,
"grad_norm": 0.03846971318125725,
"learning_rate": 5.1390797468354435e-05,
"loss": 0.0019,
"step": 15700
},
{
"epoch": 39.89955780164245,
"grad_norm": 0.009916193783283234,
"learning_rate": 5.012560759493671e-05,
"loss": 0.0022,
"step": 15800
},
{
"epoch": 40.0,
"eval_bleu": 14.2927,
"eval_gen_len": 14.3079,
"eval_loss": 5.250655651092529,
"eval_runtime": 527.1014,
"eval_samples_per_second": 11.157,
"eval_steps_per_second": 1.396,
"step": 15840
},
{
"epoch": 40.151610865445356,
"grad_norm": 0.004169174004346132,
"learning_rate": 4.886041772151898e-05,
"loss": 0.0016,
"step": 15900
},
{
"epoch": 40.404295641187616,
"grad_norm": 0.019835354760289192,
"learning_rate": 4.7595227848101274e-05,
"loss": 0.0024,
"step": 16000
},
{
"epoch": 40.65698041692988,
"grad_norm": 0.004063699394464493,
"learning_rate": 4.633003797468354e-05,
"loss": 0.0027,
"step": 16100
},
{
"epoch": 40.90966519267214,
"grad_norm": 0.023647163063287735,
"learning_rate": 4.506484810126582e-05,
"loss": 0.0014,
"step": 16200
},
{
"epoch": 41.0,
"eval_bleu": 14.2727,
"eval_gen_len": 14.2874,
"eval_loss": 5.255828380584717,
"eval_runtime": 526.7142,
"eval_samples_per_second": 11.165,
"eval_steps_per_second": 1.397,
"step": 16236
},
{
"epoch": 41.16171825647505,
"grad_norm": 0.022712113335728645,
"learning_rate": 4.3799658227848114e-05,
"loss": 0.0016,
"step": 16300
},
{
"epoch": 41.41440303221731,
"grad_norm": 0.006393743213266134,
"learning_rate": 4.253446835443038e-05,
"loss": 0.0015,
"step": 16400
},
{
"epoch": 41.66708780795957,
"grad_norm": 0.020621536299586296,
"learning_rate": 4.126927848101266e-05,
"loss": 0.0019,
"step": 16500
},
{
"epoch": 41.919772583701835,
"grad_norm": 0.06859384477138519,
"learning_rate": 4.000408860759495e-05,
"loss": 0.0021,
"step": 16600
},
{
"epoch": 42.0,
"eval_bleu": 14.1117,
"eval_gen_len": 14.1115,
"eval_loss": 5.27353572845459,
"eval_runtime": 523.6053,
"eval_samples_per_second": 11.232,
"eval_steps_per_second": 1.406,
"step": 16632
},
{
"epoch": 42.17182564750474,
"grad_norm": 0.025820279493927956,
"learning_rate": 3.873889873417722e-05,
"loss": 0.0019,
"step": 16700
},
{
"epoch": 42.424510423247,
"grad_norm": 0.01880701631307602,
"learning_rate": 3.747370886075949e-05,
"loss": 0.0018,
"step": 16800
},
{
"epoch": 42.67719519898926,
"grad_norm": 0.08723778277635574,
"learning_rate": 3.6208518987341786e-05,
"loss": 0.0016,
"step": 16900
},
{
"epoch": 42.92987997473152,
"grad_norm": 0.05921424925327301,
"learning_rate": 3.4943329113924055e-05,
"loss": 0.0013,
"step": 17000
},
{
"epoch": 43.0,
"eval_bleu": 14.4166,
"eval_gen_len": 14.1923,
"eval_loss": 5.2706618309021,
"eval_runtime": 523.861,
"eval_samples_per_second": 11.226,
"eval_steps_per_second": 1.405,
"step": 17028
},
{
"epoch": 43.18193303853443,
"grad_norm": 0.0028839095029979944,
"learning_rate": 3.367813924050633e-05,
"loss": 0.0021,
"step": 17100
},
{
"epoch": 43.43461781427669,
"grad_norm": 0.008104875683784485,
"learning_rate": 3.2425601265822784e-05,
"loss": 0.0014,
"step": 17200
},
{
"epoch": 43.68730259001895,
"grad_norm": 0.043954506516456604,
"learning_rate": 3.116041139240508e-05,
"loss": 0.0013,
"step": 17300
},
{
"epoch": 43.93998736576121,
"grad_norm": 0.003915693145245314,
"learning_rate": 2.989522151898734e-05,
"loss": 0.0021,
"step": 17400
},
{
"epoch": 44.0,
"eval_bleu": 14.4223,
"eval_gen_len": 14.2129,
"eval_loss": 5.279012203216553,
"eval_runtime": 523.2065,
"eval_samples_per_second": 11.24,
"eval_steps_per_second": 1.407,
"step": 17424
},
{
"epoch": 44.192040429564116,
"grad_norm": 0.017741164192557335,
"learning_rate": 2.8630031645569617e-05,
"loss": 0.0015,
"step": 17500
},
{
"epoch": 44.44472520530638,
"grad_norm": 0.018855927512049675,
"learning_rate": 2.736484177215189e-05,
"loss": 0.0011,
"step": 17600
},
{
"epoch": 44.697409981048644,
"grad_norm": 0.005235583987087011,
"learning_rate": 2.6099651898734178e-05,
"loss": 0.0015,
"step": 17700
},
{
"epoch": 44.950094756790904,
"grad_norm": 0.009746776893734932,
"learning_rate": 2.4834462025316453e-05,
"loss": 0.0016,
"step": 17800
},
{
"epoch": 45.0,
"eval_bleu": 14.486,
"eval_gen_len": 14.2625,
"eval_loss": 5.275847911834717,
"eval_runtime": 521.3464,
"eval_samples_per_second": 11.28,
"eval_steps_per_second": 1.412,
"step": 17820
},
{
"epoch": 45.20214782059381,
"grad_norm": 0.016078555956482887,
"learning_rate": 2.3569272151898725e-05,
"loss": 0.0026,
"step": 17900
},
{
"epoch": 45.45483259633607,
"grad_norm": 0.043376993387937546,
"learning_rate": 2.230408227848102e-05,
"loss": 0.0012,
"step": 18000
},
{
"epoch": 45.707517372078335,
"grad_norm": 0.0051286788657307625,
"learning_rate": 2.1038892405063293e-05,
"loss": 0.001,
"step": 18100
},
{
"epoch": 45.960202147820596,
"grad_norm": 0.003953118342906237,
"learning_rate": 1.9773702531645565e-05,
"loss": 0.0019,
"step": 18200
},
{
"epoch": 46.0,
"eval_bleu": 14.5501,
"eval_gen_len": 14.2695,
"eval_loss": 5.25464391708374,
"eval_runtime": 524.8843,
"eval_samples_per_second": 11.204,
"eval_steps_per_second": 1.402,
"step": 18216
},
{
"epoch": 46.2122552116235,
"grad_norm": 0.013465896248817444,
"learning_rate": 1.8508512658227857e-05,
"loss": 0.0012,
"step": 18300
},
{
"epoch": 46.46493998736576,
"grad_norm": 0.3650831878185272,
"learning_rate": 1.724332278481013e-05,
"loss": 0.0012,
"step": 18400
},
{
"epoch": 46.71762476310802,
"grad_norm": 0.0039703696966171265,
"learning_rate": 1.59781329113924e-05,
"loss": 0.0014,
"step": 18500
},
{
"epoch": 46.97030953885029,
"grad_norm": 0.00649018120020628,
"learning_rate": 1.4712943037974674e-05,
"loss": 0.0011,
"step": 18600
},
{
"epoch": 47.0,
"eval_bleu": 14.6166,
"eval_gen_len": 14.1882,
"eval_loss": 5.265357494354248,
"eval_runtime": 516.9862,
"eval_samples_per_second": 11.376,
"eval_steps_per_second": 1.424,
"step": 18612
},
{
"epoch": 47.22236260265319,
"grad_norm": 0.013269159942865372,
"learning_rate": 1.3447753164556967e-05,
"loss": 0.0012,
"step": 18700
},
{
"epoch": 47.47504737839545,
"grad_norm": 0.005190219730138779,
"learning_rate": 1.2182563291139239e-05,
"loss": 0.0008,
"step": 18800
},
{
"epoch": 47.72773215413771,
"grad_norm": 0.03904829919338226,
"learning_rate": 1.091737341772151e-05,
"loss": 0.0012,
"step": 18900
},
{
"epoch": 47.98041692987997,
"grad_norm": 0.011222691275179386,
"learning_rate": 9.652183544303805e-06,
"loss": 0.0016,
"step": 19000
},
{
"epoch": 48.0,
"eval_bleu": 14.5838,
"eval_gen_len": 14.2617,
"eval_loss": 5.261044502258301,
"eval_runtime": 523.6855,
"eval_samples_per_second": 11.23,
"eval_steps_per_second": 1.405,
"step": 19008
},
{
"epoch": 48.232469993682884,
"grad_norm": 0.006541873328387737,
"learning_rate": 8.386993670886077e-06,
"loss": 0.0011,
"step": 19100
},
{
"epoch": 48.485154769425144,
"grad_norm": 0.026967084035277367,
"learning_rate": 7.121803797468348e-06,
"loss": 0.0006,
"step": 19200
},
{
"epoch": 48.737839545167404,
"grad_norm": 0.013220852240920067,
"learning_rate": 5.856613924050642e-06,
"loss": 0.0012,
"step": 19300
},
{
"epoch": 48.990524320909664,
"grad_norm": 0.01667727902531624,
"learning_rate": 4.591424050632914e-06,
"loss": 0.0011,
"step": 19400
},
{
"epoch": 49.0,
"eval_bleu": 14.5987,
"eval_gen_len": 14.2119,
"eval_loss": 5.264169216156006,
"eval_runtime": 516.2093,
"eval_samples_per_second": 11.393,
"eval_steps_per_second": 1.426,
"step": 19404
},
{
"epoch": 49.24257738471257,
"grad_norm": 0.009594439528882504,
"learning_rate": 3.3262341772151843e-06,
"loss": 0.0007,
"step": 19500
},
{
"epoch": 49.495262160454836,
"grad_norm": 0.015515263192355633,
"learning_rate": 2.0610443037974787e-06,
"loss": 0.0009,
"step": 19600
},
{
"epoch": 49.747946936197096,
"grad_norm": 0.34153661131858826,
"learning_rate": 7.958544303797504e-07,
"loss": 0.001,
"step": 19700
},
{
"epoch": 49.87428932406822,
"eval_bleu": 14.576,
"eval_gen_len": 14.2289,
"eval_loss": 5.264520168304443,
"eval_runtime": 519.5989,
"eval_samples_per_second": 11.318,
"eval_steps_per_second": 1.416,
"step": 19750
},
{
"epoch": 49.87428932406822,
"step": 19750,
"total_flos": 6.840351903574917e+17,
"train_loss": 0.00785975846080581,
"train_runtime": 68802.4337,
"train_samples_per_second": 18.397,
"train_steps_per_second": 0.287
}
],
"logging_steps": 100,
"max_steps": 19750,
"num_input_tokens_seen": 0,
"num_train_epochs": 50,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 6.840351903574917e+17,
"train_batch_size": 16,
"trial_name": null,
"trial_params": null
}