| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.9023136246786633, | |
| "eval_steps": 26, | |
| "global_step": 468, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.005784061696658098, | |
| "grad_norm": NaN, | |
| "learning_rate": 0.0, | |
| "loss": 7.8503, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.011568123393316195, | |
| "grad_norm": 127.71849060058594, | |
| "learning_rate": 5.791505791505792e-08, | |
| "loss": 8.4022, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.017352185089974295, | |
| "grad_norm": 244.984619140625, | |
| "learning_rate": 1.7374517374517373e-07, | |
| "loss": 11.1776, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.02313624678663239, | |
| "grad_norm": 425.6427917480469, | |
| "learning_rate": 3.4749034749034746e-07, | |
| "loss": 9.7845, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.02892030848329049, | |
| "grad_norm": 238.8634490966797, | |
| "learning_rate": 5.212355212355212e-07, | |
| "loss": 8.9224, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.03470437017994859, | |
| "grad_norm": 173.4210968017578, | |
| "learning_rate": 6.949806949806949e-07, | |
| "loss": 11.1202, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.04048843187660668, | |
| "grad_norm": 77.89308166503906, | |
| "learning_rate": 8.687258687258688e-07, | |
| "loss": 7.413, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.04627249357326478, | |
| "grad_norm": Infinity, | |
| "learning_rate": 9.845559845559844e-07, | |
| "loss": 7.7803, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.05012853470437018, | |
| "eval_Qnli-dev_cosine_accuracy": 0.609375, | |
| "eval_Qnli-dev_cosine_accuracy_threshold": 0.9602330923080444, | |
| "eval_Qnli-dev_cosine_ap": 0.587737015984062, | |
| "eval_Qnli-dev_cosine_f1": 0.6358695652173914, | |
| "eval_Qnli-dev_cosine_f1_threshold": 0.7741116285324097, | |
| "eval_Qnli-dev_cosine_precision": 0.468, | |
| "eval_Qnli-dev_cosine_recall": 0.9915254237288136, | |
| "eval_Qnli-dev_dot_accuracy": 0.56640625, | |
| "eval_Qnli-dev_dot_accuracy_threshold": 750.123779296875, | |
| "eval_Qnli-dev_dot_ap": 0.5198069471819926, | |
| "eval_Qnli-dev_dot_f1": 0.6318607764390897, | |
| "eval_Qnli-dev_dot_f1_threshold": 387.37371826171875, | |
| "eval_Qnli-dev_dot_precision": 0.461839530332681, | |
| "eval_Qnli-dev_dot_recall": 1.0, | |
| "eval_Qnli-dev_euclidean_accuracy": 0.6171875, | |
| "eval_Qnli-dev_euclidean_accuracy_threshold": 7.253280162811279, | |
| "eval_Qnli-dev_euclidean_ap": 0.5993562103857255, | |
| "eval_Qnli-dev_euclidean_f1": 0.6376021798365122, | |
| "eval_Qnli-dev_euclidean_f1_threshold": 17.42431640625, | |
| "eval_Qnli-dev_euclidean_precision": 0.46987951807228917, | |
| "eval_Qnli-dev_euclidean_recall": 0.9915254237288136, | |
| "eval_Qnli-dev_manhattan_accuracy": 0.625, | |
| "eval_Qnli-dev_manhattan_accuracy_threshold": 212.0604705810547, | |
| "eval_Qnli-dev_manhattan_ap": 0.609409682371795, | |
| "eval_Qnli-dev_manhattan_f1": 0.6358695652173914, | |
| "eval_Qnli-dev_manhattan_f1_threshold": 461.66192626953125, | |
| "eval_Qnli-dev_manhattan_precision": 0.468, | |
| "eval_Qnli-dev_manhattan_recall": 0.9915254237288136, | |
| "eval_Qnli-dev_max_accuracy": 0.625, | |
| "eval_Qnli-dev_max_accuracy_threshold": 750.123779296875, | |
| "eval_Qnli-dev_max_ap": 0.609409682371795, | |
| "eval_Qnli-dev_max_f1": 0.6376021798365122, | |
| "eval_Qnli-dev_max_f1_threshold": 461.66192626953125, | |
| "eval_Qnli-dev_max_precision": 0.46987951807228917, | |
| "eval_Qnli-dev_max_recall": 1.0, | |
| "eval_allNLI-dev_cosine_accuracy": 0.666015625, | |
| "eval_allNLI-dev_cosine_accuracy_threshold": 0.9931652545928955, | |
| "eval_allNLI-dev_cosine_ap": 0.3330277035108763, | |
| "eval_allNLI-dev_cosine_f1": 0.5088235294117648, | |
| "eval_allNLI-dev_cosine_f1_threshold": 0.40660202503204346, | |
| "eval_allNLI-dev_cosine_precision": 0.34122287968441817, | |
| "eval_allNLI-dev_cosine_recall": 1.0, | |
| "eval_allNLI-dev_dot_accuracy": 0.662109375, | |
| "eval_allNLI-dev_dot_accuracy_threshold": 810.7718505859375, | |
| "eval_allNLI-dev_dot_ap": 0.2980492563498839, | |
| "eval_allNLI-dev_dot_f1": 0.5088235294117648, | |
| "eval_allNLI-dev_dot_f1_threshold": 294.67071533203125, | |
| "eval_allNLI-dev_dot_precision": 0.34122287968441817, | |
| "eval_allNLI-dev_dot_recall": 1.0, | |
| "eval_allNLI-dev_euclidean_accuracy": 0.666015625, | |
| "eval_allNLI-dev_euclidean_accuracy_threshold": 3.3851852416992188, | |
| "eval_allNLI-dev_euclidean_ap": 0.33790858355605125, | |
| "eval_allNLI-dev_euclidean_f1": 0.5095729013254787, | |
| "eval_allNLI-dev_euclidean_f1_threshold": 28.78042984008789, | |
| "eval_allNLI-dev_euclidean_precision": 0.34189723320158105, | |
| "eval_allNLI-dev_euclidean_recall": 1.0, | |
| "eval_allNLI-dev_manhattan_accuracy": 0.666015625, | |
| "eval_allNLI-dev_manhattan_accuracy_threshold": 98.2369613647461, | |
| "eval_allNLI-dev_manhattan_ap": 0.3452833632811151, | |
| "eval_allNLI-dev_manhattan_f1": 0.5091463414634146, | |
| "eval_allNLI-dev_manhattan_f1_threshold": 626.5145874023438, | |
| "eval_allNLI-dev_manhattan_precision": 0.34575569358178054, | |
| "eval_allNLI-dev_manhattan_recall": 0.9653179190751445, | |
| "eval_allNLI-dev_max_accuracy": 0.666015625, | |
| "eval_allNLI-dev_max_accuracy_threshold": 810.7718505859375, | |
| "eval_allNLI-dev_max_ap": 0.3452833632811151, | |
| "eval_allNLI-dev_max_f1": 0.5095729013254787, | |
| "eval_allNLI-dev_max_f1_threshold": 626.5145874023438, | |
| "eval_allNLI-dev_max_precision": 0.34575569358178054, | |
| "eval_allNLI-dev_max_recall": 1.0, | |
| "eval_sequential_score": 0.609409682371795, | |
| "eval_sts-test_pearson_cosine": 0.24255956652335486, | |
| "eval_sts-test_pearson_dot": 0.18022025813169615, | |
| "eval_sts-test_pearson_euclidean": 0.28657386934462287, | |
| "eval_sts-test_pearson_manhattan": 0.32587411772454034, | |
| "eval_sts-test_pearson_max": 0.32587411772454034, | |
| "eval_sts-test_spearman_cosine": 0.32225123770630715, | |
| "eval_sts-test_spearman_dot": 0.17184726031108927, | |
| "eval_sts-test_spearman_euclidean": 0.3279675400801376, | |
| "eval_sts-test_spearman_manhattan": 0.3635547276663291, | |
| "eval_sts-test_spearman_max": 0.3635547276663291, | |
| "eval_vitaminc-pairs_loss": 4.122440338134766, | |
| "eval_vitaminc-pairs_runtime": 4.7337, | |
| "eval_vitaminc-pairs_samples_per_second": 27.04, | |
| "eval_vitaminc-pairs_steps_per_second": 0.211, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.05012853470437018, | |
| "eval_negation-triplets_loss": 3.917590379714966, | |
| "eval_negation-triplets_runtime": 3.3679, | |
| "eval_negation-triplets_samples_per_second": 38.006, | |
| "eval_negation-triplets_steps_per_second": 0.297, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.05012853470437018, | |
| "eval_scitail-pairs-pos_loss": 2.5062460899353027, | |
| "eval_scitail-pairs-pos_runtime": 2.7274, | |
| "eval_scitail-pairs-pos_samples_per_second": 46.931, | |
| "eval_scitail-pairs-pos_steps_per_second": 0.367, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.05012853470437018, | |
| "eval_scitail-pairs-qa_loss": 2.978612184524536, | |
| "eval_scitail-pairs-qa_runtime": 2.2452, | |
| "eval_scitail-pairs-qa_samples_per_second": 57.011, | |
| "eval_scitail-pairs-qa_steps_per_second": 0.445, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.05012853470437018, | |
| "eval_xsum-pairs_loss": 5.394700050354004, | |
| "eval_xsum-pairs_runtime": 3.0886, | |
| "eval_xsum-pairs_samples_per_second": 41.442, | |
| "eval_xsum-pairs_steps_per_second": 0.324, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.05012853470437018, | |
| "eval_sciq_pairs_loss": 0.8977569937705994, | |
| "eval_sciq_pairs_runtime": 4.7516, | |
| "eval_sciq_pairs_samples_per_second": 26.938, | |
| "eval_sciq_pairs_steps_per_second": 0.21, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.05012853470437018, | |
| "eval_qasc_pairs_loss": 6.6808977127075195, | |
| "eval_qasc_pairs_runtime": 2.1087, | |
| "eval_qasc_pairs_samples_per_second": 60.701, | |
| "eval_qasc_pairs_steps_per_second": 0.474, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.05012853470437018, | |
| "eval_openbookqa_pairs_loss": 6.402599334716797, | |
| "eval_openbookqa_pairs_runtime": 2.2222, | |
| "eval_openbookqa_pairs_samples_per_second": 57.601, | |
| "eval_openbookqa_pairs_steps_per_second": 0.45, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.05012853470437018, | |
| "eval_msmarco_pairs_loss": 15.782674789428711, | |
| "eval_msmarco_pairs_runtime": 2.307, | |
| "eval_msmarco_pairs_samples_per_second": 55.483, | |
| "eval_msmarco_pairs_steps_per_second": 0.433, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.05012853470437018, | |
| "eval_nq_pairs_loss": 10.447425842285156, | |
| "eval_nq_pairs_runtime": 3.6239, | |
| "eval_nq_pairs_samples_per_second": 35.321, | |
| "eval_nq_pairs_steps_per_second": 0.276, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.05012853470437018, | |
| "eval_trivia_pairs_loss": 6.107333183288574, | |
| "eval_trivia_pairs_runtime": 3.1299, | |
| "eval_trivia_pairs_samples_per_second": 40.895, | |
| "eval_trivia_pairs_steps_per_second": 0.319, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.05012853470437018, | |
| "eval_gooaq_pairs_loss": 8.095273971557617, | |
| "eval_gooaq_pairs_runtime": 2.1098, | |
| "eval_gooaq_pairs_samples_per_second": 60.669, | |
| "eval_gooaq_pairs_steps_per_second": 0.474, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.05012853470437018, | |
| "eval_paws-pos_loss": 1.6077361106872559, | |
| "eval_paws-pos_runtime": 2.2412, | |
| "eval_paws-pos_samples_per_second": 57.112, | |
| "eval_paws-pos_steps_per_second": 0.446, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.05012853470437018, | |
| "eval_global_dataset_loss": 6.348417282104492, | |
| "eval_global_dataset_runtime": 8.3683, | |
| "eval_global_dataset_samples_per_second": 38.837, | |
| "eval_global_dataset_steps_per_second": 0.358, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.05205655526992288, | |
| "grad_norm": 172.1571807861328, | |
| "learning_rate": 1.1583011583011583e-06, | |
| "loss": 7.9729, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.05784061696658098, | |
| "grad_norm": 106.302001953125, | |
| "learning_rate": 1.3320463320463321e-06, | |
| "loss": 6.0587, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.06362467866323908, | |
| "grad_norm": 72.31057739257812, | |
| "learning_rate": 1.5057915057915057e-06, | |
| "loss": 5.6742, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.06940874035989718, | |
| "grad_norm": 258.2577209472656, | |
| "learning_rate": 1.6795366795366796e-06, | |
| "loss": 6.5406, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.07519280205655526, | |
| "grad_norm": 156.0019073486328, | |
| "learning_rate": 1.7953667953667953e-06, | |
| "loss": 5.4429, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.08097686375321336, | |
| "grad_norm": 227.6062469482422, | |
| "learning_rate": 1.969111969111969e-06, | |
| "loss": 6.7855, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.08676092544987146, | |
| "grad_norm": 221.12347412109375, | |
| "learning_rate": 2.1428571428571427e-06, | |
| "loss": 5.3403, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.09254498714652956, | |
| "grad_norm": 363.6319274902344, | |
| "learning_rate": 2.3166023166023166e-06, | |
| "loss": 4.2282, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.09832904884318766, | |
| "grad_norm": 119.22310638427734, | |
| "learning_rate": 2.4903474903474904e-06, | |
| "loss": 4.7411, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.10025706940874037, | |
| "eval_Qnli-dev_cosine_accuracy": 0.58984375, | |
| "eval_Qnli-dev_cosine_accuracy_threshold": 0.9407284259796143, | |
| "eval_Qnli-dev_cosine_ap": 0.5957496162735323, | |
| "eval_Qnli-dev_cosine_f1": 0.6314363143631436, | |
| "eval_Qnli-dev_cosine_f1_threshold": 0.7475543022155762, | |
| "eval_Qnli-dev_cosine_precision": 0.4641434262948207, | |
| "eval_Qnli-dev_cosine_recall": 0.9872881355932204, | |
| "eval_Qnli-dev_dot_accuracy": 0.591796875, | |
| "eval_Qnli-dev_dot_accuracy_threshold": 787.171630859375, | |
| "eval_Qnli-dev_dot_ap": 0.556540262374452, | |
| "eval_Qnli-dev_dot_f1": 0.6328767123287672, | |
| "eval_Qnli-dev_dot_f1_threshold": 576.0565795898438, | |
| "eval_Qnli-dev_dot_precision": 0.4676113360323887, | |
| "eval_Qnli-dev_dot_recall": 0.9788135593220338, | |
| "eval_Qnli-dev_euclidean_accuracy": 0.587890625, | |
| "eval_Qnli-dev_euclidean_accuracy_threshold": 9.92269515991211, | |
| "eval_Qnli-dev_euclidean_ap": 0.5913529071853763, | |
| "eval_Qnli-dev_euclidean_f1": 0.6322930800542741, | |
| "eval_Qnli-dev_euclidean_f1_threshold": 19.546676635742188, | |
| "eval_Qnli-dev_euclidean_precision": 0.46506986027944114, | |
| "eval_Qnli-dev_euclidean_recall": 0.9872881355932204, | |
| "eval_Qnli-dev_manhattan_accuracy": 0.591796875, | |
| "eval_Qnli-dev_manhattan_accuracy_threshold": 314.9248352050781, | |
| "eval_Qnli-dev_manhattan_ap": 0.5944730102616272, | |
| "eval_Qnli-dev_manhattan_f1": 0.632952691680261, | |
| "eval_Qnli-dev_manhattan_f1_threshold": 424.9676818847656, | |
| "eval_Qnli-dev_manhattan_precision": 0.5145888594164456, | |
| "eval_Qnli-dev_manhattan_recall": 0.8220338983050848, | |
| "eval_Qnli-dev_max_accuracy": 0.591796875, | |
| "eval_Qnli-dev_max_accuracy_threshold": 787.171630859375, | |
| "eval_Qnli-dev_max_ap": 0.5957496162735323, | |
| "eval_Qnli-dev_max_f1": 0.632952691680261, | |
| "eval_Qnli-dev_max_f1_threshold": 576.0565795898438, | |
| "eval_Qnli-dev_max_precision": 0.5145888594164456, | |
| "eval_Qnli-dev_max_recall": 0.9872881355932204, | |
| "eval_allNLI-dev_cosine_accuracy": 0.666015625, | |
| "eval_allNLI-dev_cosine_accuracy_threshold": 0.9878466129302979, | |
| "eval_allNLI-dev_cosine_ap": 0.3898337405414008, | |
| "eval_allNLI-dev_cosine_f1": 0.5065885797950219, | |
| "eval_allNLI-dev_cosine_f1_threshold": 0.6534674167633057, | |
| "eval_allNLI-dev_cosine_precision": 0.3392156862745098, | |
| "eval_allNLI-dev_cosine_recall": 1.0, | |
| "eval_allNLI-dev_dot_accuracy": 0.666015625, | |
| "eval_allNLI-dev_dot_accuracy_threshold": 925.5989990234375, | |
| "eval_allNLI-dev_dot_ap": 0.3391870470132937, | |
| "eval_allNLI-dev_dot_f1": 0.5046153846153847, | |
| "eval_allNLI-dev_dot_f1_threshold": 608.7012939453125, | |
| "eval_allNLI-dev_dot_precision": 0.3438155136268344, | |
| "eval_allNLI-dev_dot_recall": 0.9479768786127167, | |
| "eval_allNLI-dev_euclidean_accuracy": 0.666015625, | |
| "eval_allNLI-dev_euclidean_accuracy_threshold": 4.708977699279785, | |
| "eval_allNLI-dev_euclidean_ap": 0.39556533875401867, | |
| "eval_allNLI-dev_euclidean_f1": 0.5080763582966226, | |
| "eval_allNLI-dev_euclidean_f1_threshold": 23.160057067871094, | |
| "eval_allNLI-dev_euclidean_precision": 0.3405511811023622, | |
| "eval_allNLI-dev_euclidean_recall": 1.0, | |
| "eval_allNLI-dev_manhattan_accuracy": 0.666015625, | |
| "eval_allNLI-dev_manhattan_accuracy_threshold": 142.7260284423828, | |
| "eval_allNLI-dev_manhattan_ap": 0.4050661514301679, | |
| "eval_allNLI-dev_manhattan_f1": 0.5118483412322274, | |
| "eval_allNLI-dev_manhattan_f1_threshold": 504.7154846191406, | |
| "eval_allNLI-dev_manhattan_precision": 0.3521739130434783, | |
| "eval_allNLI-dev_manhattan_recall": 0.9364161849710982, | |
| "eval_allNLI-dev_max_accuracy": 0.666015625, | |
| "eval_allNLI-dev_max_accuracy_threshold": 925.5989990234375, | |
| "eval_allNLI-dev_max_ap": 0.4050661514301679, | |
| "eval_allNLI-dev_max_f1": 0.5118483412322274, | |
| "eval_allNLI-dev_max_f1_threshold": 608.7012939453125, | |
| "eval_allNLI-dev_max_precision": 0.3521739130434783, | |
| "eval_allNLI-dev_max_recall": 1.0, | |
| "eval_sequential_score": 0.5957496162735323, | |
| "eval_sts-test_pearson_cosine": 0.5067228548513025, | |
| "eval_sts-test_pearson_dot": 0.28867418819353924, | |
| "eval_sts-test_pearson_euclidean": 0.5315956249457565, | |
| "eval_sts-test_pearson_manhattan": 0.5620907203006935, | |
| "eval_sts-test_pearson_max": 0.5620907203006935, | |
| "eval_sts-test_spearman_cosine": 0.5257409961657202, | |
| "eval_sts-test_spearman_dot": 0.280153994622986, | |
| "eval_sts-test_spearman_euclidean": 0.5306475244646732, | |
| "eval_sts-test_spearman_manhattan": 0.5605804796765843, | |
| "eval_sts-test_spearman_max": 0.5605804796765843, | |
| "eval_vitaminc-pairs_loss": 3.929797887802124, | |
| "eval_vitaminc-pairs_runtime": 4.6018, | |
| "eval_vitaminc-pairs_samples_per_second": 27.815, | |
| "eval_vitaminc-pairs_steps_per_second": 0.217, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.10025706940874037, | |
| "eval_negation-triplets_loss": 3.267047166824341, | |
| "eval_negation-triplets_runtime": 3.2813, | |
| "eval_negation-triplets_samples_per_second": 39.009, | |
| "eval_negation-triplets_steps_per_second": 0.305, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.10025706940874037, | |
| "eval_scitail-pairs-pos_loss": 1.134178876876831, | |
| "eval_scitail-pairs-pos_runtime": 2.6261, | |
| "eval_scitail-pairs-pos_samples_per_second": 48.742, | |
| "eval_scitail-pairs-pos_steps_per_second": 0.381, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.10025706940874037, | |
| "eval_scitail-pairs-qa_loss": 2.0085601806640625, | |
| "eval_scitail-pairs-qa_runtime": 2.2161, | |
| "eval_scitail-pairs-qa_samples_per_second": 57.759, | |
| "eval_scitail-pairs-qa_steps_per_second": 0.451, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.10025706940874037, | |
| "eval_xsum-pairs_loss": 4.462350845336914, | |
| "eval_xsum-pairs_runtime": 3.1126, | |
| "eval_xsum-pairs_samples_per_second": 41.124, | |
| "eval_xsum-pairs_steps_per_second": 0.321, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.10025706940874037, | |
| "eval_sciq_pairs_loss": 0.71834796667099, | |
| "eval_sciq_pairs_runtime": 4.7505, | |
| "eval_sciq_pairs_samples_per_second": 26.944, | |
| "eval_sciq_pairs_steps_per_second": 0.211, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.10025706940874037, | |
| "eval_qasc_pairs_loss": 4.889453887939453, | |
| "eval_qasc_pairs_runtime": 2.1101, | |
| "eval_qasc_pairs_samples_per_second": 60.662, | |
| "eval_qasc_pairs_steps_per_second": 0.474, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.10025706940874037, | |
| "eval_openbookqa_pairs_loss": 3.4273369312286377, | |
| "eval_openbookqa_pairs_runtime": 2.2311, | |
| "eval_openbookqa_pairs_samples_per_second": 57.371, | |
| "eval_openbookqa_pairs_steps_per_second": 0.448, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.10025706940874037, | |
| "eval_msmarco_pairs_loss": 7.532540321350098, | |
| "eval_msmarco_pairs_runtime": 2.2959, | |
| "eval_msmarco_pairs_samples_per_second": 55.751, | |
| "eval_msmarco_pairs_steps_per_second": 0.436, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.10025706940874037, | |
| "eval_nq_pairs_loss": 6.29541015625, | |
| "eval_nq_pairs_runtime": 3.661, | |
| "eval_nq_pairs_samples_per_second": 34.963, | |
| "eval_nq_pairs_steps_per_second": 0.273, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.10025706940874037, | |
| "eval_trivia_pairs_loss": 5.789977073669434, | |
| "eval_trivia_pairs_runtime": 3.0826, | |
| "eval_trivia_pairs_samples_per_second": 41.523, | |
| "eval_trivia_pairs_steps_per_second": 0.324, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.10025706940874037, | |
| "eval_gooaq_pairs_loss": 5.785163879394531, | |
| "eval_gooaq_pairs_runtime": 2.1382, | |
| "eval_gooaq_pairs_samples_per_second": 59.863, | |
| "eval_gooaq_pairs_steps_per_second": 0.468, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.10025706940874037, | |
| "eval_paws-pos_loss": 0.47830095887184143, | |
| "eval_paws-pos_runtime": 2.3113, | |
| "eval_paws-pos_samples_per_second": 55.379, | |
| "eval_paws-pos_steps_per_second": 0.433, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.10025706940874037, | |
| "eval_global_dataset_loss": 3.8913991451263428, | |
| "eval_global_dataset_runtime": 8.3623, | |
| "eval_global_dataset_samples_per_second": 38.865, | |
| "eval_global_dataset_steps_per_second": 0.359, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.10411311053984576, | |
| "grad_norm": 55.83478546142578, | |
| "learning_rate": 2.6640926640926642e-06, | |
| "loss": 3.9082, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.10989717223650386, | |
| "grad_norm": 57.4188117980957, | |
| "learning_rate": 2.7799227799227797e-06, | |
| "loss": 4.3922, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.11568123393316196, | |
| "grad_norm": 116.89701843261719, | |
| "learning_rate": 2.9536679536679535e-06, | |
| "loss": 3.2655, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.12146529562982006, | |
| "grad_norm": 53.49580001831055, | |
| "learning_rate": 3.1274131274131274e-06, | |
| "loss": 3.1043, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.12724935732647816, | |
| "grad_norm": 51.044410705566406, | |
| "learning_rate": 3.3011583011583012e-06, | |
| "loss": 2.2074, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.13303341902313626, | |
| "grad_norm": 139.9288787841797, | |
| "learning_rate": 3.474903474903475e-06, | |
| "loss": 1.4414, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.13881748071979436, | |
| "grad_norm": 37.221168518066406, | |
| "learning_rate": 3.5907335907335905e-06, | |
| "loss": 1.5937, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.14460154241645246, | |
| "grad_norm": 25.855661392211914, | |
| "learning_rate": 3.7644787644787644e-06, | |
| "loss": 1.0306, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.15038560411311053, | |
| "grad_norm": 29.79730796813965, | |
| "learning_rate": 3.938223938223938e-06, | |
| "loss": 1.0784, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.15038560411311053, | |
| "eval_Qnli-dev_cosine_accuracy": 0.638671875, | |
| "eval_Qnli-dev_cosine_accuracy_threshold": 0.856202244758606, | |
| "eval_Qnli-dev_cosine_ap": 0.6292803059732273, | |
| "eval_Qnli-dev_cosine_f1": 0.6408839779005525, | |
| "eval_Qnli-dev_cosine_f1_threshold": 0.6206527352333069, | |
| "eval_Qnli-dev_cosine_precision": 0.47540983606557374, | |
| "eval_Qnli-dev_cosine_recall": 0.9830508474576272, | |
| "eval_Qnli-dev_dot_accuracy": 0.626953125, | |
| "eval_Qnli-dev_dot_accuracy_threshold": 887.830810546875, | |
| "eval_Qnli-dev_dot_ap": 0.592075047542057, | |
| "eval_Qnli-dev_dot_f1": 0.640990371389271, | |
| "eval_Qnli-dev_dot_f1_threshold": 601.355224609375, | |
| "eval_Qnli-dev_dot_precision": 0.4745417515274949, | |
| "eval_Qnli-dev_dot_recall": 0.9872881355932204, | |
| "eval_Qnli-dev_euclidean_accuracy": 0.638671875, | |
| "eval_Qnli-dev_euclidean_accuracy_threshold": 17.094371795654297, | |
| "eval_Qnli-dev_euclidean_ap": 0.6289192660680781, | |
| "eval_Qnli-dev_euclidean_f1": 0.6423357664233577, | |
| "eval_Qnli-dev_euclidean_f1_threshold": 24.894683837890625, | |
| "eval_Qnli-dev_euclidean_precision": 0.48997772828507796, | |
| "eval_Qnli-dev_euclidean_recall": 0.9322033898305084, | |
| "eval_Qnli-dev_manhattan_accuracy": 0.640625, | |
| "eval_Qnli-dev_manhattan_accuracy_threshold": 526.0111083984375, | |
| "eval_Qnli-dev_manhattan_ap": 0.6256143574081243, | |
| "eval_Qnli-dev_manhattan_f1": 0.6417704011065007, | |
| "eval_Qnli-dev_manhattan_f1_threshold": 853.242431640625, | |
| "eval_Qnli-dev_manhattan_precision": 0.47638603696098564, | |
| "eval_Qnli-dev_manhattan_recall": 0.9830508474576272, | |
| "eval_Qnli-dev_max_accuracy": 0.640625, | |
| "eval_Qnli-dev_max_accuracy_threshold": 887.830810546875, | |
| "eval_Qnli-dev_max_ap": 0.6292803059732273, | |
| "eval_Qnli-dev_max_f1": 0.6423357664233577, | |
| "eval_Qnli-dev_max_f1_threshold": 853.242431640625, | |
| "eval_Qnli-dev_max_precision": 0.48997772828507796, | |
| "eval_Qnli-dev_max_recall": 0.9872881355932204, | |
| "eval_allNLI-dev_cosine_accuracy": 0.6796875, | |
| "eval_allNLI-dev_cosine_accuracy_threshold": 0.9426460862159729, | |
| "eval_allNLI-dev_cosine_ap": 0.473508692504868, | |
| "eval_allNLI-dev_cosine_f1": 0.5328947368421053, | |
| "eval_allNLI-dev_cosine_f1_threshold": 0.6884603500366211, | |
| "eval_allNLI-dev_cosine_precision": 0.3724137931034483, | |
| "eval_allNLI-dev_cosine_recall": 0.9364161849710982, | |
| "eval_allNLI-dev_dot_accuracy": 0.673828125, | |
| "eval_allNLI-dev_dot_accuracy_threshold": 980.9593505859375, | |
| "eval_allNLI-dev_dot_ap": 0.44188020452374555, | |
| "eval_allNLI-dev_dot_f1": 0.53156146179402, | |
| "eval_allNLI-dev_dot_f1_threshold": 683.7774658203125, | |
| "eval_allNLI-dev_dot_precision": 0.372960372960373, | |
| "eval_allNLI-dev_dot_recall": 0.9248554913294798, | |
| "eval_allNLI-dev_euclidean_accuracy": 0.6796875, | |
| "eval_allNLI-dev_euclidean_accuracy_threshold": 10.647405624389648, | |
| "eval_allNLI-dev_euclidean_ap": 0.47440611686548606, | |
| "eval_allNLI-dev_euclidean_f1": 0.5292207792207793, | |
| "eval_allNLI-dev_euclidean_f1_threshold": 25.08370590209961, | |
| "eval_allNLI-dev_euclidean_precision": 0.36794582392776526, | |
| "eval_allNLI-dev_euclidean_recall": 0.9421965317919075, | |
| "eval_allNLI-dev_manhattan_accuracy": 0.6796875, | |
| "eval_allNLI-dev_manhattan_accuracy_threshold": 318.408203125, | |
| "eval_allNLI-dev_manhattan_ap": 0.4747137405292771, | |
| "eval_allNLI-dev_manhattan_f1": 0.5305343511450381, | |
| "eval_allNLI-dev_manhattan_f1_threshold": 697.8466796875, | |
| "eval_allNLI-dev_manhattan_precision": 0.396011396011396, | |
| "eval_allNLI-dev_manhattan_recall": 0.8034682080924855, | |
| "eval_allNLI-dev_max_accuracy": 0.6796875, | |
| "eval_allNLI-dev_max_accuracy_threshold": 980.9593505859375, | |
| "eval_allNLI-dev_max_ap": 0.4747137405292771, | |
| "eval_allNLI-dev_max_f1": 0.5328947368421053, | |
| "eval_allNLI-dev_max_f1_threshold": 697.8466796875, | |
| "eval_allNLI-dev_max_precision": 0.396011396011396, | |
| "eval_allNLI-dev_max_recall": 0.9421965317919075, | |
| "eval_sequential_score": 0.6292803059732273, | |
| "eval_sts-test_pearson_cosine": 0.8216305666941104, | |
| "eval_sts-test_pearson_dot": 0.7893058557928705, | |
| "eval_sts-test_pearson_euclidean": 0.841701309677291, | |
| "eval_sts-test_pearson_manhattan": 0.8438559225680149, | |
| "eval_sts-test_pearson_max": 0.8438559225680149, | |
| "eval_sts-test_spearman_cosine": 0.8291168678373173, | |
| "eval_sts-test_spearman_dot": 0.7782024179758329, | |
| "eval_sts-test_spearman_euclidean": 0.829029110613729, | |
| "eval_sts-test_spearman_manhattan": 0.8308224279114665, | |
| "eval_sts-test_spearman_max": 0.8308224279114665, | |
| "eval_vitaminc-pairs_loss": 3.763850212097168, | |
| "eval_vitaminc-pairs_runtime": 4.6142, | |
| "eval_vitaminc-pairs_samples_per_second": 27.741, | |
| "eval_vitaminc-pairs_steps_per_second": 0.217, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.15038560411311053, | |
| "eval_negation-triplets_loss": 1.4772971868515015, | |
| "eval_negation-triplets_runtime": 3.4145, | |
| "eval_negation-triplets_samples_per_second": 37.488, | |
| "eval_negation-triplets_steps_per_second": 0.293, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.15038560411311053, | |
| "eval_scitail-pairs-pos_loss": 0.11086193472146988, | |
| "eval_scitail-pairs-pos_runtime": 2.8091, | |
| "eval_scitail-pairs-pos_samples_per_second": 45.565, | |
| "eval_scitail-pairs-pos_steps_per_second": 0.356, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.15038560411311053, | |
| "eval_scitail-pairs-qa_loss": 0.12822097539901733, | |
| "eval_scitail-pairs-qa_runtime": 2.312, | |
| "eval_scitail-pairs-qa_samples_per_second": 55.364, | |
| "eval_scitail-pairs-qa_steps_per_second": 0.433, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.15038560411311053, | |
| "eval_xsum-pairs_loss": 0.43598121404647827, | |
| "eval_xsum-pairs_runtime": 3.1471, | |
| "eval_xsum-pairs_samples_per_second": 40.673, | |
| "eval_xsum-pairs_steps_per_second": 0.318, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.15038560411311053, | |
| "eval_sciq_pairs_loss": 0.20338551700115204, | |
| "eval_sciq_pairs_runtime": 4.8592, | |
| "eval_sciq_pairs_samples_per_second": 26.342, | |
| "eval_sciq_pairs_steps_per_second": 0.206, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.15038560411311053, | |
| "eval_qasc_pairs_loss": 1.3059492111206055, | |
| "eval_qasc_pairs_runtime": 2.1191, | |
| "eval_qasc_pairs_samples_per_second": 60.403, | |
| "eval_qasc_pairs_steps_per_second": 0.472, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.15038560411311053, | |
| "eval_openbookqa_pairs_loss": 1.1711227893829346, | |
| "eval_openbookqa_pairs_runtime": 2.2105, | |
| "eval_openbookqa_pairs_samples_per_second": 57.907, | |
| "eval_openbookqa_pairs_steps_per_second": 0.452, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.15038560411311053, | |
| "eval_msmarco_pairs_loss": 1.3856205940246582, | |
| "eval_msmarco_pairs_runtime": 2.3065, | |
| "eval_msmarco_pairs_samples_per_second": 55.494, | |
| "eval_msmarco_pairs_steps_per_second": 0.434, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.15038560411311053, | |
| "eval_nq_pairs_loss": 1.2713654041290283, | |
| "eval_nq_pairs_runtime": 3.6729, | |
| "eval_nq_pairs_samples_per_second": 34.85, | |
| "eval_nq_pairs_steps_per_second": 0.272, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.15038560411311053, | |
| "eval_trivia_pairs_loss": 1.082784652709961, | |
| "eval_trivia_pairs_runtime": 3.145, | |
| "eval_trivia_pairs_samples_per_second": 40.7, | |
| "eval_trivia_pairs_steps_per_second": 0.318, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.15038560411311053, | |
| "eval_gooaq_pairs_loss": 0.703183650970459, | |
| "eval_gooaq_pairs_runtime": 2.1975, | |
| "eval_gooaq_pairs_samples_per_second": 58.248, | |
| "eval_gooaq_pairs_steps_per_second": 0.455, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.15038560411311053, | |
| "eval_paws-pos_loss": 0.04489962384104729, | |
| "eval_paws-pos_runtime": 2.3636, | |
| "eval_paws-pos_samples_per_second": 54.155, | |
| "eval_paws-pos_steps_per_second": 0.423, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.15038560411311053, | |
| "eval_global_dataset_loss": 0.6713590621948242, | |
| "eval_global_dataset_runtime": 8.4648, | |
| "eval_global_dataset_samples_per_second": 38.394, | |
| "eval_global_dataset_steps_per_second": 0.354, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.15616966580976863, | |
| "grad_norm": 22.163389205932617, | |
| "learning_rate": 4.1119691119691125e-06, | |
| "loss": 0.9674, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.16195372750642673, | |
| "grad_norm": 25.121131896972656, | |
| "learning_rate": 4.2857142857142855e-06, | |
| "loss": 0.9335, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.16773778920308482, | |
| "grad_norm": 44.308528900146484, | |
| "learning_rate": 4.45945945945946e-06, | |
| "loss": 0.8806, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.17352185089974292, | |
| "grad_norm": 9.972362518310547, | |
| "learning_rate": 4.633204633204633e-06, | |
| "loss": 0.631, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.17930591259640102, | |
| "grad_norm": 13.609597206115723, | |
| "learning_rate": 4.806949806949808e-06, | |
| "loss": 0.3384, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.18508997429305912, | |
| "grad_norm": 6.892751693725586, | |
| "learning_rate": 4.980694980694981e-06, | |
| "loss": 0.404, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.19087403598971722, | |
| "grad_norm": 7.414526462554932, | |
| "learning_rate": 5.154440154440155e-06, | |
| "loss": 0.6488, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.19665809768637532, | |
| "grad_norm": 23.75099754333496, | |
| "learning_rate": 5.3281853281853285e-06, | |
| "loss": 0.4728, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.20051413881748073, | |
| "eval_Qnli-dev_cosine_accuracy": 0.693359375, | |
| "eval_Qnli-dev_cosine_accuracy_threshold": 0.8276629447937012, | |
| "eval_Qnli-dev_cosine_ap": 0.7127503402574797, | |
| "eval_Qnli-dev_cosine_f1": 0.6730038022813688, | |
| "eval_Qnli-dev_cosine_f1_threshold": 0.7761298418045044, | |
| "eval_Qnli-dev_cosine_precision": 0.6103448275862069, | |
| "eval_Qnli-dev_cosine_recall": 0.75, | |
| "eval_Qnli-dev_dot_accuracy": 0.681640625, | |
| "eval_Qnli-dev_dot_accuracy_threshold": 857.5540771484375, | |
| "eval_Qnli-dev_dot_ap": 0.6617256136301952, | |
| "eval_Qnli-dev_dot_f1": 0.6678141135972461, | |
| "eval_Qnli-dev_dot_f1_threshold": 756.1309204101562, | |
| "eval_Qnli-dev_dot_precision": 0.5623188405797102, | |
| "eval_Qnli-dev_dot_recall": 0.8220338983050848, | |
| "eval_Qnli-dev_euclidean_accuracy": 0.69140625, | |
| "eval_Qnli-dev_euclidean_accuracy_threshold": 18.45541763305664, | |
| "eval_Qnli-dev_euclidean_ap": 0.7154480594709788, | |
| "eval_Qnli-dev_euclidean_f1": 0.6725043782837128, | |
| "eval_Qnli-dev_euclidean_f1_threshold": 22.723068237304688, | |
| "eval_Qnli-dev_euclidean_precision": 0.573134328358209, | |
| "eval_Qnli-dev_euclidean_recall": 0.8135593220338984, | |
| "eval_Qnli-dev_manhattan_accuracy": 0.69140625, | |
| "eval_Qnli-dev_manhattan_accuracy_threshold": 585.3442993164062, | |
| "eval_Qnli-dev_manhattan_ap": 0.713289230749026, | |
| "eval_Qnli-dev_manhattan_f1": 0.6733333333333333, | |
| "eval_Qnli-dev_manhattan_f1_threshold": 733.2077026367188, | |
| "eval_Qnli-dev_manhattan_precision": 0.554945054945055, | |
| "eval_Qnli-dev_manhattan_recall": 0.8559322033898306, | |
| "eval_Qnli-dev_max_accuracy": 0.693359375, | |
| "eval_Qnli-dev_max_accuracy_threshold": 857.5540771484375, | |
| "eval_Qnli-dev_max_ap": 0.7154480594709788, | |
| "eval_Qnli-dev_max_f1": 0.6733333333333333, | |
| "eval_Qnli-dev_max_f1_threshold": 756.1309204101562, | |
| "eval_Qnli-dev_max_precision": 0.6103448275862069, | |
| "eval_Qnli-dev_max_recall": 0.8559322033898306, | |
| "eval_allNLI-dev_cosine_accuracy": 0.693359375, | |
| "eval_allNLI-dev_cosine_accuracy_threshold": 0.8820043802261353, | |
| "eval_allNLI-dev_cosine_ap": 0.5380653098565689, | |
| "eval_allNLI-dev_cosine_f1": 0.5689655172413792, | |
| "eval_allNLI-dev_cosine_f1_threshold": 0.7308224439620972, | |
| "eval_allNLI-dev_cosine_precision": 0.4536082474226804, | |
| "eval_allNLI-dev_cosine_recall": 0.7630057803468208, | |
| "eval_allNLI-dev_dot_accuracy": 0.6875, | |
| "eval_allNLI-dev_dot_accuracy_threshold": 904.0134887695312, | |
| "eval_allNLI-dev_dot_ap": 0.5175334662627952, | |
| "eval_allNLI-dev_dot_f1": 0.5675675675675675, | |
| "eval_allNLI-dev_dot_f1_threshold": 762.652587890625, | |
| "eval_allNLI-dev_dot_precision": 0.46494464944649444, | |
| "eval_allNLI-dev_dot_recall": 0.7283236994219653, | |
| "eval_allNLI-dev_euclidean_accuracy": 0.6953125, | |
| "eval_allNLI-dev_euclidean_accuracy_threshold": 15.50227165222168, | |
| "eval_allNLI-dev_euclidean_ap": 0.5410854453600531, | |
| "eval_allNLI-dev_euclidean_f1": 0.5645514223194749, | |
| "eval_allNLI-dev_euclidean_f1_threshold": 23.451454162597656, | |
| "eval_allNLI-dev_euclidean_precision": 0.45422535211267606, | |
| "eval_allNLI-dev_euclidean_recall": 0.7456647398843931, | |
| "eval_allNLI-dev_manhattan_accuracy": 0.6953125, | |
| "eval_allNLI-dev_manhattan_accuracy_threshold": 518.75146484375, | |
| "eval_allNLI-dev_manhattan_ap": 0.5430149701961973, | |
| "eval_allNLI-dev_manhattan_f1": 0.5639913232104121, | |
| "eval_allNLI-dev_manhattan_f1_threshold": 733.8897705078125, | |
| "eval_allNLI-dev_manhattan_precision": 0.4513888888888889, | |
| "eval_allNLI-dev_manhattan_recall": 0.7514450867052023, | |
| "eval_allNLI-dev_max_accuracy": 0.6953125, | |
| "eval_allNLI-dev_max_accuracy_threshold": 904.0134887695312, | |
| "eval_allNLI-dev_max_ap": 0.5430149701961973, | |
| "eval_allNLI-dev_max_f1": 0.5689655172413792, | |
| "eval_allNLI-dev_max_f1_threshold": 762.652587890625, | |
| "eval_allNLI-dev_max_precision": 0.46494464944649444, | |
| "eval_allNLI-dev_max_recall": 0.7630057803468208, | |
| "eval_sequential_score": 0.7154480594709788, | |
| "eval_sts-test_pearson_cosine": 0.8841456597076958, | |
| "eval_sts-test_pearson_dot": 0.8597263949983345, | |
| "eval_sts-test_pearson_euclidean": 0.8977156894521804, | |
| "eval_sts-test_pearson_manhattan": 0.8973099980705126, | |
| "eval_sts-test_pearson_max": 0.8977156894521804, | |
| "eval_sts-test_spearman_cosine": 0.8862653474156692, | |
| "eval_sts-test_spearman_dot": 0.8451249310521576, | |
| "eval_sts-test_spearman_euclidean": 0.8866065310325173, | |
| "eval_sts-test_spearman_manhattan": 0.8860841943095571, | |
| "eval_sts-test_spearman_max": 0.8866065310325173, | |
| "eval_vitaminc-pairs_loss": 3.314167022705078, | |
| "eval_vitaminc-pairs_runtime": 4.4886, | |
| "eval_vitaminc-pairs_samples_per_second": 28.517, | |
| "eval_vitaminc-pairs_steps_per_second": 0.223, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.20051413881748073, | |
| "eval_negation-triplets_loss": 1.3400739431381226, | |
| "eval_negation-triplets_runtime": 3.3068, | |
| "eval_negation-triplets_samples_per_second": 38.708, | |
| "eval_negation-triplets_steps_per_second": 0.302, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.20051413881748073, | |
| "eval_scitail-pairs-pos_loss": 0.07263079285621643, | |
| "eval_scitail-pairs-pos_runtime": 2.6213, | |
| "eval_scitail-pairs-pos_samples_per_second": 48.83, | |
| "eval_scitail-pairs-pos_steps_per_second": 0.381, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.20051413881748073, | |
| "eval_scitail-pairs-qa_loss": 0.048478204756975174, | |
| "eval_scitail-pairs-qa_runtime": 2.202, | |
| "eval_scitail-pairs-qa_samples_per_second": 58.13, | |
| "eval_scitail-pairs-qa_steps_per_second": 0.454, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.20051413881748073, | |
| "eval_xsum-pairs_loss": 0.14091075956821442, | |
| "eval_xsum-pairs_runtime": 3.0891, | |
| "eval_xsum-pairs_samples_per_second": 41.436, | |
| "eval_xsum-pairs_steps_per_second": 0.324, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.20051413881748073, | |
| "eval_sciq_pairs_loss": 0.1505264937877655, | |
| "eval_sciq_pairs_runtime": 4.762, | |
| "eval_sciq_pairs_samples_per_second": 26.879, | |
| "eval_sciq_pairs_steps_per_second": 0.21, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.20051413881748073, | |
| "eval_qasc_pairs_loss": 0.4244489073753357, | |
| "eval_qasc_pairs_runtime": 2.0783, | |
| "eval_qasc_pairs_samples_per_second": 61.589, | |
| "eval_qasc_pairs_steps_per_second": 0.481, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.20051413881748073, | |
| "eval_openbookqa_pairs_loss": 0.5923758745193481, | |
| "eval_openbookqa_pairs_runtime": 2.1938, | |
| "eval_openbookqa_pairs_samples_per_second": 58.345, | |
| "eval_openbookqa_pairs_steps_per_second": 0.456, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.20051413881748073, | |
| "eval_msmarco_pairs_loss": 0.5882836580276489, | |
| "eval_msmarco_pairs_runtime": 2.2483, | |
| "eval_msmarco_pairs_samples_per_second": 56.933, | |
| "eval_msmarco_pairs_steps_per_second": 0.445, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.20051413881748073, | |
| "eval_nq_pairs_loss": 0.5424538254737854, | |
| "eval_nq_pairs_runtime": 3.6424, | |
| "eval_nq_pairs_samples_per_second": 35.142, | |
| "eval_nq_pairs_steps_per_second": 0.275, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.20051413881748073, | |
| "eval_trivia_pairs_loss": 0.6437886953353882, | |
| "eval_trivia_pairs_runtime": 3.0809, | |
| "eval_trivia_pairs_samples_per_second": 41.546, | |
| "eval_trivia_pairs_steps_per_second": 0.325, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.20051413881748073, | |
| "eval_gooaq_pairs_loss": 0.22378715872764587, | |
| "eval_gooaq_pairs_runtime": 2.101, | |
| "eval_gooaq_pairs_samples_per_second": 60.924, | |
| "eval_gooaq_pairs_steps_per_second": 0.476, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.20051413881748073, | |
| "eval_paws-pos_loss": 0.03581737354397774, | |
| "eval_paws-pos_runtime": 2.2447, | |
| "eval_paws-pos_samples_per_second": 57.024, | |
| "eval_paws-pos_steps_per_second": 0.445, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.20051413881748073, | |
| "eval_global_dataset_loss": 0.3989756405353546, | |
| "eval_global_dataset_runtime": 8.402, | |
| "eval_global_dataset_samples_per_second": 38.681, | |
| "eval_global_dataset_steps_per_second": 0.357, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.20244215938303342, | |
| "grad_norm": 96.23986053466797, | |
| "learning_rate": 5.5019305019305015e-06, | |
| "loss": 0.5094, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.20822622107969152, | |
| "grad_norm": 33.88426208496094, | |
| "learning_rate": 5.675675675675676e-06, | |
| "loss": 0.8002, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.21401028277634962, | |
| "grad_norm": 7.319549560546875, | |
| "learning_rate": 5.849420849420849e-06, | |
| "loss": 0.3886, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.21979434447300772, | |
| "grad_norm": 8.65904712677002, | |
| "learning_rate": 6.023166023166023e-06, | |
| "loss": 0.6937, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.22557840616966582, | |
| "grad_norm": 376.61663818359375, | |
| "learning_rate": 6.196911196911197e-06, | |
| "loss": 0.2909, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.23136246786632392, | |
| "grad_norm": 10.809043884277344, | |
| "learning_rate": 6.370656370656371e-06, | |
| "loss": 0.3885, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.23714652956298202, | |
| "grad_norm": 8.564398765563965, | |
| "learning_rate": 6.5444015444015445e-06, | |
| "loss": 0.29, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.24293059125964012, | |
| "grad_norm": 1.5215222835540771, | |
| "learning_rate": 6.718146718146718e-06, | |
| "loss": 0.3485, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.2487146529562982, | |
| "grad_norm": 181.20809936523438, | |
| "learning_rate": 6.891891891891892e-06, | |
| "loss": 0.3931, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.2506426735218509, | |
| "eval_Qnli-dev_cosine_accuracy": 0.67578125, | |
| "eval_Qnli-dev_cosine_accuracy_threshold": 0.7986456155776978, | |
| "eval_Qnli-dev_cosine_ap": 0.6985736393833883, | |
| "eval_Qnli-dev_cosine_f1": 0.6777777777777777, | |
| "eval_Qnli-dev_cosine_f1_threshold": 0.726032018661499, | |
| "eval_Qnli-dev_cosine_precision": 0.6019736842105263, | |
| "eval_Qnli-dev_cosine_recall": 0.7754237288135594, | |
| "eval_Qnli-dev_dot_accuracy": 0.66796875, | |
| "eval_Qnli-dev_dot_accuracy_threshold": 840.5162353515625, | |
| "eval_Qnli-dev_dot_ap": 0.6656432640480556, | |
| "eval_Qnli-dev_dot_f1": 0.662020905923345, | |
| "eval_Qnli-dev_dot_f1_threshold": 715.1839599609375, | |
| "eval_Qnli-dev_dot_precision": 0.5621301775147929, | |
| "eval_Qnli-dev_dot_recall": 0.8050847457627118, | |
| "eval_Qnli-dev_euclidean_accuracy": 0.677734375, | |
| "eval_Qnli-dev_euclidean_accuracy_threshold": 20.397323608398438, | |
| "eval_Qnli-dev_euclidean_ap": 0.6992667316315335, | |
| "eval_Qnli-dev_euclidean_f1": 0.67296786389414, | |
| "eval_Qnli-dev_euclidean_f1_threshold": 23.44580841064453, | |
| "eval_Qnli-dev_euclidean_precision": 0.6075085324232082, | |
| "eval_Qnli-dev_euclidean_recall": 0.7542372881355932, | |
| "eval_Qnli-dev_manhattan_accuracy": 0.671875, | |
| "eval_Qnli-dev_manhattan_accuracy_threshold": 628.4102172851562, | |
| "eval_Qnli-dev_manhattan_ap": 0.6982133105013253, | |
| "eval_Qnli-dev_manhattan_f1": 0.673992673992674, | |
| "eval_Qnli-dev_manhattan_f1_threshold": 746.4862060546875, | |
| "eval_Qnli-dev_manhattan_precision": 0.5935483870967742, | |
| "eval_Qnli-dev_manhattan_recall": 0.7796610169491526, | |
| "eval_Qnli-dev_max_accuracy": 0.677734375, | |
| "eval_Qnli-dev_max_accuracy_threshold": 840.5162353515625, | |
| "eval_Qnli-dev_max_ap": 0.6992667316315335, | |
| "eval_Qnli-dev_max_f1": 0.6777777777777777, | |
| "eval_Qnli-dev_max_f1_threshold": 746.4862060546875, | |
| "eval_Qnli-dev_max_precision": 0.6075085324232082, | |
| "eval_Qnli-dev_max_recall": 0.8050847457627118, | |
| "eval_allNLI-dev_cosine_accuracy": 0.7109375, | |
| "eval_allNLI-dev_cosine_accuracy_threshold": 0.8466876745223999, | |
| "eval_allNLI-dev_cosine_ap": 0.5614885136859779, | |
| "eval_allNLI-dev_cosine_f1": 0.5809128630705395, | |
| "eval_allNLI-dev_cosine_f1_threshold": 0.6695160269737244, | |
| "eval_allNLI-dev_cosine_precision": 0.45307443365695793, | |
| "eval_allNLI-dev_cosine_recall": 0.8092485549132948, | |
| "eval_allNLI-dev_dot_accuracy": 0.69921875, | |
| "eval_allNLI-dev_dot_accuracy_threshold": 901.9229736328125, | |
| "eval_allNLI-dev_dot_ap": 0.5637771052921187, | |
| "eval_allNLI-dev_dot_f1": 0.5812807881773399, | |
| "eval_allNLI-dev_dot_f1_threshold": 752.6895141601562, | |
| "eval_allNLI-dev_dot_precision": 0.5064377682403434, | |
| "eval_allNLI-dev_dot_recall": 0.6820809248554913, | |
| "eval_allNLI-dev_euclidean_accuracy": 0.708984375, | |
| "eval_allNLI-dev_euclidean_accuracy_threshold": 17.844379425048828, | |
| "eval_allNLI-dev_euclidean_ap": 0.5616160818796474, | |
| "eval_allNLI-dev_euclidean_f1": 0.5838509316770187, | |
| "eval_allNLI-dev_euclidean_f1_threshold": 26.068946838378906, | |
| "eval_allNLI-dev_euclidean_precision": 0.45483870967741935, | |
| "eval_allNLI-dev_euclidean_recall": 0.815028901734104, | |
| "eval_allNLI-dev_manhattan_accuracy": 0.7109375, | |
| "eval_allNLI-dev_manhattan_accuracy_threshold": 553.34228515625, | |
| "eval_allNLI-dev_manhattan_ap": 0.5624349287984178, | |
| "eval_allNLI-dev_manhattan_f1": 0.5778688524590163, | |
| "eval_allNLI-dev_manhattan_f1_threshold": 816.2254028320312, | |
| "eval_allNLI-dev_manhattan_precision": 0.44761904761904764, | |
| "eval_allNLI-dev_manhattan_recall": 0.815028901734104, | |
| "eval_allNLI-dev_max_accuracy": 0.7109375, | |
| "eval_allNLI-dev_max_accuracy_threshold": 901.9229736328125, | |
| "eval_allNLI-dev_max_ap": 0.5637771052921187, | |
| "eval_allNLI-dev_max_f1": 0.5838509316770187, | |
| "eval_allNLI-dev_max_f1_threshold": 816.2254028320312, | |
| "eval_allNLI-dev_max_precision": 0.5064377682403434, | |
| "eval_allNLI-dev_max_recall": 0.815028901734104, | |
| "eval_sequential_score": 0.6992667316315335, | |
| "eval_sts-test_pearson_cosine": 0.9024404439516205, | |
| "eval_sts-test_pearson_dot": 0.8941554024395801, | |
| "eval_sts-test_pearson_euclidean": 0.9164387414940136, | |
| "eval_sts-test_pearson_manhattan": 0.9158406136217865, | |
| "eval_sts-test_pearson_max": 0.9164387414940136, | |
| "eval_sts-test_spearman_cosine": 0.9081672412582797, | |
| "eval_sts-test_spearman_dot": 0.8855668249766109, | |
| "eval_sts-test_spearman_euclidean": 0.9076990535616029, | |
| "eval_sts-test_spearman_manhattan": 0.907316833038059, | |
| "eval_sts-test_spearman_max": 0.9081672412582797, | |
| "eval_vitaminc-pairs_loss": 3.8795158863067627, | |
| "eval_vitaminc-pairs_runtime": 4.6143, | |
| "eval_vitaminc-pairs_samples_per_second": 27.74, | |
| "eval_vitaminc-pairs_steps_per_second": 0.217, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.2506426735218509, | |
| "eval_negation-triplets_loss": 1.1134737730026245, | |
| "eval_negation-triplets_runtime": 3.4103, | |
| "eval_negation-triplets_samples_per_second": 37.533, | |
| "eval_negation-triplets_steps_per_second": 0.293, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.2506426735218509, | |
| "eval_scitail-pairs-pos_loss": 0.045735035091638565, | |
| "eval_scitail-pairs-pos_runtime": 2.7664, | |
| "eval_scitail-pairs-pos_samples_per_second": 46.27, | |
| "eval_scitail-pairs-pos_steps_per_second": 0.361, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.2506426735218509, | |
| "eval_scitail-pairs-qa_loss": 0.036478620022535324, | |
| "eval_scitail-pairs-qa_runtime": 2.2905, | |
| "eval_scitail-pairs-qa_samples_per_second": 55.884, | |
| "eval_scitail-pairs-qa_steps_per_second": 0.437, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.2506426735218509, | |
| "eval_xsum-pairs_loss": 0.05133158713579178, | |
| "eval_xsum-pairs_runtime": 3.1323, | |
| "eval_xsum-pairs_samples_per_second": 40.865, | |
| "eval_xsum-pairs_steps_per_second": 0.319, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.2506426735218509, | |
| "eval_sciq_pairs_loss": 0.12314368784427643, | |
| "eval_sciq_pairs_runtime": 4.8477, | |
| "eval_sciq_pairs_samples_per_second": 26.404, | |
| "eval_sciq_pairs_steps_per_second": 0.206, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.2506426735218509, | |
| "eval_qasc_pairs_loss": 0.22055241465568542, | |
| "eval_qasc_pairs_runtime": 2.1243, | |
| "eval_qasc_pairs_samples_per_second": 60.255, | |
| "eval_qasc_pairs_steps_per_second": 0.471, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.2506426735218509, | |
| "eval_openbookqa_pairs_loss": 0.4567987024784088, | |
| "eval_openbookqa_pairs_runtime": 2.3078, | |
| "eval_openbookqa_pairs_samples_per_second": 55.465, | |
| "eval_openbookqa_pairs_steps_per_second": 0.433, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.2506426735218509, | |
| "eval_msmarco_pairs_loss": 0.21620944142341614, | |
| "eval_msmarco_pairs_runtime": 2.3399, | |
| "eval_msmarco_pairs_samples_per_second": 54.703, | |
| "eval_msmarco_pairs_steps_per_second": 0.427, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.2506426735218509, | |
| "eval_nq_pairs_loss": 0.26776355504989624, | |
| "eval_nq_pairs_runtime": 3.7038, | |
| "eval_nq_pairs_samples_per_second": 34.559, | |
| "eval_nq_pairs_steps_per_second": 0.27, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.2506426735218509, | |
| "eval_trivia_pairs_loss": 0.15397313237190247, | |
| "eval_trivia_pairs_runtime": 3.1256, | |
| "eval_trivia_pairs_samples_per_second": 40.952, | |
| "eval_trivia_pairs_steps_per_second": 0.32, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.2506426735218509, | |
| "eval_gooaq_pairs_loss": 0.12057099491357803, | |
| "eval_gooaq_pairs_runtime": 2.1825, | |
| "eval_gooaq_pairs_samples_per_second": 58.648, | |
| "eval_gooaq_pairs_steps_per_second": 0.458, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.2506426735218509, | |
| "eval_paws-pos_loss": 0.026105860248208046, | |
| "eval_paws-pos_runtime": 2.2995, | |
| "eval_paws-pos_samples_per_second": 55.665, | |
| "eval_paws-pos_steps_per_second": 0.435, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.2506426735218509, | |
| "eval_global_dataset_loss": 0.3591004014015198, | |
| "eval_global_dataset_runtime": 8.4153, | |
| "eval_global_dataset_samples_per_second": 38.62, | |
| "eval_global_dataset_steps_per_second": 0.356, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.2544987146529563, | |
| "grad_norm": 11.167648315429688, | |
| "learning_rate": 7.065637065637066e-06, | |
| "loss": 0.3394, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.2602827763496144, | |
| "grad_norm": 3.4576871395111084, | |
| "learning_rate": 7.239382239382239e-06, | |
| "loss": 0.1276, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.2660668380462725, | |
| "grad_norm": 9.594191551208496, | |
| "learning_rate": 7.413127413127414e-06, | |
| "loss": 0.3569, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.2718508997429306, | |
| "grad_norm": 4.255222320556641, | |
| "learning_rate": 7.586872586872587e-06, | |
| "loss": 0.1231, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.2776349614395887, | |
| "grad_norm": 6.074405670166016, | |
| "learning_rate": 7.76061776061776e-06, | |
| "loss": 0.3086, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.2834190231362468, | |
| "grad_norm": 19.442323684692383, | |
| "learning_rate": 7.934362934362936e-06, | |
| "loss": 0.3541, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.2892030848329049, | |
| "grad_norm": 5.141862392425537, | |
| "learning_rate": 8.108108108108109e-06, | |
| "loss": 0.2597, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.294987146529563, | |
| "grad_norm": 5.942401885986328, | |
| "learning_rate": 8.281853281853282e-06, | |
| "loss": 0.1585, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.30077120822622105, | |
| "grad_norm": 97.85575103759766, | |
| "learning_rate": 8.455598455598455e-06, | |
| "loss": 0.1436, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.30077120822622105, | |
| "eval_Qnli-dev_cosine_accuracy": 0.693359375, | |
| "eval_Qnli-dev_cosine_accuracy_threshold": 0.7395063638687134, | |
| "eval_Qnli-dev_cosine_ap": 0.7187155891006661, | |
| "eval_Qnli-dev_cosine_f1": 0.6837606837606838, | |
| "eval_Qnli-dev_cosine_f1_threshold": 0.660994291305542, | |
| "eval_Qnli-dev_cosine_precision": 0.5730659025787965, | |
| "eval_Qnli-dev_cosine_recall": 0.847457627118644, | |
| "eval_Qnli-dev_dot_accuracy": 0.669921875, | |
| "eval_Qnli-dev_dot_accuracy_threshold": 701.7633056640625, | |
| "eval_Qnli-dev_dot_ap": 0.6888124099944863, | |
| "eval_Qnli-dev_dot_f1": 0.6812816188870151, | |
| "eval_Qnli-dev_dot_f1_threshold": 633.31884765625, | |
| "eval_Qnli-dev_dot_precision": 0.5658263305322129, | |
| "eval_Qnli-dev_dot_recall": 0.8559322033898306, | |
| "eval_Qnli-dev_euclidean_accuracy": 0.697265625, | |
| "eval_Qnli-dev_euclidean_accuracy_threshold": 22.954120635986328, | |
| "eval_Qnli-dev_euclidean_ap": 0.721593279891293, | |
| "eval_Qnli-dev_euclidean_f1": 0.6762075134168157, | |
| "eval_Qnli-dev_euclidean_f1_threshold": 24.89653968811035, | |
| "eval_Qnli-dev_euclidean_precision": 0.5851393188854489, | |
| "eval_Qnli-dev_euclidean_recall": 0.8008474576271186, | |
| "eval_Qnli-dev_manhattan_accuracy": 0.6953125, | |
| "eval_Qnli-dev_manhattan_accuracy_threshold": 718.2297973632812, | |
| "eval_Qnli-dev_manhattan_ap": 0.7239405100010182, | |
| "eval_Qnli-dev_manhattan_f1": 0.6807017543859649, | |
| "eval_Qnli-dev_manhattan_f1_threshold": 784.701171875, | |
| "eval_Qnli-dev_manhattan_precision": 0.5808383233532934, | |
| "eval_Qnli-dev_manhattan_recall": 0.8220338983050848, | |
| "eval_Qnli-dev_max_accuracy": 0.697265625, | |
| "eval_Qnli-dev_max_accuracy_threshold": 718.2297973632812, | |
| "eval_Qnli-dev_max_ap": 0.7239405100010182, | |
| "eval_Qnli-dev_max_f1": 0.6837606837606838, | |
| "eval_Qnli-dev_max_f1_threshold": 784.701171875, | |
| "eval_Qnli-dev_max_precision": 0.5851393188854489, | |
| "eval_Qnli-dev_max_recall": 0.8559322033898306, | |
| "eval_allNLI-dev_cosine_accuracy": 0.71875, | |
| "eval_allNLI-dev_cosine_accuracy_threshold": 0.8460206985473633, | |
| "eval_allNLI-dev_cosine_ap": 0.5796621470925665, | |
| "eval_allNLI-dev_cosine_f1": 0.6139954853273137, | |
| "eval_allNLI-dev_cosine_f1_threshold": 0.7097446322441101, | |
| "eval_allNLI-dev_cosine_precision": 0.5037037037037037, | |
| "eval_allNLI-dev_cosine_recall": 0.7861271676300579, | |
| "eval_allNLI-dev_dot_accuracy": 0.7109375, | |
| "eval_allNLI-dev_dot_accuracy_threshold": 837.8741455078125, | |
| "eval_allNLI-dev_dot_ap": 0.5763352242724786, | |
| "eval_allNLI-dev_dot_f1": 0.6179245283018867, | |
| "eval_allNLI-dev_dot_f1_threshold": 712.9136352539062, | |
| "eval_allNLI-dev_dot_precision": 0.5219123505976095, | |
| "eval_allNLI-dev_dot_recall": 0.7572254335260116, | |
| "eval_allNLI-dev_euclidean_accuracy": 0.724609375, | |
| "eval_allNLI-dev_euclidean_accuracy_threshold": 18.82543182373047, | |
| "eval_allNLI-dev_euclidean_ap": 0.5803015452539461, | |
| "eval_allNLI-dev_euclidean_f1": 0.6085011185682326, | |
| "eval_allNLI-dev_euclidean_f1_threshold": 24.0869197845459, | |
| "eval_allNLI-dev_euclidean_precision": 0.49635036496350365, | |
| "eval_allNLI-dev_euclidean_recall": 0.7861271676300579, | |
| "eval_allNLI-dev_manhattan_accuracy": 0.720703125, | |
| "eval_allNLI-dev_manhattan_accuracy_threshold": 586.8082275390625, | |
| "eval_allNLI-dev_manhattan_ap": 0.5806697040885319, | |
| "eval_allNLI-dev_manhattan_f1": 0.6035242290748899, | |
| "eval_allNLI-dev_manhattan_f1_threshold": 756.8704833984375, | |
| "eval_allNLI-dev_manhattan_precision": 0.4875444839857651, | |
| "eval_allNLI-dev_manhattan_recall": 0.791907514450867, | |
| "eval_allNLI-dev_max_accuracy": 0.724609375, | |
| "eval_allNLI-dev_max_accuracy_threshold": 837.8741455078125, | |
| "eval_allNLI-dev_max_ap": 0.5806697040885319, | |
| "eval_allNLI-dev_max_f1": 0.6179245283018867, | |
| "eval_allNLI-dev_max_f1_threshold": 756.8704833984375, | |
| "eval_allNLI-dev_max_precision": 0.5219123505976095, | |
| "eval_allNLI-dev_max_recall": 0.791907514450867, | |
| "eval_sequential_score": 0.7239405100010182, | |
| "eval_sts-test_pearson_cosine": 0.9086022204177822, | |
| "eval_sts-test_pearson_dot": 0.8963707820565703, | |
| "eval_sts-test_pearson_euclidean": 0.9226263845809013, | |
| "eval_sts-test_pearson_manhattan": 0.9216624653503607, | |
| "eval_sts-test_pearson_max": 0.9226263845809013, | |
| "eval_sts-test_spearman_cosine": 0.9176886091344368, | |
| "eval_sts-test_spearman_dot": 0.8941249219513181, | |
| "eval_sts-test_spearman_euclidean": 0.9176283739366038, | |
| "eval_sts-test_spearman_manhattan": 0.9167815458172944, | |
| "eval_sts-test_spearman_max": 0.9176886091344368, | |
| "eval_vitaminc-pairs_loss": 3.8346316814422607, | |
| "eval_vitaminc-pairs_runtime": 4.4216, | |
| "eval_vitaminc-pairs_samples_per_second": 28.949, | |
| "eval_vitaminc-pairs_steps_per_second": 0.226, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.30077120822622105, | |
| "eval_negation-triplets_loss": 0.9441184997558594, | |
| "eval_negation-triplets_runtime": 3.3236, | |
| "eval_negation-triplets_samples_per_second": 38.512, | |
| "eval_negation-triplets_steps_per_second": 0.301, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.30077120822622105, | |
| "eval_scitail-pairs-pos_loss": 0.03759249672293663, | |
| "eval_scitail-pairs-pos_runtime": 2.6436, | |
| "eval_scitail-pairs-pos_samples_per_second": 48.418, | |
| "eval_scitail-pairs-pos_steps_per_second": 0.378, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.30077120822622105, | |
| "eval_scitail-pairs-qa_loss": 0.04104204475879669, | |
| "eval_scitail-pairs-qa_runtime": 2.205, | |
| "eval_scitail-pairs-qa_samples_per_second": 58.05, | |
| "eval_scitail-pairs-qa_steps_per_second": 0.454, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.30077120822622105, | |
| "eval_xsum-pairs_loss": 0.05032652989029884, | |
| "eval_xsum-pairs_runtime": 3.0619, | |
| "eval_xsum-pairs_samples_per_second": 41.805, | |
| "eval_xsum-pairs_steps_per_second": 0.327, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.30077120822622105, | |
| "eval_sciq_pairs_loss": 0.08709195256233215, | |
| "eval_sciq_pairs_runtime": 4.8405, | |
| "eval_sciq_pairs_samples_per_second": 26.444, | |
| "eval_sciq_pairs_steps_per_second": 0.207, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.30077120822622105, | |
| "eval_qasc_pairs_loss": 0.17168982326984406, | |
| "eval_qasc_pairs_runtime": 2.078, | |
| "eval_qasc_pairs_samples_per_second": 61.599, | |
| "eval_qasc_pairs_steps_per_second": 0.481, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.30077120822622105, | |
| "eval_openbookqa_pairs_loss": 0.4304167926311493, | |
| "eval_openbookqa_pairs_runtime": 2.1991, | |
| "eval_openbookqa_pairs_samples_per_second": 58.205, | |
| "eval_openbookqa_pairs_steps_per_second": 0.455, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.30077120822622105, | |
| "eval_msmarco_pairs_loss": 0.13501481711864471, | |
| "eval_msmarco_pairs_runtime": 2.2186, | |
| "eval_msmarco_pairs_samples_per_second": 57.694, | |
| "eval_msmarco_pairs_steps_per_second": 0.451, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.30077120822622105, | |
| "eval_nq_pairs_loss": 0.20035718381404877, | |
| "eval_nq_pairs_runtime": 3.6367, | |
| "eval_nq_pairs_samples_per_second": 35.197, | |
| "eval_nq_pairs_steps_per_second": 0.275, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.30077120822622105, | |
| "eval_trivia_pairs_loss": 0.08228898048400879, | |
| "eval_trivia_pairs_runtime": 3.086, | |
| "eval_trivia_pairs_samples_per_second": 41.477, | |
| "eval_trivia_pairs_steps_per_second": 0.324, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.30077120822622105, | |
| "eval_gooaq_pairs_loss": 0.0767039954662323, | |
| "eval_gooaq_pairs_runtime": 2.1123, | |
| "eval_gooaq_pairs_samples_per_second": 60.599, | |
| "eval_gooaq_pairs_steps_per_second": 0.473, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.30077120822622105, | |
| "eval_paws-pos_loss": 0.02134825475513935, | |
| "eval_paws-pos_runtime": 2.2234, | |
| "eval_paws-pos_samples_per_second": 57.571, | |
| "eval_paws-pos_steps_per_second": 0.45, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.30077120822622105, | |
| "eval_global_dataset_loss": 0.2958550453186035, | |
| "eval_global_dataset_runtime": 8.3542, | |
| "eval_global_dataset_samples_per_second": 38.903, | |
| "eval_global_dataset_steps_per_second": 0.359, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.3065552699228792, | |
| "grad_norm": 47.52979278564453, | |
| "learning_rate": 8.62934362934363e-06, | |
| "loss": 0.1941, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.31233933161953725, | |
| "grad_norm": 1234.25927734375, | |
| "learning_rate": 8.803088803088804e-06, | |
| "loss": 0.3041, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.3181233933161954, | |
| "grad_norm": 185.80238342285156, | |
| "learning_rate": 8.976833976833977e-06, | |
| "loss": 0.2358, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.32390745501285345, | |
| "grad_norm": 7.233102798461914, | |
| "learning_rate": 9.15057915057915e-06, | |
| "loss": 0.2148, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.3296915167095116, | |
| "grad_norm": 342.77685546875, | |
| "learning_rate": 9.324324324324325e-06, | |
| "loss": 0.8567, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.33547557840616965, | |
| "grad_norm": 181.07650756835938, | |
| "learning_rate": 9.498069498069498e-06, | |
| "loss": 0.3668, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.3412596401028278, | |
| "grad_norm": 198.4124755859375, | |
| "learning_rate": 9.671814671814673e-06, | |
| "loss": 0.3278, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.34704370179948585, | |
| "grad_norm": 91.80472564697266, | |
| "learning_rate": 9.845559845559846e-06, | |
| "loss": 0.474, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.3508997429305913, | |
| "eval_Qnli-dev_cosine_accuracy": 0.6953125, | |
| "eval_Qnli-dev_cosine_accuracy_threshold": 0.7320406436920166, | |
| "eval_Qnli-dev_cosine_ap": 0.7163901550352688, | |
| "eval_Qnli-dev_cosine_f1": 0.6857142857142857, | |
| "eval_Qnli-dev_cosine_f1_threshold": 0.6526604890823364, | |
| "eval_Qnli-dev_cosine_precision": 0.5682451253481894, | |
| "eval_Qnli-dev_cosine_recall": 0.864406779661017, | |
| "eval_Qnli-dev_dot_accuracy": 0.67578125, | |
| "eval_Qnli-dev_dot_accuracy_threshold": 759.5008544921875, | |
| "eval_Qnli-dev_dot_ap": 0.6857309318375049, | |
| "eval_Qnli-dev_dot_f1": 0.6857142857142857, | |
| "eval_Qnli-dev_dot_f1_threshold": 653.0701904296875, | |
| "eval_Qnli-dev_dot_precision": 0.5682451253481894, | |
| "eval_Qnli-dev_dot_recall": 0.864406779661017, | |
| "eval_Qnli-dev_euclidean_accuracy": 0.689453125, | |
| "eval_Qnli-dev_euclidean_accuracy_threshold": 23.348251342773438, | |
| "eval_Qnli-dev_euclidean_ap": 0.7186338110228241, | |
| "eval_Qnli-dev_euclidean_f1": 0.6819672131147542, | |
| "eval_Qnli-dev_euclidean_f1_threshold": 26.874767303466797, | |
| "eval_Qnli-dev_euclidean_precision": 0.5561497326203209, | |
| "eval_Qnli-dev_euclidean_recall": 0.8813559322033898, | |
| "eval_Qnli-dev_manhattan_accuracy": 0.6875, | |
| "eval_Qnli-dev_manhattan_accuracy_threshold": 695.7578125, | |
| "eval_Qnli-dev_manhattan_ap": 0.7187237394606566, | |
| "eval_Qnli-dev_manhattan_f1": 0.6815286624203821, | |
| "eval_Qnli-dev_manhattan_f1_threshold": 856.7666015625, | |
| "eval_Qnli-dev_manhattan_precision": 0.5459183673469388, | |
| "eval_Qnli-dev_manhattan_recall": 0.9067796610169492, | |
| "eval_Qnli-dev_max_accuracy": 0.6953125, | |
| "eval_Qnli-dev_max_accuracy_threshold": 759.5008544921875, | |
| "eval_Qnli-dev_max_ap": 0.7187237394606566, | |
| "eval_Qnli-dev_max_f1": 0.6857142857142857, | |
| "eval_Qnli-dev_max_f1_threshold": 856.7666015625, | |
| "eval_Qnli-dev_max_precision": 0.5682451253481894, | |
| "eval_Qnli-dev_max_recall": 0.9067796610169492, | |
| "eval_allNLI-dev_cosine_accuracy": 0.705078125, | |
| "eval_allNLI-dev_cosine_accuracy_threshold": 0.8531240224838257, | |
| "eval_allNLI-dev_cosine_ap": 0.5555367525973895, | |
| "eval_allNLI-dev_cosine_f1": 0.5972222222222222, | |
| "eval_allNLI-dev_cosine_f1_threshold": 0.7072098255157471, | |
| "eval_allNLI-dev_cosine_precision": 0.4980694980694981, | |
| "eval_allNLI-dev_cosine_recall": 0.7456647398843931, | |
| "eval_allNLI-dev_dot_accuracy": 0.705078125, | |
| "eval_allNLI-dev_dot_accuracy_threshold": 858.6700439453125, | |
| "eval_allNLI-dev_dot_ap": 0.5501313597611072, | |
| "eval_allNLI-dev_dot_f1": 0.5992217898832685, | |
| "eval_allNLI-dev_dot_f1_threshold": 657.9625244140625, | |
| "eval_allNLI-dev_dot_precision": 0.45161290322580644, | |
| "eval_allNLI-dev_dot_recall": 0.8901734104046243, | |
| "eval_allNLI-dev_euclidean_accuracy": 0.705078125, | |
| "eval_allNLI-dev_euclidean_accuracy_threshold": 16.42839813232422, | |
| "eval_allNLI-dev_euclidean_ap": 0.5519081828104085, | |
| "eval_allNLI-dev_euclidean_f1": 0.5919282511210763, | |
| "eval_allNLI-dev_euclidean_f1_threshold": 24.696884155273438, | |
| "eval_allNLI-dev_euclidean_precision": 0.4835164835164835, | |
| "eval_allNLI-dev_euclidean_recall": 0.7630057803468208, | |
| "eval_allNLI-dev_manhattan_accuracy": 0.705078125, | |
| "eval_allNLI-dev_manhattan_accuracy_threshold": 517.3240356445312, | |
| "eval_allNLI-dev_manhattan_ap": 0.5527648330520126, | |
| "eval_allNLI-dev_manhattan_f1": 0.5900900900900902, | |
| "eval_allNLI-dev_manhattan_f1_threshold": 765.703857421875, | |
| "eval_allNLI-dev_manhattan_precision": 0.4833948339483395, | |
| "eval_allNLI-dev_manhattan_recall": 0.7572254335260116, | |
| "eval_allNLI-dev_max_accuracy": 0.705078125, | |
| "eval_allNLI-dev_max_accuracy_threshold": 858.6700439453125, | |
| "eval_allNLI-dev_max_ap": 0.5555367525973895, | |
| "eval_allNLI-dev_max_f1": 0.5992217898832685, | |
| "eval_allNLI-dev_max_f1_threshold": 765.703857421875, | |
| "eval_allNLI-dev_max_precision": 0.4980694980694981, | |
| "eval_allNLI-dev_max_recall": 0.8901734104046243, | |
| "eval_sequential_score": 0.7187237394606566, | |
| "eval_sts-test_pearson_cosine": 0.8999110842263669, | |
| "eval_sts-test_pearson_dot": 0.8830335713257453, | |
| "eval_sts-test_pearson_euclidean": 0.9123289195839425, | |
| "eval_sts-test_pearson_manhattan": 0.913131760355957, | |
| "eval_sts-test_pearson_max": 0.913131760355957, | |
| "eval_sts-test_spearman_cosine": 0.9102556931248936, | |
| "eval_sts-test_spearman_dot": 0.8782251567890449, | |
| "eval_sts-test_spearman_euclidean": 0.9078544746924153, | |
| "eval_sts-test_spearman_manhattan": 0.9084604962741777, | |
| "eval_sts-test_spearman_max": 0.9102556931248936, | |
| "eval_vitaminc-pairs_loss": 3.8443844318389893, | |
| "eval_vitaminc-pairs_runtime": 4.62, | |
| "eval_vitaminc-pairs_samples_per_second": 27.706, | |
| "eval_vitaminc-pairs_steps_per_second": 0.216, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.3508997429305913, | |
| "eval_negation-triplets_loss": 0.8706199526786804, | |
| "eval_negation-triplets_runtime": 3.4865, | |
| "eval_negation-triplets_samples_per_second": 36.714, | |
| "eval_negation-triplets_steps_per_second": 0.287, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.3508997429305913, | |
| "eval_scitail-pairs-pos_loss": 0.05496061593294144, | |
| "eval_scitail-pairs-pos_runtime": 2.8238, | |
| "eval_scitail-pairs-pos_samples_per_second": 45.328, | |
| "eval_scitail-pairs-pos_steps_per_second": 0.354, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.3508997429305913, | |
| "eval_scitail-pairs-qa_loss": 0.12511999905109406, | |
| "eval_scitail-pairs-qa_runtime": 2.2972, | |
| "eval_scitail-pairs-qa_samples_per_second": 55.721, | |
| "eval_scitail-pairs-qa_steps_per_second": 0.435, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.3508997429305913, | |
| "eval_xsum-pairs_loss": 0.03821578621864319, | |
| "eval_xsum-pairs_runtime": 3.1015, | |
| "eval_xsum-pairs_samples_per_second": 41.27, | |
| "eval_xsum-pairs_steps_per_second": 0.322, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.3508997429305913, | |
| "eval_sciq_pairs_loss": 0.08246304094791412, | |
| "eval_sciq_pairs_runtime": 4.8955, | |
| "eval_sciq_pairs_samples_per_second": 26.147, | |
| "eval_sciq_pairs_steps_per_second": 0.204, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.3508997429305913, | |
| "eval_qasc_pairs_loss": 0.15105530619621277, | |
| "eval_qasc_pairs_runtime": 2.2079, | |
| "eval_qasc_pairs_samples_per_second": 57.972, | |
| "eval_qasc_pairs_steps_per_second": 0.453, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.3508997429305913, | |
| "eval_openbookqa_pairs_loss": 0.7893790602684021, | |
| "eval_openbookqa_pairs_runtime": 2.2579, | |
| "eval_openbookqa_pairs_samples_per_second": 56.69, | |
| "eval_openbookqa_pairs_steps_per_second": 0.443, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.3508997429305913, | |
| "eval_msmarco_pairs_loss": 0.41756388545036316, | |
| "eval_msmarco_pairs_runtime": 2.2908, | |
| "eval_msmarco_pairs_samples_per_second": 55.877, | |
| "eval_msmarco_pairs_steps_per_second": 0.437, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.3508997429305913, | |
| "eval_nq_pairs_loss": 0.2524656355381012, | |
| "eval_nq_pairs_runtime": 3.6631, | |
| "eval_nq_pairs_samples_per_second": 34.943, | |
| "eval_nq_pairs_steps_per_second": 0.273, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.3508997429305913, | |
| "eval_trivia_pairs_loss": 0.062444861978292465, | |
| "eval_trivia_pairs_runtime": 3.115, | |
| "eval_trivia_pairs_samples_per_second": 41.091, | |
| "eval_trivia_pairs_steps_per_second": 0.321, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.3508997429305913, | |
| "eval_gooaq_pairs_loss": 0.12118146568536758, | |
| "eval_gooaq_pairs_runtime": 2.236, | |
| "eval_gooaq_pairs_samples_per_second": 57.245, | |
| "eval_gooaq_pairs_steps_per_second": 0.447, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.3508997429305913, | |
| "eval_paws-pos_loss": 0.020900897681713104, | |
| "eval_paws-pos_runtime": 2.2981, | |
| "eval_paws-pos_samples_per_second": 55.698, | |
| "eval_paws-pos_steps_per_second": 0.435, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.3508997429305913, | |
| "eval_global_dataset_loss": 0.38996464014053345, | |
| "eval_global_dataset_runtime": 8.423, | |
| "eval_global_dataset_samples_per_second": 38.585, | |
| "eval_global_dataset_steps_per_second": 0.356, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.352827763496144, | |
| "grad_norm": 412.62738037109375, | |
| "learning_rate": 1.0019305019305019e-05, | |
| "loss": 0.5365, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.35861182519280205, | |
| "grad_norm": 693.9551391601562, | |
| "learning_rate": 1.0193050193050193e-05, | |
| "loss": 0.6902, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.3643958868894602, | |
| "grad_norm": 739.5757446289062, | |
| "learning_rate": 1.0366795366795366e-05, | |
| "loss": 0.4105, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.37017994858611825, | |
| "grad_norm": 174.94882202148438, | |
| "learning_rate": 1.0540540540540541e-05, | |
| "loss": 0.2434, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.3759640102827764, | |
| "grad_norm": 5.913711071014404, | |
| "learning_rate": 1.0714285714285714e-05, | |
| "loss": 0.1521, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.38174807197943444, | |
| "grad_norm": 34.247581481933594, | |
| "learning_rate": 1.0888030888030889e-05, | |
| "loss": 0.1878, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.38753213367609257, | |
| "grad_norm": 50.26247787475586, | |
| "learning_rate": 1.1061776061776062e-05, | |
| "loss": 0.3544, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.39331619537275064, | |
| "grad_norm": 227.15286254882812, | |
| "learning_rate": 1.1235521235521235e-05, | |
| "loss": 0.1397, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.3991002570694087, | |
| "grad_norm": 95.1203384399414, | |
| "learning_rate": 1.140926640926641e-05, | |
| "loss": 0.2982, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.40102827763496146, | |
| "eval_Qnli-dev_cosine_accuracy": 0.685546875, | |
| "eval_Qnli-dev_cosine_accuracy_threshold": 0.715149462223053, | |
| "eval_Qnli-dev_cosine_ap": 0.7217325275419479, | |
| "eval_Qnli-dev_cosine_f1": 0.6854130052724077, | |
| "eval_Qnli-dev_cosine_f1_threshold": 0.658507227897644, | |
| "eval_Qnli-dev_cosine_precision": 0.5855855855855856, | |
| "eval_Qnli-dev_cosine_recall": 0.826271186440678, | |
| "eval_Qnli-dev_dot_accuracy": 0.669921875, | |
| "eval_Qnli-dev_dot_accuracy_threshold": 717.3741455078125, | |
| "eval_Qnli-dev_dot_ap": 0.6857348319823914, | |
| "eval_Qnli-dev_dot_f1": 0.6757679180887372, | |
| "eval_Qnli-dev_dot_f1_threshold": 627.569580078125, | |
| "eval_Qnli-dev_dot_precision": 0.5657142857142857, | |
| "eval_Qnli-dev_dot_recall": 0.8389830508474576, | |
| "eval_Qnli-dev_euclidean_accuracy": 0.68359375, | |
| "eval_Qnli-dev_euclidean_accuracy_threshold": 23.550647735595703, | |
| "eval_Qnli-dev_euclidean_ap": 0.7266186529674616, | |
| "eval_Qnli-dev_euclidean_f1": 0.6959578207381371, | |
| "eval_Qnli-dev_euclidean_f1_threshold": 25.634620666503906, | |
| "eval_Qnli-dev_euclidean_precision": 0.5945945945945946, | |
| "eval_Qnli-dev_euclidean_recall": 0.8389830508474576, | |
| "eval_Qnli-dev_manhattan_accuracy": 0.6796875, | |
| "eval_Qnli-dev_manhattan_accuracy_threshold": 681.6762084960938, | |
| "eval_Qnli-dev_manhattan_ap": 0.7264245248641615, | |
| "eval_Qnli-dev_manhattan_f1": 0.6938053097345133, | |
| "eval_Qnli-dev_manhattan_f1_threshold": 793.83837890625, | |
| "eval_Qnli-dev_manhattan_precision": 0.5957446808510638, | |
| "eval_Qnli-dev_manhattan_recall": 0.8305084745762712, | |
| "eval_Qnli-dev_max_accuracy": 0.685546875, | |
| "eval_Qnli-dev_max_accuracy_threshold": 717.3741455078125, | |
| "eval_Qnli-dev_max_ap": 0.7266186529674616, | |
| "eval_Qnli-dev_max_f1": 0.6959578207381371, | |
| "eval_Qnli-dev_max_f1_threshold": 793.83837890625, | |
| "eval_Qnli-dev_max_precision": 0.5957446808510638, | |
| "eval_Qnli-dev_max_recall": 0.8389830508474576, | |
| "eval_allNLI-dev_cosine_accuracy": 0.720703125, | |
| "eval_allNLI-dev_cosine_accuracy_threshold": 0.8279609680175781, | |
| "eval_allNLI-dev_cosine_ap": 0.5853174176053627, | |
| "eval_allNLI-dev_cosine_f1": 0.6189376443418013, | |
| "eval_allNLI-dev_cosine_f1_threshold": 0.7112143039703369, | |
| "eval_allNLI-dev_cosine_precision": 0.5153846153846153, | |
| "eval_allNLI-dev_cosine_recall": 0.7745664739884393, | |
| "eval_allNLI-dev_dot_accuracy": 0.708984375, | |
| "eval_allNLI-dev_dot_accuracy_threshold": 844.9638061523438, | |
| "eval_allNLI-dev_dot_ap": 0.5767480495013521, | |
| "eval_allNLI-dev_dot_f1": 0.620525059665871, | |
| "eval_allNLI-dev_dot_f1_threshold": 698.743408203125, | |
| "eval_allNLI-dev_dot_precision": 0.5284552845528455, | |
| "eval_allNLI-dev_dot_recall": 0.7514450867052023, | |
| "eval_allNLI-dev_euclidean_accuracy": 0.71875, | |
| "eval_allNLI-dev_euclidean_accuracy_threshold": 17.55453109741211, | |
| "eval_allNLI-dev_euclidean_ap": 0.5835774601767848, | |
| "eval_allNLI-dev_euclidean_f1": 0.6063569682151589, | |
| "eval_allNLI-dev_euclidean_f1_threshold": 22.958927154541016, | |
| "eval_allNLI-dev_euclidean_precision": 0.5254237288135594, | |
| "eval_allNLI-dev_euclidean_recall": 0.7167630057803468, | |
| "eval_allNLI-dev_manhattan_accuracy": 0.71875, | |
| "eval_allNLI-dev_manhattan_accuracy_threshold": 546.237060546875, | |
| "eval_allNLI-dev_manhattan_ap": 0.5834282573106326, | |
| "eval_allNLI-dev_manhattan_f1": 0.6093023255813954, | |
| "eval_allNLI-dev_manhattan_f1_threshold": 738.0977172851562, | |
| "eval_allNLI-dev_manhattan_precision": 0.5097276264591439, | |
| "eval_allNLI-dev_manhattan_recall": 0.7572254335260116, | |
| "eval_allNLI-dev_max_accuracy": 0.720703125, | |
| "eval_allNLI-dev_max_accuracy_threshold": 844.9638061523438, | |
| "eval_allNLI-dev_max_ap": 0.5853174176053627, | |
| "eval_allNLI-dev_max_f1": 0.620525059665871, | |
| "eval_allNLI-dev_max_f1_threshold": 738.0977172851562, | |
| "eval_allNLI-dev_max_precision": 0.5284552845528455, | |
| "eval_allNLI-dev_max_recall": 0.7745664739884393, | |
| "eval_sequential_score": 0.7266186529674616, | |
| "eval_sts-test_pearson_cosine": 0.9110444991658042, | |
| "eval_sts-test_pearson_dot": 0.9027007693117498, | |
| "eval_sts-test_pearson_euclidean": 0.9255983263878029, | |
| "eval_sts-test_pearson_manhattan": 0.925028098980626, | |
| "eval_sts-test_pearson_max": 0.9255983263878029, | |
| "eval_sts-test_spearman_cosine": 0.9211899030744385, | |
| "eval_sts-test_spearman_dot": 0.9034949858127107, | |
| "eval_sts-test_spearman_euclidean": 0.9209135640353466, | |
| "eval_sts-test_spearman_manhattan": 0.9206623304538951, | |
| "eval_sts-test_spearman_max": 0.9211899030744385, | |
| "eval_vitaminc-pairs_loss": 3.210468292236328, | |
| "eval_vitaminc-pairs_runtime": 4.4941, | |
| "eval_vitaminc-pairs_samples_per_second": 28.482, | |
| "eval_vitaminc-pairs_steps_per_second": 0.223, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.40102827763496146, | |
| "eval_negation-triplets_loss": 0.6793166995048523, | |
| "eval_negation-triplets_runtime": 3.2613, | |
| "eval_negation-triplets_samples_per_second": 39.249, | |
| "eval_negation-triplets_steps_per_second": 0.307, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.40102827763496146, | |
| "eval_scitail-pairs-pos_loss": 0.020910607650876045, | |
| "eval_scitail-pairs-pos_runtime": 2.6259, | |
| "eval_scitail-pairs-pos_samples_per_second": 48.745, | |
| "eval_scitail-pairs-pos_steps_per_second": 0.381, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.40102827763496146, | |
| "eval_scitail-pairs-qa_loss": 0.024287551641464233, | |
| "eval_scitail-pairs-qa_runtime": 2.2326, | |
| "eval_scitail-pairs-qa_samples_per_second": 57.331, | |
| "eval_scitail-pairs-qa_steps_per_second": 0.448, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.40102827763496146, | |
| "eval_xsum-pairs_loss": 0.028975818306207657, | |
| "eval_xsum-pairs_runtime": 3.0795, | |
| "eval_xsum-pairs_samples_per_second": 41.565, | |
| "eval_xsum-pairs_steps_per_second": 0.325, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.40102827763496146, | |
| "eval_sciq_pairs_loss": 0.08505217730998993, | |
| "eval_sciq_pairs_runtime": 4.7361, | |
| "eval_sciq_pairs_samples_per_second": 27.026, | |
| "eval_sciq_pairs_steps_per_second": 0.211, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.40102827763496146, | |
| "eval_qasc_pairs_loss": 0.1334802210330963, | |
| "eval_qasc_pairs_runtime": 2.0711, | |
| "eval_qasc_pairs_samples_per_second": 61.804, | |
| "eval_qasc_pairs_steps_per_second": 0.483, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.40102827763496146, | |
| "eval_openbookqa_pairs_loss": 0.4361065924167633, | |
| "eval_openbookqa_pairs_runtime": 2.2402, | |
| "eval_openbookqa_pairs_samples_per_second": 57.137, | |
| "eval_openbookqa_pairs_steps_per_second": 0.446, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.40102827763496146, | |
| "eval_msmarco_pairs_loss": 0.09543590992689133, | |
| "eval_msmarco_pairs_runtime": 2.2223, | |
| "eval_msmarco_pairs_samples_per_second": 57.599, | |
| "eval_msmarco_pairs_steps_per_second": 0.45, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.40102827763496146, | |
| "eval_nq_pairs_loss": 0.24286797642707825, | |
| "eval_nq_pairs_runtime": 3.617, | |
| "eval_nq_pairs_samples_per_second": 35.389, | |
| "eval_nq_pairs_steps_per_second": 0.276, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.40102827763496146, | |
| "eval_trivia_pairs_loss": 0.053676724433898926, | |
| "eval_trivia_pairs_runtime": 3.0737, | |
| "eval_trivia_pairs_samples_per_second": 41.644, | |
| "eval_trivia_pairs_steps_per_second": 0.325, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.40102827763496146, | |
| "eval_gooaq_pairs_loss": 0.07278460264205933, | |
| "eval_gooaq_pairs_runtime": 2.1378, | |
| "eval_gooaq_pairs_samples_per_second": 59.874, | |
| "eval_gooaq_pairs_steps_per_second": 0.468, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.40102827763496146, | |
| "eval_paws-pos_loss": 0.022397885099053383, | |
| "eval_paws-pos_runtime": 2.253, | |
| "eval_paws-pos_samples_per_second": 56.812, | |
| "eval_paws-pos_steps_per_second": 0.444, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.40102827763496146, | |
| "eval_global_dataset_loss": 0.250370055437088, | |
| "eval_global_dataset_runtime": 8.321, | |
| "eval_global_dataset_samples_per_second": 39.058, | |
| "eval_global_dataset_steps_per_second": 0.361, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.40488431876606684, | |
| "grad_norm": 10.200668334960938, | |
| "learning_rate": 1.1583011583011584e-05, | |
| "loss": 0.1875, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.4106683804627249, | |
| "grad_norm": 25.712194442749023, | |
| "learning_rate": 1.1756756756756757e-05, | |
| "loss": 0.169, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.41645244215938304, | |
| "grad_norm": 48.711666107177734, | |
| "learning_rate": 1.193050193050193e-05, | |
| "loss": 0.2341, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.4222365038560411, | |
| "grad_norm": 13.946932792663574, | |
| "learning_rate": 1.2104247104247105e-05, | |
| "loss": 0.1806, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 0.42802056555269924, | |
| "grad_norm": 6.094577312469482, | |
| "learning_rate": 1.227799227799228e-05, | |
| "loss": 0.2736, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 0.4338046272493573, | |
| "grad_norm": 6.224375247955322, | |
| "learning_rate": 1.2451737451737452e-05, | |
| "loss": 0.1772, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.43958868894601544, | |
| "grad_norm": 34.21040725708008, | |
| "learning_rate": 1.2625482625482625e-05, | |
| "loss": 0.131, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 0.4453727506426735, | |
| "grad_norm": 3.1434619426727295, | |
| "learning_rate": 1.2799227799227798e-05, | |
| "loss": 0.0825, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 0.45115681233933164, | |
| "grad_norm": 11.084990501403809, | |
| "learning_rate": 1.2972972972972975e-05, | |
| "loss": 0.2745, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.45115681233933164, | |
| "eval_Qnli-dev_cosine_accuracy": 0.681640625, | |
| "eval_Qnli-dev_cosine_accuracy_threshold": 0.7570502758026123, | |
| "eval_Qnli-dev_cosine_ap": 0.7209954522759622, | |
| "eval_Qnli-dev_cosine_f1": 0.6952054794520549, | |
| "eval_Qnli-dev_cosine_f1_threshold": 0.6451430320739746, | |
| "eval_Qnli-dev_cosine_precision": 0.5833333333333334, | |
| "eval_Qnli-dev_cosine_recall": 0.8601694915254238, | |
| "eval_Qnli-dev_dot_accuracy": 0.689453125, | |
| "eval_Qnli-dev_dot_accuracy_threshold": 690.3031616210938, | |
| "eval_Qnli-dev_dot_ap": 0.6892678092033461, | |
| "eval_Qnli-dev_dot_f1": 0.6910994764397904, | |
| "eval_Qnli-dev_dot_f1_threshold": 616.900390625, | |
| "eval_Qnli-dev_dot_precision": 0.5875370919881305, | |
| "eval_Qnli-dev_dot_recall": 0.8389830508474576, | |
| "eval_Qnli-dev_euclidean_accuracy": 0.685546875, | |
| "eval_Qnli-dev_euclidean_accuracy_threshold": 22.101879119873047, | |
| "eval_Qnli-dev_euclidean_ap": 0.725163804769685, | |
| "eval_Qnli-dev_euclidean_f1": 0.6919275123558485, | |
| "eval_Qnli-dev_euclidean_f1_threshold": 26.458560943603516, | |
| "eval_Qnli-dev_euclidean_precision": 0.5660377358490566, | |
| "eval_Qnli-dev_euclidean_recall": 0.8898305084745762, | |
| "eval_Qnli-dev_manhattan_accuracy": 0.685546875, | |
| "eval_Qnli-dev_manhattan_accuracy_threshold": 683.6396484375, | |
| "eval_Qnli-dev_manhattan_ap": 0.7242560148825838, | |
| "eval_Qnli-dev_manhattan_f1": 0.6919275123558485, | |
| "eval_Qnli-dev_manhattan_f1_threshold": 830.1368408203125, | |
| "eval_Qnli-dev_manhattan_precision": 0.5660377358490566, | |
| "eval_Qnli-dev_manhattan_recall": 0.8898305084745762, | |
| "eval_Qnli-dev_max_accuracy": 0.689453125, | |
| "eval_Qnli-dev_max_accuracy_threshold": 690.3031616210938, | |
| "eval_Qnli-dev_max_ap": 0.725163804769685, | |
| "eval_Qnli-dev_max_f1": 0.6952054794520549, | |
| "eval_Qnli-dev_max_f1_threshold": 830.1368408203125, | |
| "eval_Qnli-dev_max_precision": 0.5875370919881305, | |
| "eval_Qnli-dev_max_recall": 0.8898305084745762, | |
| "eval_allNLI-dev_cosine_accuracy": 0.72265625, | |
| "eval_allNLI-dev_cosine_accuracy_threshold": 0.8315098285675049, | |
| "eval_allNLI-dev_cosine_ap": 0.579111348491065, | |
| "eval_allNLI-dev_cosine_f1": 0.6045454545454545, | |
| "eval_allNLI-dev_cosine_f1_threshold": 0.7200944423675537, | |
| "eval_allNLI-dev_cosine_precision": 0.49812734082397003, | |
| "eval_allNLI-dev_cosine_recall": 0.7687861271676301, | |
| "eval_allNLI-dev_dot_accuracy": 0.7109375, | |
| "eval_allNLI-dev_dot_accuracy_threshold": 813.0138549804688, | |
| "eval_allNLI-dev_dot_ap": 0.5580493245872203, | |
| "eval_allNLI-dev_dot_f1": 0.6124401913875598, | |
| "eval_allNLI-dev_dot_f1_threshold": 703.9876708984375, | |
| "eval_allNLI-dev_dot_precision": 0.5224489795918368, | |
| "eval_allNLI-dev_dot_recall": 0.7398843930635838, | |
| "eval_allNLI-dev_euclidean_accuracy": 0.732421875, | |
| "eval_allNLI-dev_euclidean_accuracy_threshold": 18.200069427490234, | |
| "eval_allNLI-dev_euclidean_ap": 0.579946296621263, | |
| "eval_allNLI-dev_euclidean_f1": 0.6084905660377359, | |
| "eval_allNLI-dev_euclidean_f1_threshold": 22.654571533203125, | |
| "eval_allNLI-dev_euclidean_precision": 0.5139442231075697, | |
| "eval_allNLI-dev_euclidean_recall": 0.7456647398843931, | |
| "eval_allNLI-dev_manhattan_accuracy": 0.728515625, | |
| "eval_allNLI-dev_manhattan_accuracy_threshold": 571.3721313476562, | |
| "eval_allNLI-dev_manhattan_ap": 0.5810042621969582, | |
| "eval_allNLI-dev_manhattan_f1": 0.6150234741784038, | |
| "eval_allNLI-dev_manhattan_f1_threshold": 708.9692993164062, | |
| "eval_allNLI-dev_manhattan_precision": 0.5177865612648221, | |
| "eval_allNLI-dev_manhattan_recall": 0.7572254335260116, | |
| "eval_allNLI-dev_max_accuracy": 0.732421875, | |
| "eval_allNLI-dev_max_accuracy_threshold": 813.0138549804688, | |
| "eval_allNLI-dev_max_ap": 0.5810042621969582, | |
| "eval_allNLI-dev_max_f1": 0.6150234741784038, | |
| "eval_allNLI-dev_max_f1_threshold": 708.9692993164062, | |
| "eval_allNLI-dev_max_precision": 0.5224489795918368, | |
| "eval_allNLI-dev_max_recall": 0.7687861271676301, | |
| "eval_sequential_score": 0.725163804769685, | |
| "eval_sts-test_pearson_cosine": 0.9079355833912655, | |
| "eval_sts-test_pearson_dot": 0.8900251358598403, | |
| "eval_sts-test_pearson_euclidean": 0.9252346568054488, | |
| "eval_sts-test_pearson_manhattan": 0.9248533505767098, | |
| "eval_sts-test_pearson_max": 0.9252346568054488, | |
| "eval_sts-test_spearman_cosine": 0.91858412662614, | |
| "eval_sts-test_spearman_dot": 0.8870906501782496, | |
| "eval_sts-test_spearman_euclidean": 0.919076345750184, | |
| "eval_sts-test_spearman_manhattan": 0.9185557095751564, | |
| "eval_sts-test_spearman_max": 0.919076345750184, | |
| "eval_vitaminc-pairs_loss": 3.4478557109832764, | |
| "eval_vitaminc-pairs_runtime": 4.5288, | |
| "eval_vitaminc-pairs_samples_per_second": 28.264, | |
| "eval_vitaminc-pairs_steps_per_second": 0.221, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.45115681233933164, | |
| "eval_negation-triplets_loss": 0.636644184589386, | |
| "eval_negation-triplets_runtime": 3.402, | |
| "eval_negation-triplets_samples_per_second": 37.625, | |
| "eval_negation-triplets_steps_per_second": 0.294, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.45115681233933164, | |
| "eval_scitail-pairs-pos_loss": 0.026369448751211166, | |
| "eval_scitail-pairs-pos_runtime": 2.7594, | |
| "eval_scitail-pairs-pos_samples_per_second": 46.386, | |
| "eval_scitail-pairs-pos_steps_per_second": 0.362, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.45115681233933164, | |
| "eval_scitail-pairs-qa_loss": 0.023621411994099617, | |
| "eval_scitail-pairs-qa_runtime": 2.2993, | |
| "eval_scitail-pairs-qa_samples_per_second": 55.669, | |
| "eval_scitail-pairs-qa_steps_per_second": 0.435, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.45115681233933164, | |
| "eval_xsum-pairs_loss": 0.03148656710982323, | |
| "eval_xsum-pairs_runtime": 3.1312, | |
| "eval_xsum-pairs_samples_per_second": 40.878, | |
| "eval_xsum-pairs_steps_per_second": 0.319, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.45115681233933164, | |
| "eval_sciq_pairs_loss": 0.08541633188724518, | |
| "eval_sciq_pairs_runtime": 4.8654, | |
| "eval_sciq_pairs_samples_per_second": 26.308, | |
| "eval_sciq_pairs_steps_per_second": 0.206, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.45115681233933164, | |
| "eval_qasc_pairs_loss": 0.12370316684246063, | |
| "eval_qasc_pairs_runtime": 2.2058, | |
| "eval_qasc_pairs_samples_per_second": 58.029, | |
| "eval_qasc_pairs_steps_per_second": 0.453, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.45115681233933164, | |
| "eval_openbookqa_pairs_loss": 0.41401803493499756, | |
| "eval_openbookqa_pairs_runtime": 2.338, | |
| "eval_openbookqa_pairs_samples_per_second": 54.747, | |
| "eval_openbookqa_pairs_steps_per_second": 0.428, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.45115681233933164, | |
| "eval_msmarco_pairs_loss": 0.031952742487192154, | |
| "eval_msmarco_pairs_runtime": 2.3775, | |
| "eval_msmarco_pairs_samples_per_second": 53.839, | |
| "eval_msmarco_pairs_steps_per_second": 0.421, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.45115681233933164, | |
| "eval_nq_pairs_loss": 0.20804765820503235, | |
| "eval_nq_pairs_runtime": 3.7039, | |
| "eval_nq_pairs_samples_per_second": 34.558, | |
| "eval_nq_pairs_steps_per_second": 0.27, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.45115681233933164, | |
| "eval_trivia_pairs_loss": 0.053435422480106354, | |
| "eval_trivia_pairs_runtime": 3.1705, | |
| "eval_trivia_pairs_samples_per_second": 40.372, | |
| "eval_trivia_pairs_steps_per_second": 0.315, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.45115681233933164, | |
| "eval_gooaq_pairs_loss": 0.06229045242071152, | |
| "eval_gooaq_pairs_runtime": 2.2703, | |
| "eval_gooaq_pairs_samples_per_second": 56.38, | |
| "eval_gooaq_pairs_steps_per_second": 0.44, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.45115681233933164, | |
| "eval_paws-pos_loss": 0.018832631409168243, | |
| "eval_paws-pos_runtime": 2.3285, | |
| "eval_paws-pos_samples_per_second": 54.972, | |
| "eval_paws-pos_steps_per_second": 0.429, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.45115681233933164, | |
| "eval_global_dataset_loss": 0.24586248397827148, | |
| "eval_global_dataset_runtime": 8.4553, | |
| "eval_global_dataset_samples_per_second": 38.438, | |
| "eval_global_dataset_steps_per_second": 0.355, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 0.4569408740359897, | |
| "grad_norm": 12.329869270324707, | |
| "learning_rate": 1.3146718146718148e-05, | |
| "loss": 0.131, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 0.46272493573264784, | |
| "grad_norm": 19.341089248657227, | |
| "learning_rate": 1.332046332046332e-05, | |
| "loss": 0.2012, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.4685089974293059, | |
| "grad_norm": 2.5328197479248047, | |
| "learning_rate": 1.3494208494208494e-05, | |
| "loss": 0.0855, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 0.47429305912596403, | |
| "grad_norm": 7.81564474105835, | |
| "learning_rate": 1.3667953667953668e-05, | |
| "loss": 0.1181, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 0.4800771208226221, | |
| "grad_norm": 4.904431343078613, | |
| "learning_rate": 1.3841698841698843e-05, | |
| "loss": 0.0992, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 0.48586118251928023, | |
| "grad_norm": 2.585756540298462, | |
| "learning_rate": 1.4015444015444016e-05, | |
| "loss": 0.0375, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 0.4916452442159383, | |
| "grad_norm": 67.9809799194336, | |
| "learning_rate": 1.4189189189189189e-05, | |
| "loss": 0.1503, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.4974293059125964, | |
| "grad_norm": 8.545217514038086, | |
| "learning_rate": 1.4362934362934362e-05, | |
| "loss": 0.1239, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 0.5012853470437018, | |
| "eval_Qnli-dev_cosine_accuracy": 0.69140625, | |
| "eval_Qnli-dev_cosine_accuracy_threshold": 0.6921107769012451, | |
| "eval_Qnli-dev_cosine_ap": 0.7392224932862383, | |
| "eval_Qnli-dev_cosine_f1": 0.6944444444444444, | |
| "eval_Qnli-dev_cosine_f1_threshold": 0.6230607032775879, | |
| "eval_Qnli-dev_cosine_precision": 0.5882352941176471, | |
| "eval_Qnli-dev_cosine_recall": 0.847457627118644, | |
| "eval_Qnli-dev_dot_accuracy": 0.6875, | |
| "eval_Qnli-dev_dot_accuracy_threshold": 697.8347778320312, | |
| "eval_Qnli-dev_dot_ap": 0.7145385015601986, | |
| "eval_Qnli-dev_dot_f1": 0.6952054794520549, | |
| "eval_Qnli-dev_dot_f1_threshold": 581.857666015625, | |
| "eval_Qnli-dev_dot_precision": 0.5833333333333334, | |
| "eval_Qnli-dev_dot_recall": 0.8601694915254238, | |
| "eval_Qnli-dev_euclidean_accuracy": 0.689453125, | |
| "eval_Qnli-dev_euclidean_accuracy_threshold": 23.611530303955078, | |
| "eval_Qnli-dev_euclidean_ap": 0.7374932215461754, | |
| "eval_Qnli-dev_euclidean_f1": 0.697594501718213, | |
| "eval_Qnli-dev_euclidean_f1_threshold": 26.889122009277344, | |
| "eval_Qnli-dev_euclidean_precision": 0.5867052023121387, | |
| "eval_Qnli-dev_euclidean_recall": 0.8601694915254238, | |
| "eval_Qnli-dev_manhattan_accuracy": 0.6875, | |
| "eval_Qnli-dev_manhattan_accuracy_threshold": 731.7583618164062, | |
| "eval_Qnli-dev_manhattan_ap": 0.737286744524734, | |
| "eval_Qnli-dev_manhattan_f1": 0.6998223801065719, | |
| "eval_Qnli-dev_manhattan_f1_threshold": 821.986328125, | |
| "eval_Qnli-dev_manhattan_precision": 0.6024464831804281, | |
| "eval_Qnli-dev_manhattan_recall": 0.8347457627118644, | |
| "eval_Qnli-dev_max_accuracy": 0.69140625, | |
| "eval_Qnli-dev_max_accuracy_threshold": 731.7583618164062, | |
| "eval_Qnli-dev_max_ap": 0.7392224932862383, | |
| "eval_Qnli-dev_max_f1": 0.6998223801065719, | |
| "eval_Qnli-dev_max_f1_threshold": 821.986328125, | |
| "eval_Qnli-dev_max_precision": 0.6024464831804281, | |
| "eval_Qnli-dev_max_recall": 0.8601694915254238, | |
| "eval_allNLI-dev_cosine_accuracy": 0.716796875, | |
| "eval_allNLI-dev_cosine_accuracy_threshold": 0.842121958732605, | |
| "eval_allNLI-dev_cosine_ap": 0.5873681355801845, | |
| "eval_allNLI-dev_cosine_f1": 0.6113744075829384, | |
| "eval_allNLI-dev_cosine_f1_threshold": 0.7030021548271179, | |
| "eval_allNLI-dev_cosine_precision": 0.5180722891566265, | |
| "eval_allNLI-dev_cosine_recall": 0.7456647398843931, | |
| "eval_allNLI-dev_dot_accuracy": 0.720703125, | |
| "eval_allNLI-dev_dot_accuracy_threshold": 750.5220947265625, | |
| "eval_allNLI-dev_dot_ap": 0.5799311198634003, | |
| "eval_allNLI-dev_dot_f1": 0.6201923076923077, | |
| "eval_allNLI-dev_dot_f1_threshold": 695.0645751953125, | |
| "eval_allNLI-dev_dot_precision": 0.5308641975308642, | |
| "eval_allNLI-dev_dot_recall": 0.7456647398843931, | |
| "eval_allNLI-dev_euclidean_accuracy": 0.71875, | |
| "eval_allNLI-dev_euclidean_accuracy_threshold": 17.74507713317871, | |
| "eval_allNLI-dev_euclidean_ap": 0.5836049221301078, | |
| "eval_allNLI-dev_euclidean_f1": 0.6175771971496437, | |
| "eval_allNLI-dev_euclidean_f1_threshold": 23.809024810791016, | |
| "eval_allNLI-dev_euclidean_precision": 0.5241935483870968, | |
| "eval_allNLI-dev_euclidean_recall": 0.7514450867052023, | |
| "eval_allNLI-dev_manhattan_accuracy": 0.716796875, | |
| "eval_allNLI-dev_manhattan_accuracy_threshold": 556.4411010742188, | |
| "eval_allNLI-dev_manhattan_ap": 0.5840998034727366, | |
| "eval_allNLI-dev_manhattan_f1": 0.6146572104018913, | |
| "eval_allNLI-dev_manhattan_f1_threshold": 746.3275756835938, | |
| "eval_allNLI-dev_manhattan_precision": 0.52, | |
| "eval_allNLI-dev_manhattan_recall": 0.7514450867052023, | |
| "eval_allNLI-dev_max_accuracy": 0.720703125, | |
| "eval_allNLI-dev_max_accuracy_threshold": 750.5220947265625, | |
| "eval_allNLI-dev_max_ap": 0.5873681355801845, | |
| "eval_allNLI-dev_max_f1": 0.6201923076923077, | |
| "eval_allNLI-dev_max_f1_threshold": 746.3275756835938, | |
| "eval_allNLI-dev_max_precision": 0.5308641975308642, | |
| "eval_allNLI-dev_max_recall": 0.7514450867052023, | |
| "eval_sequential_score": 0.7392224932862383, | |
| "eval_sts-test_pearson_cosine": 0.9106833903275314, | |
| "eval_sts-test_pearson_dot": 0.8992284856445683, | |
| "eval_sts-test_pearson_euclidean": 0.9250592315631505, | |
| "eval_sts-test_pearson_manhattan": 0.9246839540884049, | |
| "eval_sts-test_pearson_max": 0.9250592315631505, | |
| "eval_sts-test_spearman_cosine": 0.920613820086547, | |
| "eval_sts-test_spearman_dot": 0.8966107650190875, | |
| "eval_sts-test_spearman_euclidean": 0.9195501721057961, | |
| "eval_sts-test_spearman_manhattan": 0.9193704510400483, | |
| "eval_sts-test_spearman_max": 0.920613820086547, | |
| "eval_vitaminc-pairs_loss": 2.7884156703948975, | |
| "eval_vitaminc-pairs_runtime": 4.4544, | |
| "eval_vitaminc-pairs_samples_per_second": 28.736, | |
| "eval_vitaminc-pairs_steps_per_second": 0.224, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.5012853470437018, | |
| "eval_negation-triplets_loss": 0.6253538727760315, | |
| "eval_negation-triplets_runtime": 3.2937, | |
| "eval_negation-triplets_samples_per_second": 38.862, | |
| "eval_negation-triplets_steps_per_second": 0.304, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.5012853470437018, | |
| "eval_scitail-pairs-pos_loss": 0.01729702576994896, | |
| "eval_scitail-pairs-pos_runtime": 2.6618, | |
| "eval_scitail-pairs-pos_samples_per_second": 48.088, | |
| "eval_scitail-pairs-pos_steps_per_second": 0.376, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.5012853470437018, | |
| "eval_scitail-pairs-qa_loss": 0.025592364370822906, | |
| "eval_scitail-pairs-qa_runtime": 2.204, | |
| "eval_scitail-pairs-qa_samples_per_second": 58.075, | |
| "eval_scitail-pairs-qa_steps_per_second": 0.454, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.5012853470437018, | |
| "eval_xsum-pairs_loss": 0.012989089824259281, | |
| "eval_xsum-pairs_runtime": 3.0884, | |
| "eval_xsum-pairs_samples_per_second": 41.445, | |
| "eval_xsum-pairs_steps_per_second": 0.324, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.5012853470437018, | |
| "eval_sciq_pairs_loss": 0.0916769877076149, | |
| "eval_sciq_pairs_runtime": 4.7492, | |
| "eval_sciq_pairs_samples_per_second": 26.952, | |
| "eval_sciq_pairs_steps_per_second": 0.211, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.5012853470437018, | |
| "eval_qasc_pairs_loss": 0.10503728687763214, | |
| "eval_qasc_pairs_runtime": 2.1169, | |
| "eval_qasc_pairs_samples_per_second": 60.466, | |
| "eval_qasc_pairs_steps_per_second": 0.472, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.5012853470437018, | |
| "eval_openbookqa_pairs_loss": 0.439107745885849, | |
| "eval_openbookqa_pairs_runtime": 2.1928, | |
| "eval_openbookqa_pairs_samples_per_second": 58.372, | |
| "eval_openbookqa_pairs_steps_per_second": 0.456, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.5012853470437018, | |
| "eval_msmarco_pairs_loss": 0.4504755735397339, | |
| "eval_msmarco_pairs_runtime": 2.2269, | |
| "eval_msmarco_pairs_samples_per_second": 57.48, | |
| "eval_msmarco_pairs_steps_per_second": 0.449, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.5012853470437018, | |
| "eval_nq_pairs_loss": 0.14197981357574463, | |
| "eval_nq_pairs_runtime": 3.633, | |
| "eval_nq_pairs_samples_per_second": 35.232, | |
| "eval_nq_pairs_steps_per_second": 0.275, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.5012853470437018, | |
| "eval_trivia_pairs_loss": 0.05544201284646988, | |
| "eval_trivia_pairs_runtime": 3.0694, | |
| "eval_trivia_pairs_samples_per_second": 41.701, | |
| "eval_trivia_pairs_steps_per_second": 0.326, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.5012853470437018, | |
| "eval_gooaq_pairs_loss": 0.0432414636015892, | |
| "eval_gooaq_pairs_runtime": 2.096, | |
| "eval_gooaq_pairs_samples_per_second": 61.069, | |
| "eval_gooaq_pairs_steps_per_second": 0.477, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.5012853470437018, | |
| "eval_paws-pos_loss": 0.02002798579633236, | |
| "eval_paws-pos_runtime": 2.2043, | |
| "eval_paws-pos_samples_per_second": 58.068, | |
| "eval_paws-pos_steps_per_second": 0.454, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.5012853470437018, | |
| "eval_global_dataset_loss": 0.2467704862356186, | |
| "eval_global_dataset_runtime": 8.3491, | |
| "eval_global_dataset_samples_per_second": 38.926, | |
| "eval_global_dataset_steps_per_second": 0.359, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.5032133676092545, | |
| "grad_norm": 7.293272495269775, | |
| "learning_rate": 1.4536679536679537e-05, | |
| "loss": 0.1849, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 0.5089974293059126, | |
| "grad_norm": 4.900815010070801, | |
| "learning_rate": 1.4710424710424711e-05, | |
| "loss": 0.0453, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 0.5147814910025706, | |
| "grad_norm": 130.37289428710938, | |
| "learning_rate": 1.4884169884169884e-05, | |
| "loss": 0.118, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 0.5205655526992288, | |
| "grad_norm": 139.6493377685547, | |
| "learning_rate": 1.4999944826454462e-05, | |
| "loss": 0.2285, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.5263496143958869, | |
| "grad_norm": 8.455353736877441, | |
| "learning_rate": 1.4999117241310476e-05, | |
| "loss": 0.072, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 0.532133676092545, | |
| "grad_norm": 37.64395523071289, | |
| "learning_rate": 1.4997296673048672e-05, | |
| "loss": 0.1449, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 0.537917737789203, | |
| "grad_norm": 35.890071868896484, | |
| "learning_rate": 1.4994483389525345e-05, | |
| "loss": 0.2179, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 0.5437017994858612, | |
| "grad_norm": 4.571140289306641, | |
| "learning_rate": 1.4990677804652856e-05, | |
| "loss": 0.0687, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 0.5494858611825193, | |
| "grad_norm": 4.596847057342529, | |
| "learning_rate": 1.4985880478338728e-05, | |
| "loss": 0.071, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.5514138817480719, | |
| "eval_Qnli-dev_cosine_accuracy": 0.6953125, | |
| "eval_Qnli-dev_cosine_accuracy_threshold": 0.7168486714363098, | |
| "eval_Qnli-dev_cosine_ap": 0.7438559407478242, | |
| "eval_Qnli-dev_cosine_f1": 0.7035175879396984, | |
| "eval_Qnli-dev_cosine_f1_threshold": 0.6049751043319702, | |
| "eval_Qnli-dev_cosine_precision": 0.5817174515235457, | |
| "eval_Qnli-dev_cosine_recall": 0.8898305084745762, | |
| "eval_Qnli-dev_dot_accuracy": 0.689453125, | |
| "eval_Qnli-dev_dot_accuracy_threshold": 723.0029296875, | |
| "eval_Qnli-dev_dot_ap": 0.715173938792371, | |
| "eval_Qnli-dev_dot_f1": 0.6942148760330578, | |
| "eval_Qnli-dev_dot_f1_threshold": 577.0636596679688, | |
| "eval_Qnli-dev_dot_precision": 0.5691056910569106, | |
| "eval_Qnli-dev_dot_recall": 0.8898305084745762, | |
| "eval_Qnli-dev_euclidean_accuracy": 0.693359375, | |
| "eval_Qnli-dev_euclidean_accuracy_threshold": 23.278547286987305, | |
| "eval_Qnli-dev_euclidean_ap": 0.7463211078255994, | |
| "eval_Qnli-dev_euclidean_f1": 0.6970227670753064, | |
| "eval_Qnli-dev_euclidean_f1_threshold": 27.014936447143555, | |
| "eval_Qnli-dev_euclidean_precision": 0.5940298507462687, | |
| "eval_Qnli-dev_euclidean_recall": 0.8432203389830508, | |
| "eval_Qnli-dev_manhattan_accuracy": 0.693359375, | |
| "eval_Qnli-dev_manhattan_accuracy_threshold": 722.95458984375, | |
| "eval_Qnli-dev_manhattan_ap": 0.7459762632488858, | |
| "eval_Qnli-dev_manhattan_f1": 0.7010309278350516, | |
| "eval_Qnli-dev_manhattan_f1_threshold": 851.98779296875, | |
| "eval_Qnli-dev_manhattan_precision": 0.5895953757225434, | |
| "eval_Qnli-dev_manhattan_recall": 0.864406779661017, | |
| "eval_Qnli-dev_max_accuracy": 0.6953125, | |
| "eval_Qnli-dev_max_accuracy_threshold": 723.0029296875, | |
| "eval_Qnli-dev_max_ap": 0.7463211078255994, | |
| "eval_Qnli-dev_max_f1": 0.7035175879396984, | |
| "eval_Qnli-dev_max_f1_threshold": 851.98779296875, | |
| "eval_Qnli-dev_max_precision": 0.5940298507462687, | |
| "eval_Qnli-dev_max_recall": 0.8898305084745762, | |
| "eval_allNLI-dev_cosine_accuracy": 0.716796875, | |
| "eval_allNLI-dev_cosine_accuracy_threshold": 0.8314377069473267, | |
| "eval_allNLI-dev_cosine_ap": 0.5914689901349341, | |
| "eval_allNLI-dev_cosine_f1": 0.6323529411764707, | |
| "eval_allNLI-dev_cosine_f1_threshold": 0.7238098978996277, | |
| "eval_allNLI-dev_cosine_precision": 0.548936170212766, | |
| "eval_allNLI-dev_cosine_recall": 0.7456647398843931, | |
| "eval_allNLI-dev_dot_accuracy": 0.71484375, | |
| "eval_allNLI-dev_dot_accuracy_threshold": 810.9595947265625, | |
| "eval_allNLI-dev_dot_ap": 0.5839807817494149, | |
| "eval_allNLI-dev_dot_f1": 0.6192468619246863, | |
| "eval_allNLI-dev_dot_f1_threshold": 654.8411254882812, | |
| "eval_allNLI-dev_dot_precision": 0.4852459016393443, | |
| "eval_allNLI-dev_dot_recall": 0.8554913294797688, | |
| "eval_allNLI-dev_euclidean_accuracy": 0.720703125, | |
| "eval_allNLI-dev_euclidean_accuracy_threshold": 18.951290130615234, | |
| "eval_allNLI-dev_euclidean_ap": 0.5919673869334235, | |
| "eval_allNLI-dev_euclidean_f1": 0.6300715990453462, | |
| "eval_allNLI-dev_euclidean_f1_threshold": 23.576871871948242, | |
| "eval_allNLI-dev_euclidean_precision": 0.5365853658536586, | |
| "eval_allNLI-dev_euclidean_recall": 0.7630057803468208, | |
| "eval_allNLI-dev_manhattan_accuracy": 0.72265625, | |
| "eval_allNLI-dev_manhattan_accuracy_threshold": 587.756103515625, | |
| "eval_allNLI-dev_manhattan_ap": 0.5937706599658953, | |
| "eval_allNLI-dev_manhattan_f1": 0.631578947368421, | |
| "eval_allNLI-dev_manhattan_f1_threshold": 734.0830078125, | |
| "eval_allNLI-dev_manhattan_precision": 0.5387755102040817, | |
| "eval_allNLI-dev_manhattan_recall": 0.7630057803468208, | |
| "eval_allNLI-dev_max_accuracy": 0.72265625, | |
| "eval_allNLI-dev_max_accuracy_threshold": 810.9595947265625, | |
| "eval_allNLI-dev_max_ap": 0.5937706599658953, | |
| "eval_allNLI-dev_max_f1": 0.6323529411764707, | |
| "eval_allNLI-dev_max_f1_threshold": 734.0830078125, | |
| "eval_allNLI-dev_max_precision": 0.548936170212766, | |
| "eval_allNLI-dev_max_recall": 0.8554913294797688, | |
| "eval_sequential_score": 0.7463211078255994, | |
| "eval_sts-test_pearson_cosine": 0.9045390674341153, | |
| "eval_sts-test_pearson_dot": 0.8941766888408116, | |
| "eval_sts-test_pearson_euclidean": 0.9222411970618043, | |
| "eval_sts-test_pearson_manhattan": 0.9215732932835548, | |
| "eval_sts-test_pearson_max": 0.9222411970618043, | |
| "eval_sts-test_spearman_cosine": 0.9179395294570584, | |
| "eval_sts-test_spearman_dot": 0.8987898829412804, | |
| "eval_sts-test_spearman_euclidean": 0.9189799962844084, | |
| "eval_sts-test_spearman_manhattan": 0.9181845426131761, | |
| "eval_sts-test_spearman_max": 0.9189799962844084, | |
| "eval_vitaminc-pairs_loss": 3.9507529735565186, | |
| "eval_vitaminc-pairs_runtime": 4.5266, | |
| "eval_vitaminc-pairs_samples_per_second": 28.277, | |
| "eval_vitaminc-pairs_steps_per_second": 0.221, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.5514138817480719, | |
| "eval_negation-triplets_loss": 0.6451127529144287, | |
| "eval_negation-triplets_runtime": 3.3941, | |
| "eval_negation-triplets_samples_per_second": 37.713, | |
| "eval_negation-triplets_steps_per_second": 0.295, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.5514138817480719, | |
| "eval_scitail-pairs-pos_loss": 0.018346838653087616, | |
| "eval_scitail-pairs-pos_runtime": 2.7523, | |
| "eval_scitail-pairs-pos_samples_per_second": 46.506, | |
| "eval_scitail-pairs-pos_steps_per_second": 0.363, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.5514138817480719, | |
| "eval_scitail-pairs-qa_loss": 0.021733948960900307, | |
| "eval_scitail-pairs-qa_runtime": 2.3512, | |
| "eval_scitail-pairs-qa_samples_per_second": 54.439, | |
| "eval_scitail-pairs-qa_steps_per_second": 0.425, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.5514138817480719, | |
| "eval_xsum-pairs_loss": 0.006104973144829273, | |
| "eval_xsum-pairs_runtime": 3.1114, | |
| "eval_xsum-pairs_samples_per_second": 41.139, | |
| "eval_xsum-pairs_steps_per_second": 0.321, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.5514138817480719, | |
| "eval_sciq_pairs_loss": 0.09672096371650696, | |
| "eval_sciq_pairs_runtime": 4.7885, | |
| "eval_sciq_pairs_samples_per_second": 26.731, | |
| "eval_sciq_pairs_steps_per_second": 0.209, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.5514138817480719, | |
| "eval_qasc_pairs_loss": 0.12122269719839096, | |
| "eval_qasc_pairs_runtime": 2.2255, | |
| "eval_qasc_pairs_samples_per_second": 57.515, | |
| "eval_qasc_pairs_steps_per_second": 0.449, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.5514138817480719, | |
| "eval_openbookqa_pairs_loss": 0.4551953971385956, | |
| "eval_openbookqa_pairs_runtime": 2.3841, | |
| "eval_openbookqa_pairs_samples_per_second": 53.688, | |
| "eval_openbookqa_pairs_steps_per_second": 0.419, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.5514138817480719, | |
| "eval_msmarco_pairs_loss": 0.029557496309280396, | |
| "eval_msmarco_pairs_runtime": 2.2846, | |
| "eval_msmarco_pairs_samples_per_second": 56.027, | |
| "eval_msmarco_pairs_steps_per_second": 0.438, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.5514138817480719, | |
| "eval_nq_pairs_loss": 0.16617023944854736, | |
| "eval_nq_pairs_runtime": 3.6481, | |
| "eval_nq_pairs_samples_per_second": 35.087, | |
| "eval_nq_pairs_steps_per_second": 0.274, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.5514138817480719, | |
| "eval_trivia_pairs_loss": 0.06793519854545593, | |
| "eval_trivia_pairs_runtime": 3.1576, | |
| "eval_trivia_pairs_samples_per_second": 40.537, | |
| "eval_trivia_pairs_steps_per_second": 0.317, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.5514138817480719, | |
| "eval_gooaq_pairs_loss": 0.05080220475792885, | |
| "eval_gooaq_pairs_runtime": 2.2483, | |
| "eval_gooaq_pairs_samples_per_second": 56.931, | |
| "eval_gooaq_pairs_steps_per_second": 0.445, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.5514138817480719, | |
| "eval_paws-pos_loss": 0.02359372191131115, | |
| "eval_paws-pos_runtime": 2.2829, | |
| "eval_paws-pos_samples_per_second": 56.07, | |
| "eval_paws-pos_steps_per_second": 0.438, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.5514138817480719, | |
| "eval_global_dataset_loss": 0.2985983192920685, | |
| "eval_global_dataset_runtime": 8.4308, | |
| "eval_global_dataset_samples_per_second": 38.549, | |
| "eval_global_dataset_steps_per_second": 0.356, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 0.5552699228791774, | |
| "grad_norm": 5.148947238922119, | |
| "learning_rate": 1.4980092116403276e-05, | |
| "loss": 0.0855, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.5610539845758354, | |
| "grad_norm": 13.268424034118652, | |
| "learning_rate": 1.4973313570475753e-05, | |
| "loss": 0.103, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 0.5668380462724936, | |
| "grad_norm": 6.761672496795654, | |
| "learning_rate": 1.4965545837869058e-05, | |
| "loss": 0.1555, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 0.5726221079691517, | |
| "grad_norm": 3.2632031440734863, | |
| "learning_rate": 1.4956790061432996e-05, | |
| "loss": 0.1069, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 0.5784061696658098, | |
| "grad_norm": 7.57591438293457, | |
| "learning_rate": 1.494704752938614e-05, | |
| "loss": 0.2014, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.5841902313624678, | |
| "grad_norm": 3.9399232864379883, | |
| "learning_rate": 1.49363196751263e-05, | |
| "loss": 0.1028, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 0.589974293059126, | |
| "grad_norm": 6.177419662475586, | |
| "learning_rate": 1.4924608077019618e-05, | |
| "loss": 0.2425, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 0.5957583547557841, | |
| "grad_norm": 3.4256973266601562, | |
| "learning_rate": 1.4911914458168361e-05, | |
| "loss": 0.1639, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 0.6015424164524421, | |
| "grad_norm": 3.3635151386260986, | |
| "learning_rate": 1.4898240686157396e-05, | |
| "loss": 0.099, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.6015424164524421, | |
| "eval_Qnli-dev_cosine_accuracy": 0.69921875, | |
| "eval_Qnli-dev_cosine_accuracy_threshold": 0.7296463251113892, | |
| "eval_Qnli-dev_cosine_ap": 0.7499128775998081, | |
| "eval_Qnli-dev_cosine_f1": 0.700525394045534, | |
| "eval_Qnli-dev_cosine_f1_threshold": 0.6390261650085449, | |
| "eval_Qnli-dev_cosine_precision": 0.5970149253731343, | |
| "eval_Qnli-dev_cosine_recall": 0.847457627118644, | |
| "eval_Qnli-dev_dot_accuracy": 0.685546875, | |
| "eval_Qnli-dev_dot_accuracy_threshold": 729.7808837890625, | |
| "eval_Qnli-dev_dot_ap": 0.7267806921631229, | |
| "eval_Qnli-dev_dot_f1": 0.6901408450704225, | |
| "eval_Qnli-dev_dot_f1_threshold": 627.7941284179688, | |
| "eval_Qnli-dev_dot_precision": 0.5903614457831325, | |
| "eval_Qnli-dev_dot_recall": 0.8305084745762712, | |
| "eval_Qnli-dev_euclidean_accuracy": 0.703125, | |
| "eval_Qnli-dev_euclidean_accuracy_threshold": 23.62594223022461, | |
| "eval_Qnli-dev_euclidean_ap": 0.7513960328625653, | |
| "eval_Qnli-dev_euclidean_f1": 0.6930693069306931, | |
| "eval_Qnli-dev_euclidean_f1_threshold": 27.77108383178711, | |
| "eval_Qnli-dev_euclidean_precision": 0.5675675675675675, | |
| "eval_Qnli-dev_euclidean_recall": 0.8898305084745762, | |
| "eval_Qnli-dev_manhattan_accuracy": 0.703125, | |
| "eval_Qnli-dev_manhattan_accuracy_threshold": 719.592529296875, | |
| "eval_Qnli-dev_manhattan_ap": 0.7518755537519403, | |
| "eval_Qnli-dev_manhattan_f1": 0.6963696369636965, | |
| "eval_Qnli-dev_manhattan_f1_threshold": 864.5909423828125, | |
| "eval_Qnli-dev_manhattan_precision": 0.5702702702702702, | |
| "eval_Qnli-dev_manhattan_recall": 0.8940677966101694, | |
| "eval_Qnli-dev_max_accuracy": 0.703125, | |
| "eval_Qnli-dev_max_accuracy_threshold": 729.7808837890625, | |
| "eval_Qnli-dev_max_ap": 0.7518755537519403, | |
| "eval_Qnli-dev_max_f1": 0.700525394045534, | |
| "eval_Qnli-dev_max_f1_threshold": 864.5909423828125, | |
| "eval_Qnli-dev_max_precision": 0.5970149253731343, | |
| "eval_Qnli-dev_max_recall": 0.8940677966101694, | |
| "eval_allNLI-dev_cosine_accuracy": 0.716796875, | |
| "eval_allNLI-dev_cosine_accuracy_threshold": 0.744063138961792, | |
| "eval_allNLI-dev_cosine_ap": 0.5872524450594222, | |
| "eval_allNLI-dev_cosine_f1": 0.6287128712871288, | |
| "eval_allNLI-dev_cosine_f1_threshold": 0.7192724943161011, | |
| "eval_allNLI-dev_cosine_precision": 0.5497835497835498, | |
| "eval_allNLI-dev_cosine_recall": 0.7341040462427746, | |
| "eval_allNLI-dev_dot_accuracy": 0.716796875, | |
| "eval_allNLI-dev_dot_accuracy_threshold": 740.0152587890625, | |
| "eval_allNLI-dev_dot_ap": 0.5749750816372319, | |
| "eval_allNLI-dev_dot_f1": 0.6201923076923077, | |
| "eval_allNLI-dev_dot_f1_threshold": 690.6895751953125, | |
| "eval_allNLI-dev_dot_precision": 0.5308641975308642, | |
| "eval_allNLI-dev_dot_recall": 0.7456647398843931, | |
| "eval_allNLI-dev_euclidean_accuracy": 0.71484375, | |
| "eval_allNLI-dev_euclidean_accuracy_threshold": 22.877256393432617, | |
| "eval_allNLI-dev_euclidean_ap": 0.587161106665968, | |
| "eval_allNLI-dev_euclidean_f1": 0.6275510204081632, | |
| "eval_allNLI-dev_euclidean_f1_threshold": 22.877256393432617, | |
| "eval_allNLI-dev_euclidean_precision": 0.5616438356164384, | |
| "eval_allNLI-dev_euclidean_recall": 0.7109826589595376, | |
| "eval_allNLI-dev_manhattan_accuracy": 0.71484375, | |
| "eval_allNLI-dev_manhattan_accuracy_threshold": 710.319091796875, | |
| "eval_allNLI-dev_manhattan_ap": 0.588942704782977, | |
| "eval_allNLI-dev_manhattan_f1": 0.6294416243654821, | |
| "eval_allNLI-dev_manhattan_f1_threshold": 713.067138671875, | |
| "eval_allNLI-dev_manhattan_precision": 0.5610859728506787, | |
| "eval_allNLI-dev_manhattan_recall": 0.7167630057803468, | |
| "eval_allNLI-dev_max_accuracy": 0.716796875, | |
| "eval_allNLI-dev_max_accuracy_threshold": 740.0152587890625, | |
| "eval_allNLI-dev_max_ap": 0.588942704782977, | |
| "eval_allNLI-dev_max_f1": 0.6294416243654821, | |
| "eval_allNLI-dev_max_f1_threshold": 713.067138671875, | |
| "eval_allNLI-dev_max_precision": 0.5616438356164384, | |
| "eval_allNLI-dev_max_recall": 0.7456647398843931, | |
| "eval_sequential_score": 0.7518755537519403, | |
| "eval_sts-test_pearson_cosine": 0.9043388760494557, | |
| "eval_sts-test_pearson_dot": 0.8886027908855041, | |
| "eval_sts-test_pearson_euclidean": 0.9241059779670555, | |
| "eval_sts-test_pearson_manhattan": 0.9234022512717295, | |
| "eval_sts-test_pearson_max": 0.9241059779670555, | |
| "eval_sts-test_spearman_cosine": 0.919076121993877, | |
| "eval_sts-test_spearman_dot": 0.8892642189440298, | |
| "eval_sts-test_spearman_euclidean": 0.9203842013643476, | |
| "eval_sts-test_spearman_manhattan": 0.9199692676687267, | |
| "eval_sts-test_spearman_max": 0.9203842013643476, | |
| "eval_vitaminc-pairs_loss": 3.3625028133392334, | |
| "eval_vitaminc-pairs_runtime": 4.4127, | |
| "eval_vitaminc-pairs_samples_per_second": 29.007, | |
| "eval_vitaminc-pairs_steps_per_second": 0.227, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.6015424164524421, | |
| "eval_negation-triplets_loss": 0.5403304696083069, | |
| "eval_negation-triplets_runtime": 3.3216, | |
| "eval_negation-triplets_samples_per_second": 38.535, | |
| "eval_negation-triplets_steps_per_second": 0.301, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.6015424164524421, | |
| "eval_scitail-pairs-pos_loss": 0.007662723306566477, | |
| "eval_scitail-pairs-pos_runtime": 2.6317, | |
| "eval_scitail-pairs-pos_samples_per_second": 48.637, | |
| "eval_scitail-pairs-pos_steps_per_second": 0.38, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.6015424164524421, | |
| "eval_scitail-pairs-qa_loss": 0.014555388130247593, | |
| "eval_scitail-pairs-qa_runtime": 2.1734, | |
| "eval_scitail-pairs-qa_samples_per_second": 58.895, | |
| "eval_scitail-pairs-qa_steps_per_second": 0.46, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.6015424164524421, | |
| "eval_xsum-pairs_loss": 0.004975470248609781, | |
| "eval_xsum-pairs_runtime": 3.0489, | |
| "eval_xsum-pairs_samples_per_second": 41.982, | |
| "eval_xsum-pairs_steps_per_second": 0.328, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.6015424164524421, | |
| "eval_sciq_pairs_loss": 0.08704940229654312, | |
| "eval_sciq_pairs_runtime": 4.7639, | |
| "eval_sciq_pairs_samples_per_second": 26.869, | |
| "eval_sciq_pairs_steps_per_second": 0.21, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.6015424164524421, | |
| "eval_qasc_pairs_loss": 0.15072381496429443, | |
| "eval_qasc_pairs_runtime": 2.0766, | |
| "eval_qasc_pairs_samples_per_second": 61.638, | |
| "eval_qasc_pairs_steps_per_second": 0.482, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.6015424164524421, | |
| "eval_openbookqa_pairs_loss": 0.4214599132537842, | |
| "eval_openbookqa_pairs_runtime": 2.1857, | |
| "eval_openbookqa_pairs_samples_per_second": 58.564, | |
| "eval_openbookqa_pairs_steps_per_second": 0.458, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.6015424164524421, | |
| "eval_msmarco_pairs_loss": 0.04896237328648567, | |
| "eval_msmarco_pairs_runtime": 2.2125, | |
| "eval_msmarco_pairs_samples_per_second": 57.853, | |
| "eval_msmarco_pairs_steps_per_second": 0.452, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.6015424164524421, | |
| "eval_nq_pairs_loss": 0.08789835125207901, | |
| "eval_nq_pairs_runtime": 3.6548, | |
| "eval_nq_pairs_samples_per_second": 35.022, | |
| "eval_nq_pairs_steps_per_second": 0.274, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.6015424164524421, | |
| "eval_trivia_pairs_loss": 0.04830099269747734, | |
| "eval_trivia_pairs_runtime": 3.0616, | |
| "eval_trivia_pairs_samples_per_second": 41.808, | |
| "eval_trivia_pairs_steps_per_second": 0.327, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.6015424164524421, | |
| "eval_gooaq_pairs_loss": 0.04206620529294014, | |
| "eval_gooaq_pairs_runtime": 2.1174, | |
| "eval_gooaq_pairs_samples_per_second": 60.452, | |
| "eval_gooaq_pairs_steps_per_second": 0.472, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.6015424164524421, | |
| "eval_paws-pos_loss": 0.024975091218948364, | |
| "eval_paws-pos_runtime": 2.201, | |
| "eval_paws-pos_samples_per_second": 58.154, | |
| "eval_paws-pos_steps_per_second": 0.454, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.6015424164524421, | |
| "eval_global_dataset_loss": 0.23649312555789948, | |
| "eval_global_dataset_runtime": 8.3828, | |
| "eval_global_dataset_samples_per_second": 38.77, | |
| "eval_global_dataset_steps_per_second": 0.358, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 0.6073264781491002, | |
| "grad_norm": 2.4958090782165527, | |
| "learning_rate": 1.4883588772779416e-05, | |
| "loss": 0.1218, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.6131105398457584, | |
| "grad_norm": 3.888019323348999, | |
| "learning_rate": 1.4867960873738955e-05, | |
| "loss": 0.1575, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 0.6188946015424165, | |
| "grad_norm": 5.112899303436279, | |
| "learning_rate": 1.485135928833522e-05, | |
| "loss": 0.178, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 0.6246786632390745, | |
| "grad_norm": 2.7898402214050293, | |
| "learning_rate": 1.4833786459123798e-05, | |
| "loss": 0.0777, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 0.6304627249357326, | |
| "grad_norm": 6.2421488761901855, | |
| "learning_rate": 1.4815244971557292e-05, | |
| "loss": 0.0696, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 0.6362467866323908, | |
| "grad_norm": 133.5791015625, | |
| "learning_rate": 1.4795737553604924e-05, | |
| "loss": 0.1206, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.6420308483290489, | |
| "grad_norm": 9.727315902709961, | |
| "learning_rate": 1.477526707535119e-05, | |
| "loss": 0.0926, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 0.6478149100257069, | |
| "grad_norm": 6.652628421783447, | |
| "learning_rate": 1.4753836548573564e-05, | |
| "loss": 0.1357, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 0.6516709511568124, | |
| "eval_Qnli-dev_cosine_accuracy": 0.703125, | |
| "eval_Qnli-dev_cosine_accuracy_threshold": 0.7029188275337219, | |
| "eval_Qnli-dev_cosine_ap": 0.7383384263937729, | |
| "eval_Qnli-dev_cosine_f1": 0.7160940325497287, | |
| "eval_Qnli-dev_cosine_f1_threshold": 0.6273617744445801, | |
| "eval_Qnli-dev_cosine_precision": 0.6246056782334385, | |
| "eval_Qnli-dev_cosine_recall": 0.8389830508474576, | |
| "eval_Qnli-dev_dot_accuracy": 0.69921875, | |
| "eval_Qnli-dev_dot_accuracy_threshold": 647.5323486328125, | |
| "eval_Qnli-dev_dot_ap": 0.7214295624305991, | |
| "eval_Qnli-dev_dot_f1": 0.6965517241379311, | |
| "eval_Qnli-dev_dot_f1_threshold": 570.4014282226562, | |
| "eval_Qnli-dev_dot_precision": 0.5872093023255814, | |
| "eval_Qnli-dev_dot_recall": 0.8559322033898306, | |
| "eval_Qnli-dev_euclidean_accuracy": 0.70703125, | |
| "eval_Qnli-dev_euclidean_accuracy_threshold": 23.107650756835938, | |
| "eval_Qnli-dev_euclidean_ap": 0.7395303199325498, | |
| "eval_Qnli-dev_euclidean_f1": 0.7045871559633027, | |
| "eval_Qnli-dev_euclidean_f1_threshold": 26.295127868652344, | |
| "eval_Qnli-dev_euclidean_precision": 0.6213592233009708, | |
| "eval_Qnli-dev_euclidean_recall": 0.8135593220338984, | |
| "eval_Qnli-dev_manhattan_accuracy": 0.701171875, | |
| "eval_Qnli-dev_manhattan_accuracy_threshold": 702.7601928710938, | |
| "eval_Qnli-dev_manhattan_ap": 0.7390235736074865, | |
| "eval_Qnli-dev_manhattan_f1": 0.7094474153297684, | |
| "eval_Qnli-dev_manhattan_f1_threshold": 828.67724609375, | |
| "eval_Qnli-dev_manhattan_precision": 0.6123076923076923, | |
| "eval_Qnli-dev_manhattan_recall": 0.8432203389830508, | |
| "eval_Qnli-dev_max_accuracy": 0.70703125, | |
| "eval_Qnli-dev_max_accuracy_threshold": 702.7601928710938, | |
| "eval_Qnli-dev_max_ap": 0.7395303199325498, | |
| "eval_Qnli-dev_max_f1": 0.7160940325497287, | |
| "eval_Qnli-dev_max_f1_threshold": 828.67724609375, | |
| "eval_Qnli-dev_max_precision": 0.6246056782334385, | |
| "eval_Qnli-dev_max_recall": 0.8559322033898306, | |
| "eval_allNLI-dev_cosine_accuracy": 0.716796875, | |
| "eval_allNLI-dev_cosine_accuracy_threshold": 0.8458511233329773, | |
| "eval_allNLI-dev_cosine_ap": 0.596148092754136, | |
| "eval_allNLI-dev_cosine_f1": 0.6157517899761337, | |
| "eval_allNLI-dev_cosine_f1_threshold": 0.7034751176834106, | |
| "eval_allNLI-dev_cosine_precision": 0.524390243902439, | |
| "eval_allNLI-dev_cosine_recall": 0.7456647398843931, | |
| "eval_allNLI-dev_dot_accuracy": 0.7109375, | |
| "eval_allNLI-dev_dot_accuracy_threshold": 726.9520263671875, | |
| "eval_allNLI-dev_dot_ap": 0.5809559752676249, | |
| "eval_allNLI-dev_dot_f1": 0.6153846153846154, | |
| "eval_allNLI-dev_dot_f1_threshold": 624.0245361328125, | |
| "eval_allNLI-dev_dot_precision": 0.488135593220339, | |
| "eval_allNLI-dev_dot_recall": 0.8323699421965318, | |
| "eval_allNLI-dev_euclidean_accuracy": 0.716796875, | |
| "eval_allNLI-dev_euclidean_accuracy_threshold": 17.03449249267578, | |
| "eval_allNLI-dev_euclidean_ap": 0.5966915737527797, | |
| "eval_allNLI-dev_euclidean_f1": 0.6265060240963856, | |
| "eval_allNLI-dev_euclidean_f1_threshold": 23.620059967041016, | |
| "eval_allNLI-dev_euclidean_precision": 0.5371900826446281, | |
| "eval_allNLI-dev_euclidean_recall": 0.7514450867052023, | |
| "eval_allNLI-dev_manhattan_accuracy": 0.716796875, | |
| "eval_allNLI-dev_manhattan_accuracy_threshold": 531.2730712890625, | |
| "eval_allNLI-dev_manhattan_ap": 0.598298778219752, | |
| "eval_allNLI-dev_manhattan_f1": 0.6231884057971014, | |
| "eval_allNLI-dev_manhattan_f1_threshold": 734.1275634765625, | |
| "eval_allNLI-dev_manhattan_precision": 0.5352697095435685, | |
| "eval_allNLI-dev_manhattan_recall": 0.7456647398843931, | |
| "eval_allNLI-dev_max_accuracy": 0.716796875, | |
| "eval_allNLI-dev_max_accuracy_threshold": 726.9520263671875, | |
| "eval_allNLI-dev_max_ap": 0.598298778219752, | |
| "eval_allNLI-dev_max_f1": 0.6265060240963856, | |
| "eval_allNLI-dev_max_f1_threshold": 734.1275634765625, | |
| "eval_allNLI-dev_max_precision": 0.5371900826446281, | |
| "eval_allNLI-dev_max_recall": 0.8323699421965318, | |
| "eval_sequential_score": 0.7395303199325498, | |
| "eval_sts-test_pearson_cosine": 0.9132656286571847, | |
| "eval_sts-test_pearson_dot": 0.900423684668929, | |
| "eval_sts-test_pearson_euclidean": 0.930599039345619, | |
| "eval_sts-test_pearson_manhattan": 0.9300256250399432, | |
| "eval_sts-test_pearson_max": 0.930599039345619, | |
| "eval_sts-test_spearman_cosine": 0.9248599987724085, | |
| "eval_sts-test_spearman_dot": 0.8983566907310115, | |
| "eval_sts-test_spearman_euclidean": 0.926492435285444, | |
| "eval_sts-test_spearman_manhattan": 0.9259926291445923, | |
| "eval_sts-test_spearman_max": 0.926492435285444, | |
| "eval_vitaminc-pairs_loss": 3.174186944961548, | |
| "eval_vitaminc-pairs_runtime": 4.5721, | |
| "eval_vitaminc-pairs_samples_per_second": 27.996, | |
| "eval_vitaminc-pairs_steps_per_second": 0.219, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.6516709511568124, | |
| "eval_negation-triplets_loss": 0.5640384554862976, | |
| "eval_negation-triplets_runtime": 3.4331, | |
| "eval_negation-triplets_samples_per_second": 37.285, | |
| "eval_negation-triplets_steps_per_second": 0.291, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.6516709511568124, | |
| "eval_scitail-pairs-pos_loss": 0.010583501309156418, | |
| "eval_scitail-pairs-pos_runtime": 2.746, | |
| "eval_scitail-pairs-pos_samples_per_second": 46.613, | |
| "eval_scitail-pairs-pos_steps_per_second": 0.364, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.6516709511568124, | |
| "eval_scitail-pairs-qa_loss": 0.0013579510850831866, | |
| "eval_scitail-pairs-qa_runtime": 2.2771, | |
| "eval_scitail-pairs-qa_samples_per_second": 56.211, | |
| "eval_scitail-pairs-qa_steps_per_second": 0.439, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.6516709511568124, | |
| "eval_xsum-pairs_loss": 0.005872013047337532, | |
| "eval_xsum-pairs_runtime": 3.1545, | |
| "eval_xsum-pairs_samples_per_second": 40.577, | |
| "eval_xsum-pairs_steps_per_second": 0.317, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.6516709511568124, | |
| "eval_sciq_pairs_loss": 0.0752023309469223, | |
| "eval_sciq_pairs_runtime": 4.8292, | |
| "eval_sciq_pairs_samples_per_second": 26.505, | |
| "eval_sciq_pairs_steps_per_second": 0.207, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.6516709511568124, | |
| "eval_qasc_pairs_loss": 0.1399243026971817, | |
| "eval_qasc_pairs_runtime": 2.1004, | |
| "eval_qasc_pairs_samples_per_second": 60.942, | |
| "eval_qasc_pairs_steps_per_second": 0.476, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.6516709511568124, | |
| "eval_openbookqa_pairs_loss": 0.4193750321865082, | |
| "eval_openbookqa_pairs_runtime": 2.2958, | |
| "eval_openbookqa_pairs_samples_per_second": 55.754, | |
| "eval_openbookqa_pairs_steps_per_second": 0.436, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.6516709511568124, | |
| "eval_msmarco_pairs_loss": 0.037960685789585114, | |
| "eval_msmarco_pairs_runtime": 2.2487, | |
| "eval_msmarco_pairs_samples_per_second": 56.922, | |
| "eval_msmarco_pairs_steps_per_second": 0.445, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.6516709511568124, | |
| "eval_nq_pairs_loss": 0.11820158362388611, | |
| "eval_nq_pairs_runtime": 3.6602, | |
| "eval_nq_pairs_samples_per_second": 34.971, | |
| "eval_nq_pairs_steps_per_second": 0.273, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.6516709511568124, | |
| "eval_trivia_pairs_loss": 0.028769271448254585, | |
| "eval_trivia_pairs_runtime": 3.1269, | |
| "eval_trivia_pairs_samples_per_second": 40.935, | |
| "eval_trivia_pairs_steps_per_second": 0.32, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.6516709511568124, | |
| "eval_gooaq_pairs_loss": 0.0657971203327179, | |
| "eval_gooaq_pairs_runtime": 2.1776, | |
| "eval_gooaq_pairs_samples_per_second": 58.779, | |
| "eval_gooaq_pairs_steps_per_second": 0.459, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.6516709511568124, | |
| "eval_paws-pos_loss": 0.02519122324883938, | |
| "eval_paws-pos_runtime": 2.3434, | |
| "eval_paws-pos_samples_per_second": 54.622, | |
| "eval_paws-pos_steps_per_second": 0.427, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.6516709511568124, | |
| "eval_global_dataset_loss": 0.19101615250110626, | |
| "eval_global_dataset_runtime": 8.4497, | |
| "eval_global_dataset_samples_per_second": 38.463, | |
| "eval_global_dataset_steps_per_second": 0.355, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 0.653598971722365, | |
| "grad_norm": 45.49017333984375, | |
| "learning_rate": 1.4731449126299404e-05, | |
| "loss": 0.1234, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 0.6593830334190232, | |
| "grad_norm": 7.209679126739502, | |
| "learning_rate": 1.4708108102342047e-05, | |
| "loss": 0.1056, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 0.6651670951156813, | |
| "grad_norm": 6.202587604522705, | |
| "learning_rate": 1.468381691081619e-05, | |
| "loss": 0.1525, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.6709511568123393, | |
| "grad_norm": 7.8031535148620605, | |
| "learning_rate": 1.4658579125632652e-05, | |
| "loss": 0.1425, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 0.6767352185089974, | |
| "grad_norm": 4.834568500518799, | |
| "learning_rate": 1.4632398459972538e-05, | |
| "loss": 0.1401, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 0.6825192802056556, | |
| "grad_norm": 5.427046298980713, | |
| "learning_rate": 1.4605278765740929e-05, | |
| "loss": 0.1271, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 0.6883033419023136, | |
| "grad_norm": 1.4662452936172485, | |
| "learning_rate": 1.457722403300017e-05, | |
| "loss": 0.0598, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 0.6940874035989717, | |
| "grad_norm": 2.993671417236328, | |
| "learning_rate": 1.4548238389382802e-05, | |
| "loss": 0.0681, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.6998714652956298, | |
| "grad_norm": 10.515313148498535, | |
| "learning_rate": 1.4518326099484293e-05, | |
| "loss": 0.1132, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 0.7017994858611826, | |
| "eval_Qnli-dev_cosine_accuracy": 0.69921875, | |
| "eval_Qnli-dev_cosine_accuracy_threshold": 0.7072766423225403, | |
| "eval_Qnli-dev_cosine_ap": 0.7245987545496009, | |
| "eval_Qnli-dev_cosine_f1": 0.7016274864376131, | |
| "eval_Qnli-dev_cosine_f1_threshold": 0.6138921976089478, | |
| "eval_Qnli-dev_cosine_precision": 0.61198738170347, | |
| "eval_Qnli-dev_cosine_recall": 0.8220338983050848, | |
| "eval_Qnli-dev_dot_accuracy": 0.68359375, | |
| "eval_Qnli-dev_dot_accuracy_threshold": 638.5958251953125, | |
| "eval_Qnli-dev_dot_ap": 0.6982591186827876, | |
| "eval_Qnli-dev_dot_f1": 0.6946308724832215, | |
| "eval_Qnli-dev_dot_f1_threshold": 531.0111083984375, | |
| "eval_Qnli-dev_dot_precision": 0.575, | |
| "eval_Qnli-dev_dot_recall": 0.8771186440677966, | |
| "eval_Qnli-dev_euclidean_accuracy": 0.70703125, | |
| "eval_Qnli-dev_euclidean_accuracy_threshold": 23.832258224487305, | |
| "eval_Qnli-dev_euclidean_ap": 0.7287738238939392, | |
| "eval_Qnli-dev_euclidean_f1": 0.6975088967971529, | |
| "eval_Qnli-dev_euclidean_f1_threshold": 27.3236083984375, | |
| "eval_Qnli-dev_euclidean_precision": 0.6012269938650306, | |
| "eval_Qnli-dev_euclidean_recall": 0.8305084745762712, | |
| "eval_Qnli-dev_manhattan_accuracy": 0.7109375, | |
| "eval_Qnli-dev_manhattan_accuracy_threshold": 737.96826171875, | |
| "eval_Qnli-dev_manhattan_ap": 0.7286289416450926, | |
| "eval_Qnli-dev_manhattan_f1": 0.7001795332136445, | |
| "eval_Qnli-dev_manhattan_f1_threshold": 847.50732421875, | |
| "eval_Qnli-dev_manhattan_precision": 0.6074766355140186, | |
| "eval_Qnli-dev_manhattan_recall": 0.826271186440678, | |
| "eval_Qnli-dev_max_accuracy": 0.7109375, | |
| "eval_Qnli-dev_max_accuracy_threshold": 737.96826171875, | |
| "eval_Qnli-dev_max_ap": 0.7287738238939392, | |
| "eval_Qnli-dev_max_f1": 0.7016274864376131, | |
| "eval_Qnli-dev_max_f1_threshold": 847.50732421875, | |
| "eval_Qnli-dev_max_precision": 0.61198738170347, | |
| "eval_Qnli-dev_max_recall": 0.8771186440677966, | |
| "eval_allNLI-dev_cosine_accuracy": 0.716796875, | |
| "eval_allNLI-dev_cosine_accuracy_threshold": 0.793137788772583, | |
| "eval_allNLI-dev_cosine_ap": 0.5846176608505069, | |
| "eval_allNLI-dev_cosine_f1": 0.6329670329670329, | |
| "eval_allNLI-dev_cosine_f1_threshold": 0.6309775114059448, | |
| "eval_allNLI-dev_cosine_precision": 0.5106382978723404, | |
| "eval_allNLI-dev_cosine_recall": 0.8323699421965318, | |
| "eval_allNLI-dev_dot_accuracy": 0.712890625, | |
| "eval_allNLI-dev_dot_accuracy_threshold": 794.71240234375, | |
| "eval_allNLI-dev_dot_ap": 0.5801144123861134, | |
| "eval_allNLI-dev_dot_f1": 0.6205357142857143, | |
| "eval_allNLI-dev_dot_f1_threshold": 594.3692626953125, | |
| "eval_allNLI-dev_dot_precision": 0.5054545454545455, | |
| "eval_allNLI-dev_dot_recall": 0.8034682080924855, | |
| "eval_allNLI-dev_euclidean_accuracy": 0.716796875, | |
| "eval_allNLI-dev_euclidean_accuracy_threshold": 18.532958984375, | |
| "eval_allNLI-dev_euclidean_ap": 0.5858640058497049, | |
| "eval_allNLI-dev_euclidean_f1": 0.6361556064073227, | |
| "eval_allNLI-dev_euclidean_f1_threshold": 25.700441360473633, | |
| "eval_allNLI-dev_euclidean_precision": 0.5265151515151515, | |
| "eval_allNLI-dev_euclidean_recall": 0.8034682080924855, | |
| "eval_allNLI-dev_manhattan_accuracy": 0.716796875, | |
| "eval_allNLI-dev_manhattan_accuracy_threshold": 577.34521484375, | |
| "eval_allNLI-dev_manhattan_ap": 0.585729227241877, | |
| "eval_allNLI-dev_manhattan_f1": 0.6339285714285714, | |
| "eval_allNLI-dev_manhattan_f1_threshold": 813.3930053710938, | |
| "eval_allNLI-dev_manhattan_precision": 0.5163636363636364, | |
| "eval_allNLI-dev_manhattan_recall": 0.8208092485549133, | |
| "eval_allNLI-dev_max_accuracy": 0.716796875, | |
| "eval_allNLI-dev_max_accuracy_threshold": 794.71240234375, | |
| "eval_allNLI-dev_max_ap": 0.5858640058497049, | |
| "eval_allNLI-dev_max_f1": 0.6361556064073227, | |
| "eval_allNLI-dev_max_f1_threshold": 813.3930053710938, | |
| "eval_allNLI-dev_max_precision": 0.5265151515151515, | |
| "eval_allNLI-dev_max_recall": 0.8323699421965318, | |
| "eval_sequential_score": 0.7287738238939392, | |
| "eval_sts-test_pearson_cosine": 0.9168734909278741, | |
| "eval_sts-test_pearson_dot": 0.9060722657495308, | |
| "eval_sts-test_pearson_euclidean": 0.9292788993338806, | |
| "eval_sts-test_pearson_manhattan": 0.9286667562357451, | |
| "eval_sts-test_pearson_max": 0.9292788993338806, | |
| "eval_sts-test_spearman_cosine": 0.923743231044386, | |
| "eval_sts-test_spearman_dot": 0.9006938701084388, | |
| "eval_sts-test_spearman_euclidean": 0.9245718006490479, | |
| "eval_sts-test_spearman_manhattan": 0.9238214114980368, | |
| "eval_sts-test_spearman_max": 0.9245718006490479, | |
| "eval_vitaminc-pairs_loss": 3.7763519287109375, | |
| "eval_vitaminc-pairs_runtime": 4.4308, | |
| "eval_vitaminc-pairs_samples_per_second": 28.889, | |
| "eval_vitaminc-pairs_steps_per_second": 0.226, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.7017994858611826, | |
| "eval_negation-triplets_loss": 0.5828245282173157, | |
| "eval_negation-triplets_runtime": 3.3268, | |
| "eval_negation-triplets_samples_per_second": 38.476, | |
| "eval_negation-triplets_steps_per_second": 0.301, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.7017994858611826, | |
| "eval_scitail-pairs-pos_loss": 0.011268602684140205, | |
| "eval_scitail-pairs-pos_runtime": 2.6437, | |
| "eval_scitail-pairs-pos_samples_per_second": 48.416, | |
| "eval_scitail-pairs-pos_steps_per_second": 0.378, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.7017994858611826, | |
| "eval_scitail-pairs-qa_loss": 0.0031084802467375994, | |
| "eval_scitail-pairs-qa_runtime": 2.2151, | |
| "eval_scitail-pairs-qa_samples_per_second": 57.785, | |
| "eval_scitail-pairs-qa_steps_per_second": 0.451, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.7017994858611826, | |
| "eval_xsum-pairs_loss": 0.01026161853224039, | |
| "eval_xsum-pairs_runtime": 3.0546, | |
| "eval_xsum-pairs_samples_per_second": 41.904, | |
| "eval_xsum-pairs_steps_per_second": 0.327, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.7017994858611826, | |
| "eval_sciq_pairs_loss": 0.07422904670238495, | |
| "eval_sciq_pairs_runtime": 4.7698, | |
| "eval_sciq_pairs_samples_per_second": 26.836, | |
| "eval_sciq_pairs_steps_per_second": 0.21, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.7017994858611826, | |
| "eval_qasc_pairs_loss": 0.06358247995376587, | |
| "eval_qasc_pairs_runtime": 2.0983, | |
| "eval_qasc_pairs_samples_per_second": 61.002, | |
| "eval_qasc_pairs_steps_per_second": 0.477, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.7017994858611826, | |
| "eval_openbookqa_pairs_loss": 0.34939464926719666, | |
| "eval_openbookqa_pairs_runtime": 2.1938, | |
| "eval_openbookqa_pairs_samples_per_second": 58.348, | |
| "eval_openbookqa_pairs_steps_per_second": 0.456, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.7017994858611826, | |
| "eval_msmarco_pairs_loss": 0.02596309781074524, | |
| "eval_msmarco_pairs_runtime": 2.2095, | |
| "eval_msmarco_pairs_samples_per_second": 57.931, | |
| "eval_msmarco_pairs_steps_per_second": 0.453, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.7017994858611826, | |
| "eval_nq_pairs_loss": 0.026090342551469803, | |
| "eval_nq_pairs_runtime": 3.6272, | |
| "eval_nq_pairs_samples_per_second": 35.289, | |
| "eval_nq_pairs_steps_per_second": 0.276, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.7017994858611826, | |
| "eval_trivia_pairs_loss": 0.019855637103319168, | |
| "eval_trivia_pairs_runtime": 3.0864, | |
| "eval_trivia_pairs_samples_per_second": 41.472, | |
| "eval_trivia_pairs_steps_per_second": 0.324, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.7017994858611826, | |
| "eval_gooaq_pairs_loss": 0.07295343279838562, | |
| "eval_gooaq_pairs_runtime": 2.1333, | |
| "eval_gooaq_pairs_samples_per_second": 60.0, | |
| "eval_gooaq_pairs_steps_per_second": 0.469, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.7017994858611826, | |
| "eval_paws-pos_loss": 0.02318389154970646, | |
| "eval_paws-pos_runtime": 2.2203, | |
| "eval_paws-pos_samples_per_second": 57.65, | |
| "eval_paws-pos_steps_per_second": 0.45, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.7017994858611826, | |
| "eval_global_dataset_loss": 0.24246200919151306, | |
| "eval_global_dataset_runtime": 8.3489, | |
| "eval_global_dataset_samples_per_second": 38.927, | |
| "eval_global_dataset_steps_per_second": 0.359, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 0.705655526992288, | |
| "grad_norm": 3.3271055221557617, | |
| "learning_rate": 1.448749156423558e-05, | |
| "loss": 0.0714, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 0.711439588688946, | |
| "grad_norm": 5.8299336433410645, | |
| "learning_rate": 1.4455739320255581e-05, | |
| "loss": 0.1816, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 0.7172236503856041, | |
| "grad_norm": 5.665037155151367, | |
| "learning_rate": 1.4423074039183725e-05, | |
| "loss": 0.1609, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 0.7230077120822622, | |
| "grad_norm": 3.767242193222046, | |
| "learning_rate": 1.4389500526992632e-05, | |
| "loss": 0.1123, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.7287917737789203, | |
| "grad_norm": 6.201554298400879, | |
| "learning_rate": 1.4355023723281005e-05, | |
| "loss": 0.1906, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 0.7345758354755784, | |
| "grad_norm": 6.10802698135376, | |
| "learning_rate": 1.4319648700546897e-05, | |
| "loss": 0.0689, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 0.7403598971722365, | |
| "grad_norm": 3.1739063262939453, | |
| "learning_rate": 1.4283380663441395e-05, | |
| "loss": 0.1897, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 0.7461439588688946, | |
| "grad_norm": 4.408695220947266, | |
| "learning_rate": 1.4246224948002864e-05, | |
| "loss": 0.1268, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 0.7519280205655527, | |
| "grad_norm": 4.3629350662231445, | |
| "learning_rate": 1.4208187020871878e-05, | |
| "loss": 0.1256, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.7519280205655527, | |
| "eval_Qnli-dev_cosine_accuracy": 0.71875, | |
| "eval_Qnli-dev_cosine_accuracy_threshold": 0.687468409538269, | |
| "eval_Qnli-dev_cosine_ap": 0.7382336178190824, | |
| "eval_Qnli-dev_cosine_f1": 0.7061224489795919, | |
| "eval_Qnli-dev_cosine_f1_threshold": 0.6693283319473267, | |
| "eval_Qnli-dev_cosine_precision": 0.6811023622047244, | |
| "eval_Qnli-dev_cosine_recall": 0.7330508474576272, | |
| "eval_Qnli-dev_dot_accuracy": 0.703125, | |
| "eval_Qnli-dev_dot_accuracy_threshold": 671.2976684570312, | |
| "eval_Qnli-dev_dot_ap": 0.7084828885948017, | |
| "eval_Qnli-dev_dot_f1": 0.6944908180300502, | |
| "eval_Qnli-dev_dot_f1_threshold": 556.4210205078125, | |
| "eval_Qnli-dev_dot_precision": 0.5730027548209367, | |
| "eval_Qnli-dev_dot_recall": 0.8813559322033898, | |
| "eval_Qnli-dev_euclidean_accuracy": 0.71875, | |
| "eval_Qnli-dev_euclidean_accuracy_threshold": 24.559717178344727, | |
| "eval_Qnli-dev_euclidean_ap": 0.739908359646434, | |
| "eval_Qnli-dev_euclidean_f1": 0.705223880597015, | |
| "eval_Qnli-dev_euclidean_f1_threshold": 25.997222900390625, | |
| "eval_Qnli-dev_euclidean_precision": 0.63, | |
| "eval_Qnli-dev_euclidean_recall": 0.8008474576271186, | |
| "eval_Qnli-dev_manhattan_accuracy": 0.716796875, | |
| "eval_Qnli-dev_manhattan_accuracy_threshold": 754.3623046875, | |
| "eval_Qnli-dev_manhattan_ap": 0.7392277169186197, | |
| "eval_Qnli-dev_manhattan_f1": 0.7067669172932329, | |
| "eval_Qnli-dev_manhattan_f1_threshold": 808.9886474609375, | |
| "eval_Qnli-dev_manhattan_precision": 0.6351351351351351, | |
| "eval_Qnli-dev_manhattan_recall": 0.7966101694915254, | |
| "eval_Qnli-dev_max_accuracy": 0.71875, | |
| "eval_Qnli-dev_max_accuracy_threshold": 754.3623046875, | |
| "eval_Qnli-dev_max_ap": 0.739908359646434, | |
| "eval_Qnli-dev_max_f1": 0.7067669172932329, | |
| "eval_Qnli-dev_max_f1_threshold": 808.9886474609375, | |
| "eval_Qnli-dev_max_precision": 0.6811023622047244, | |
| "eval_Qnli-dev_max_recall": 0.8813559322033898, | |
| "eval_allNLI-dev_cosine_accuracy": 0.720703125, | |
| "eval_allNLI-dev_cosine_accuracy_threshold": 0.8184170126914978, | |
| "eval_allNLI-dev_cosine_ap": 0.5969272075792788, | |
| "eval_allNLI-dev_cosine_f1": 0.6310904872389791, | |
| "eval_allNLI-dev_cosine_f1_threshold": 0.6732250452041626, | |
| "eval_allNLI-dev_cosine_precision": 0.5271317829457365, | |
| "eval_allNLI-dev_cosine_recall": 0.7861271676300579, | |
| "eval_allNLI-dev_dot_accuracy": 0.7265625, | |
| "eval_allNLI-dev_dot_accuracy_threshold": 728.9887084960938, | |
| "eval_allNLI-dev_dot_ap": 0.5950023435373721, | |
| "eval_allNLI-dev_dot_f1": 0.6263498920086393, | |
| "eval_allNLI-dev_dot_f1_threshold": 599.548583984375, | |
| "eval_allNLI-dev_dot_precision": 0.5, | |
| "eval_allNLI-dev_dot_recall": 0.838150289017341, | |
| "eval_allNLI-dev_euclidean_accuracy": 0.720703125, | |
| "eval_allNLI-dev_euclidean_accuracy_threshold": 19.761653900146484, | |
| "eval_allNLI-dev_euclidean_ap": 0.5951396590656841, | |
| "eval_allNLI-dev_euclidean_f1": 0.631336405529954, | |
| "eval_allNLI-dev_euclidean_f1_threshold": 24.669166564941406, | |
| "eval_allNLI-dev_euclidean_precision": 0.524904214559387, | |
| "eval_allNLI-dev_euclidean_recall": 0.791907514450867, | |
| "eval_allNLI-dev_manhattan_accuracy": 0.720703125, | |
| "eval_allNLI-dev_manhattan_accuracy_threshold": 621.3643798828125, | |
| "eval_allNLI-dev_manhattan_ap": 0.5949760338392235, | |
| "eval_allNLI-dev_manhattan_f1": 0.6415094339622641, | |
| "eval_allNLI-dev_manhattan_f1_threshold": 759.7421264648438, | |
| "eval_allNLI-dev_manhattan_precision": 0.5418326693227091, | |
| "eval_allNLI-dev_manhattan_recall": 0.7861271676300579, | |
| "eval_allNLI-dev_max_accuracy": 0.7265625, | |
| "eval_allNLI-dev_max_accuracy_threshold": 728.9887084960938, | |
| "eval_allNLI-dev_max_ap": 0.5969272075792788, | |
| "eval_allNLI-dev_max_f1": 0.6415094339622641, | |
| "eval_allNLI-dev_max_f1_threshold": 759.7421264648438, | |
| "eval_allNLI-dev_max_precision": 0.5418326693227091, | |
| "eval_allNLI-dev_max_recall": 0.838150289017341, | |
| "eval_sequential_score": 0.739908359646434, | |
| "eval_sts-test_pearson_cosine": 0.9171550146572618, | |
| "eval_sts-test_pearson_dot": 0.9078360871894045, | |
| "eval_sts-test_pearson_euclidean": 0.9297785604600608, | |
| "eval_sts-test_pearson_manhattan": 0.9289608083051153, | |
| "eval_sts-test_pearson_max": 0.9297785604600608, | |
| "eval_sts-test_spearman_cosine": 0.9273106673487241, | |
| "eval_sts-test_spearman_dot": 0.9100156921100516, | |
| "eval_sts-test_spearman_euclidean": 0.927380166057665, | |
| "eval_sts-test_spearman_manhattan": 0.9263161600668233, | |
| "eval_sts-test_spearman_max": 0.927380166057665, | |
| "eval_vitaminc-pairs_loss": 3.1046688556671143, | |
| "eval_vitaminc-pairs_runtime": 4.5292, | |
| "eval_vitaminc-pairs_samples_per_second": 28.261, | |
| "eval_vitaminc-pairs_steps_per_second": 0.221, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.7519280205655527, | |
| "eval_negation-triplets_loss": 0.575215756893158, | |
| "eval_negation-triplets_runtime": 3.4581, | |
| "eval_negation-triplets_samples_per_second": 37.014, | |
| "eval_negation-triplets_steps_per_second": 0.289, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.7519280205655527, | |
| "eval_scitail-pairs-pos_loss": 0.008397433906793594, | |
| "eval_scitail-pairs-pos_runtime": 2.7763, | |
| "eval_scitail-pairs-pos_samples_per_second": 46.104, | |
| "eval_scitail-pairs-pos_steps_per_second": 0.36, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.7519280205655527, | |
| "eval_scitail-pairs-qa_loss": 0.0021291954908519983, | |
| "eval_scitail-pairs-qa_runtime": 2.2643, | |
| "eval_scitail-pairs-qa_samples_per_second": 56.53, | |
| "eval_scitail-pairs-qa_steps_per_second": 0.442, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.7519280205655527, | |
| "eval_xsum-pairs_loss": 0.006353999022394419, | |
| "eval_xsum-pairs_runtime": 3.1689, | |
| "eval_xsum-pairs_samples_per_second": 40.393, | |
| "eval_xsum-pairs_steps_per_second": 0.316, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.7519280205655527, | |
| "eval_sciq_pairs_loss": 0.06674782186746597, | |
| "eval_sciq_pairs_runtime": 4.8334, | |
| "eval_sciq_pairs_samples_per_second": 26.483, | |
| "eval_sciq_pairs_steps_per_second": 0.207, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.7519280205655527, | |
| "eval_qasc_pairs_loss": 0.0450245626270771, | |
| "eval_qasc_pairs_runtime": 2.1448, | |
| "eval_qasc_pairs_samples_per_second": 59.68, | |
| "eval_qasc_pairs_steps_per_second": 0.466, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.7519280205655527, | |
| "eval_openbookqa_pairs_loss": 0.34624671936035156, | |
| "eval_openbookqa_pairs_runtime": 2.2437, | |
| "eval_openbookqa_pairs_samples_per_second": 57.048, | |
| "eval_openbookqa_pairs_steps_per_second": 0.446, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.7519280205655527, | |
| "eval_msmarco_pairs_loss": 0.0238041914999485, | |
| "eval_msmarco_pairs_runtime": 2.222, | |
| "eval_msmarco_pairs_samples_per_second": 57.606, | |
| "eval_msmarco_pairs_steps_per_second": 0.45, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.7519280205655527, | |
| "eval_nq_pairs_loss": 0.04486466199159622, | |
| "eval_nq_pairs_runtime": 3.632, | |
| "eval_nq_pairs_samples_per_second": 35.243, | |
| "eval_nq_pairs_steps_per_second": 0.275, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.7519280205655527, | |
| "eval_trivia_pairs_loss": 0.02883004955947399, | |
| "eval_trivia_pairs_runtime": 3.1146, | |
| "eval_trivia_pairs_samples_per_second": 41.097, | |
| "eval_trivia_pairs_steps_per_second": 0.321, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.7519280205655527, | |
| "eval_gooaq_pairs_loss": 0.05297350883483887, | |
| "eval_gooaq_pairs_runtime": 2.1744, | |
| "eval_gooaq_pairs_samples_per_second": 58.867, | |
| "eval_gooaq_pairs_steps_per_second": 0.46, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.7519280205655527, | |
| "eval_paws-pos_loss": 0.024740247055888176, | |
| "eval_paws-pos_runtime": 2.3365, | |
| "eval_paws-pos_samples_per_second": 54.782, | |
| "eval_paws-pos_steps_per_second": 0.428, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.7519280205655527, | |
| "eval_global_dataset_loss": 0.18419289588928223, | |
| "eval_global_dataset_runtime": 8.4479, | |
| "eval_global_dataset_samples_per_second": 38.471, | |
| "eval_global_dataset_steps_per_second": 0.355, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.7577120822622108, | |
| "grad_norm": 4.638159275054932, | |
| "learning_rate": 1.416927247848693e-05, | |
| "loss": 0.1201, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 0.7634961439588689, | |
| "grad_norm": 2.0510141849517822, | |
| "learning_rate": 1.4129487046261021e-05, | |
| "loss": 0.0995, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 0.769280205655527, | |
| "grad_norm": 3.916620969772339, | |
| "learning_rate": 1.4088836577739306e-05, | |
| "loss": 0.0825, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 0.7750642673521851, | |
| "grad_norm": 1.7055885791778564, | |
| "learning_rate": 1.4047327053737877e-05, | |
| "loss": 0.0778, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 0.7808483290488432, | |
| "grad_norm": 4.174785614013672, | |
| "learning_rate": 1.4004964581463804e-05, | |
| "loss": 0.0696, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.7866323907455013, | |
| "grad_norm": 4.478493690490723, | |
| "learning_rate": 1.3961755393616608e-05, | |
| "loss": 0.1332, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 0.7924164524421594, | |
| "grad_norm": 116.97537994384766, | |
| "learning_rate": 1.3917705847471257e-05, | |
| "loss": 0.0684, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 0.7982005141388174, | |
| "grad_norm": 7.2861433029174805, | |
| "learning_rate": 1.3872822423942823e-05, | |
| "loss": 0.2002, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 0.8020565552699229, | |
| "eval_Qnli-dev_cosine_accuracy": 0.71875, | |
| "eval_Qnli-dev_cosine_accuracy_threshold": 0.7014390230178833, | |
| "eval_Qnli-dev_cosine_ap": 0.7508144524343823, | |
| "eval_Qnli-dev_cosine_f1": 0.70703125, | |
| "eval_Qnli-dev_cosine_f1_threshold": 0.6568057537078857, | |
| "eval_Qnli-dev_cosine_precision": 0.6557971014492754, | |
| "eval_Qnli-dev_cosine_recall": 0.7669491525423728, | |
| "eval_Qnli-dev_dot_accuracy": 0.703125, | |
| "eval_Qnli-dev_dot_accuracy_threshold": 700.2732543945312, | |
| "eval_Qnli-dev_dot_ap": 0.7261430213596531, | |
| "eval_Qnli-dev_dot_f1": 0.7074380165289257, | |
| "eval_Qnli-dev_dot_f1_threshold": 554.1097412109375, | |
| "eval_Qnli-dev_dot_precision": 0.5799457994579946, | |
| "eval_Qnli-dev_dot_recall": 0.9067796610169492, | |
| "eval_Qnli-dev_euclidean_accuracy": 0.720703125, | |
| "eval_Qnli-dev_euclidean_accuracy_threshold": 23.440505981445312, | |
| "eval_Qnli-dev_euclidean_ap": 0.7517128865086153, | |
| "eval_Qnli-dev_euclidean_f1": 0.7045454545454546, | |
| "eval_Qnli-dev_euclidean_f1_threshold": 25.780654907226562, | |
| "eval_Qnli-dev_euclidean_precision": 0.636986301369863, | |
| "eval_Qnli-dev_euclidean_recall": 0.788135593220339, | |
| "eval_Qnli-dev_manhattan_accuracy": 0.720703125, | |
| "eval_Qnli-dev_manhattan_accuracy_threshold": 725.7116088867188, | |
| "eval_Qnli-dev_manhattan_ap": 0.7520524924838988, | |
| "eval_Qnli-dev_manhattan_f1": 0.7114624505928854, | |
| "eval_Qnli-dev_manhattan_f1_threshold": 791.727294921875, | |
| "eval_Qnli-dev_manhattan_precision": 0.6666666666666666, | |
| "eval_Qnli-dev_manhattan_recall": 0.7627118644067796, | |
| "eval_Qnli-dev_max_accuracy": 0.720703125, | |
| "eval_Qnli-dev_max_accuracy_threshold": 725.7116088867188, | |
| "eval_Qnli-dev_max_ap": 0.7520524924838988, | |
| "eval_Qnli-dev_max_f1": 0.7114624505928854, | |
| "eval_Qnli-dev_max_f1_threshold": 791.727294921875, | |
| "eval_Qnli-dev_max_precision": 0.6666666666666666, | |
| "eval_Qnli-dev_max_recall": 0.9067796610169492, | |
| "eval_allNLI-dev_cosine_accuracy": 0.7265625, | |
| "eval_allNLI-dev_cosine_accuracy_threshold": 0.7971610426902771, | |
| "eval_allNLI-dev_cosine_ap": 0.6029939953883172, | |
| "eval_allNLI-dev_cosine_f1": 0.6285714285714286, | |
| "eval_allNLI-dev_cosine_f1_threshold": 0.6836909055709839, | |
| "eval_allNLI-dev_cosine_precision": 0.5344129554655871, | |
| "eval_allNLI-dev_cosine_recall": 0.7630057803468208, | |
| "eval_allNLI-dev_dot_accuracy": 0.736328125, | |
| "eval_allNLI-dev_dot_accuracy_threshold": 756.0765380859375, | |
| "eval_allNLI-dev_dot_ap": 0.6094208828475387, | |
| "eval_allNLI-dev_dot_f1": 0.6301969365426696, | |
| "eval_allNLI-dev_dot_f1_threshold": 619.8289794921875, | |
| "eval_allNLI-dev_dot_precision": 0.5070422535211268, | |
| "eval_allNLI-dev_dot_recall": 0.8323699421965318, | |
| "eval_allNLI-dev_euclidean_accuracy": 0.724609375, | |
| "eval_allNLI-dev_euclidean_accuracy_threshold": 19.75946807861328, | |
| "eval_allNLI-dev_euclidean_ap": 0.5979771944382657, | |
| "eval_allNLI-dev_euclidean_f1": 0.6224256292906178, | |
| "eval_allNLI-dev_euclidean_f1_threshold": 24.92770767211914, | |
| "eval_allNLI-dev_euclidean_precision": 0.5151515151515151, | |
| "eval_allNLI-dev_euclidean_recall": 0.7861271676300579, | |
| "eval_allNLI-dev_manhattan_accuracy": 0.72265625, | |
| "eval_allNLI-dev_manhattan_accuracy_threshold": 590.9429321289062, | |
| "eval_allNLI-dev_manhattan_ap": 0.5975736439492428, | |
| "eval_allNLI-dev_manhattan_f1": 0.6261261261261263, | |
| "eval_allNLI-dev_manhattan_f1_threshold": 779.2568359375, | |
| "eval_allNLI-dev_manhattan_precision": 0.5129151291512916, | |
| "eval_allNLI-dev_manhattan_recall": 0.8034682080924855, | |
| "eval_allNLI-dev_max_accuracy": 0.736328125, | |
| "eval_allNLI-dev_max_accuracy_threshold": 756.0765380859375, | |
| "eval_allNLI-dev_max_ap": 0.6094208828475387, | |
| "eval_allNLI-dev_max_f1": 0.6301969365426696, | |
| "eval_allNLI-dev_max_f1_threshold": 779.2568359375, | |
| "eval_allNLI-dev_max_precision": 0.5344129554655871, | |
| "eval_allNLI-dev_max_recall": 0.8323699421965318, | |
| "eval_sequential_score": 0.7520524924838988, | |
| "eval_sts-test_pearson_cosine": 0.9150548197025481, | |
| "eval_sts-test_pearson_dot": 0.9041984235640055, | |
| "eval_sts-test_pearson_euclidean": 0.929928477203424, | |
| "eval_sts-test_pearson_manhattan": 0.9292292029223367, | |
| "eval_sts-test_pearson_max": 0.929928477203424, | |
| "eval_sts-test_spearman_cosine": 0.9273011800813092, | |
| "eval_sts-test_spearman_dot": 0.9082826547614081, | |
| "eval_sts-test_spearman_euclidean": 0.9273154557336931, | |
| "eval_sts-test_spearman_manhattan": 0.9263550489129724, | |
| "eval_sts-test_spearman_max": 0.9273154557336931, | |
| "eval_vitaminc-pairs_loss": 2.4590253829956055, | |
| "eval_vitaminc-pairs_runtime": 4.4297, | |
| "eval_vitaminc-pairs_samples_per_second": 28.896, | |
| "eval_vitaminc-pairs_steps_per_second": 0.226, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.8020565552699229, | |
| "eval_negation-triplets_loss": 0.5903122425079346, | |
| "eval_negation-triplets_runtime": 3.2993, | |
| "eval_negation-triplets_samples_per_second": 38.796, | |
| "eval_negation-triplets_steps_per_second": 0.303, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.8020565552699229, | |
| "eval_scitail-pairs-pos_loss": 0.008961405605077744, | |
| "eval_scitail-pairs-pos_runtime": 2.6102, | |
| "eval_scitail-pairs-pos_samples_per_second": 49.039, | |
| "eval_scitail-pairs-pos_steps_per_second": 0.383, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.8020565552699229, | |
| "eval_scitail-pairs-qa_loss": 0.0034254209604114294, | |
| "eval_scitail-pairs-qa_runtime": 2.2159, | |
| "eval_scitail-pairs-qa_samples_per_second": 57.763, | |
| "eval_scitail-pairs-qa_steps_per_second": 0.451, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.8020565552699229, | |
| "eval_xsum-pairs_loss": 0.00398410065099597, | |
| "eval_xsum-pairs_runtime": 3.0705, | |
| "eval_xsum-pairs_samples_per_second": 41.687, | |
| "eval_xsum-pairs_steps_per_second": 0.326, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.8020565552699229, | |
| "eval_sciq_pairs_loss": 0.07794960588216782, | |
| "eval_sciq_pairs_runtime": 4.7751, | |
| "eval_sciq_pairs_samples_per_second": 26.806, | |
| "eval_sciq_pairs_steps_per_second": 0.209, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.8020565552699229, | |
| "eval_qasc_pairs_loss": 0.08308541029691696, | |
| "eval_qasc_pairs_runtime": 2.0713, | |
| "eval_qasc_pairs_samples_per_second": 61.797, | |
| "eval_qasc_pairs_steps_per_second": 0.483, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.8020565552699229, | |
| "eval_openbookqa_pairs_loss": 0.36576414108276367, | |
| "eval_openbookqa_pairs_runtime": 2.2061, | |
| "eval_openbookqa_pairs_samples_per_second": 58.022, | |
| "eval_openbookqa_pairs_steps_per_second": 0.453, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.8020565552699229, | |
| "eval_msmarco_pairs_loss": 0.01674315147101879, | |
| "eval_msmarco_pairs_runtime": 2.2271, | |
| "eval_msmarco_pairs_samples_per_second": 57.473, | |
| "eval_msmarco_pairs_steps_per_second": 0.449, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.8020565552699229, | |
| "eval_nq_pairs_loss": 0.05656155198812485, | |
| "eval_nq_pairs_runtime": 3.648, | |
| "eval_nq_pairs_samples_per_second": 35.088, | |
| "eval_nq_pairs_steps_per_second": 0.274, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.8020565552699229, | |
| "eval_trivia_pairs_loss": 0.029444541782140732, | |
| "eval_trivia_pairs_runtime": 3.0578, | |
| "eval_trivia_pairs_samples_per_second": 41.86, | |
| "eval_trivia_pairs_steps_per_second": 0.327, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.8020565552699229, | |
| "eval_gooaq_pairs_loss": 0.0676947683095932, | |
| "eval_gooaq_pairs_runtime": 2.107, | |
| "eval_gooaq_pairs_samples_per_second": 60.75, | |
| "eval_gooaq_pairs_steps_per_second": 0.475, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.8020565552699229, | |
| "eval_paws-pos_loss": 0.02595301903784275, | |
| "eval_paws-pos_runtime": 2.2011, | |
| "eval_paws-pos_samples_per_second": 58.152, | |
| "eval_paws-pos_steps_per_second": 0.454, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.8020565552699229, | |
| "eval_global_dataset_loss": 0.14871826767921448, | |
| "eval_global_dataset_runtime": 8.3457, | |
| "eval_global_dataset_samples_per_second": 38.942, | |
| "eval_global_dataset_steps_per_second": 0.359, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.8039845758354756, | |
| "grad_norm": 2.4795706272125244, | |
| "learning_rate": 1.3827111726632967e-05, | |
| "loss": 0.0455, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 0.8097686375321337, | |
| "grad_norm": 8.65510082244873, | |
| "learning_rate": 1.3780580480858366e-05, | |
| "loss": 0.15, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.8155526992287918, | |
| "grad_norm": 6.457799434661865, | |
| "learning_rate": 1.373323553266123e-05, | |
| "loss": 0.1232, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 0.8213367609254498, | |
| "grad_norm": 4.571555137634277, | |
| "learning_rate": 1.3685083847802057e-05, | |
| "loss": 0.193, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 0.827120822622108, | |
| "grad_norm": 6.843802452087402, | |
| "learning_rate": 1.3636132510734774e-05, | |
| "loss": 0.1856, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 0.8329048843187661, | |
| "grad_norm": 10.976419448852539, | |
| "learning_rate": 1.3586388723564424e-05, | |
| "loss": 0.0898, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 0.8386889460154242, | |
| "grad_norm": 0.5411176085472107, | |
| "learning_rate": 1.3535859804987526e-05, | |
| "loss": 0.0242, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.8444730077120822, | |
| "grad_norm": 5.699883937835693, | |
| "learning_rate": 1.34845531892153e-05, | |
| "loss": 0.0693, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 0.8502570694087404, | |
| "grad_norm": 5.475403308868408, | |
| "learning_rate": 1.3432476424879881e-05, | |
| "loss": 0.1245, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 0.8521850899742931, | |
| "eval_Qnli-dev_cosine_accuracy": 0.720703125, | |
| "eval_Qnli-dev_cosine_accuracy_threshold": 0.6795437335968018, | |
| "eval_Qnli-dev_cosine_ap": 0.7439191659873163, | |
| "eval_Qnli-dev_cosine_f1": 0.7072691552062867, | |
| "eval_Qnli-dev_cosine_f1_threshold": 0.6607608795166016, | |
| "eval_Qnli-dev_cosine_precision": 0.6593406593406593, | |
| "eval_Qnli-dev_cosine_recall": 0.7627118644067796, | |
| "eval_Qnli-dev_dot_accuracy": 0.701171875, | |
| "eval_Qnli-dev_dot_accuracy_threshold": 683.7807006835938, | |
| "eval_Qnli-dev_dot_ap": 0.7138952944956131, | |
| "eval_Qnli-dev_dot_f1": 0.6950092421441774, | |
| "eval_Qnli-dev_dot_f1_threshold": 608.1612548828125, | |
| "eval_Qnli-dev_dot_precision": 0.6163934426229508, | |
| "eval_Qnli-dev_dot_recall": 0.7966101694915254, | |
| "eval_Qnli-dev_euclidean_accuracy": 0.71875, | |
| "eval_Qnli-dev_euclidean_accuracy_threshold": 23.998138427734375, | |
| "eval_Qnli-dev_euclidean_ap": 0.7473177380956941, | |
| "eval_Qnli-dev_euclidean_f1": 0.7063339731285989, | |
| "eval_Qnli-dev_euclidean_f1_threshold": 25.749244689941406, | |
| "eval_Qnli-dev_euclidean_precision": 0.6456140350877193, | |
| "eval_Qnli-dev_euclidean_recall": 0.7796610169491526, | |
| "eval_Qnli-dev_manhattan_accuracy": 0.720703125, | |
| "eval_Qnli-dev_manhattan_accuracy_threshold": 744.572509765625, | |
| "eval_Qnli-dev_manhattan_ap": 0.7478733392923025, | |
| "eval_Qnli-dev_manhattan_f1": 0.7135506003430532, | |
| "eval_Qnli-dev_manhattan_f1_threshold": 845.7078857421875, | |
| "eval_Qnli-dev_manhattan_precision": 0.5994236311239193, | |
| "eval_Qnli-dev_manhattan_recall": 0.8813559322033898, | |
| "eval_Qnli-dev_max_accuracy": 0.720703125, | |
| "eval_Qnli-dev_max_accuracy_threshold": 744.572509765625, | |
| "eval_Qnli-dev_max_ap": 0.7478733392923025, | |
| "eval_Qnli-dev_max_f1": 0.7135506003430532, | |
| "eval_Qnli-dev_max_f1_threshold": 845.7078857421875, | |
| "eval_Qnli-dev_max_precision": 0.6593406593406593, | |
| "eval_Qnli-dev_max_recall": 0.8813559322033898, | |
| "eval_allNLI-dev_cosine_accuracy": 0.7265625, | |
| "eval_allNLI-dev_cosine_accuracy_threshold": 0.7983643412590027, | |
| "eval_allNLI-dev_cosine_ap": 0.6021691080084255, | |
| "eval_allNLI-dev_cosine_f1": 0.6365795724465557, | |
| "eval_allNLI-dev_cosine_f1_threshold": 0.6869878172874451, | |
| "eval_allNLI-dev_cosine_precision": 0.5403225806451613, | |
| "eval_allNLI-dev_cosine_recall": 0.7745664739884393, | |
| "eval_allNLI-dev_dot_accuracy": 0.728515625, | |
| "eval_allNLI-dev_dot_accuracy_threshold": 743.5009765625, | |
| "eval_allNLI-dev_dot_ap": 0.6046964324835868, | |
| "eval_allNLI-dev_dot_f1": 0.6241457858769931, | |
| "eval_allNLI-dev_dot_f1_threshold": 630.758056640625, | |
| "eval_allNLI-dev_dot_precision": 0.5150375939849624, | |
| "eval_allNLI-dev_dot_recall": 0.791907514450867, | |
| "eval_allNLI-dev_euclidean_accuracy": 0.724609375, | |
| "eval_allNLI-dev_euclidean_accuracy_threshold": 19.04987144470215, | |
| "eval_allNLI-dev_euclidean_ap": 0.5993095223669318, | |
| "eval_allNLI-dev_euclidean_f1": 0.6382978723404256, | |
| "eval_allNLI-dev_euclidean_f1_threshold": 24.33264923095703, | |
| "eval_allNLI-dev_euclidean_precision": 0.54, | |
| "eval_allNLI-dev_euclidean_recall": 0.7803468208092486, | |
| "eval_allNLI-dev_manhattan_accuracy": 0.72265625, | |
| "eval_allNLI-dev_manhattan_accuracy_threshold": 599.6419677734375, | |
| "eval_allNLI-dev_manhattan_ap": 0.6009156396302207, | |
| "eval_allNLI-dev_manhattan_f1": 0.6400000000000001, | |
| "eval_allNLI-dev_manhattan_f1_threshold": 763.5076904296875, | |
| "eval_allNLI-dev_manhattan_precision": 0.5396825396825397, | |
| "eval_allNLI-dev_manhattan_recall": 0.7861271676300579, | |
| "eval_allNLI-dev_max_accuracy": 0.728515625, | |
| "eval_allNLI-dev_max_accuracy_threshold": 743.5009765625, | |
| "eval_allNLI-dev_max_ap": 0.6046964324835868, | |
| "eval_allNLI-dev_max_f1": 0.6400000000000001, | |
| "eval_allNLI-dev_max_f1_threshold": 763.5076904296875, | |
| "eval_allNLI-dev_max_precision": 0.5403225806451613, | |
| "eval_allNLI-dev_max_recall": 0.791907514450867, | |
| "eval_sequential_score": 0.7478733392923025, | |
| "eval_sts-test_pearson_cosine": 0.9138402547612807, | |
| "eval_sts-test_pearson_dot": 0.9059823959468752, | |
| "eval_sts-test_pearson_euclidean": 0.9303888224436159, | |
| "eval_sts-test_pearson_manhattan": 0.9299076277394032, | |
| "eval_sts-test_pearson_max": 0.9303888224436159, | |
| "eval_sts-test_spearman_cosine": 0.926297498790823, | |
| "eval_sts-test_spearman_dot": 0.9113238162317836, | |
| "eval_sts-test_spearman_euclidean": 0.9254560408228805, | |
| "eval_sts-test_spearman_manhattan": 0.9249534841574552, | |
| "eval_sts-test_spearman_max": 0.926297498790823, | |
| "eval_vitaminc-pairs_loss": 2.7577602863311768, | |
| "eval_vitaminc-pairs_runtime": 4.5823, | |
| "eval_vitaminc-pairs_samples_per_second": 27.933, | |
| "eval_vitaminc-pairs_steps_per_second": 0.218, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.8521850899742931, | |
| "eval_negation-triplets_loss": 0.5208465456962585, | |
| "eval_negation-triplets_runtime": 3.4824, | |
| "eval_negation-triplets_samples_per_second": 36.756, | |
| "eval_negation-triplets_steps_per_second": 0.287, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.8521850899742931, | |
| "eval_scitail-pairs-pos_loss": 0.00855955295264721, | |
| "eval_scitail-pairs-pos_runtime": 2.7579, | |
| "eval_scitail-pairs-pos_samples_per_second": 46.413, | |
| "eval_scitail-pairs-pos_steps_per_second": 0.363, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.8521850899742931, | |
| "eval_scitail-pairs-qa_loss": 0.0023300235625356436, | |
| "eval_scitail-pairs-qa_runtime": 2.3124, | |
| "eval_scitail-pairs-qa_samples_per_second": 55.353, | |
| "eval_scitail-pairs-qa_steps_per_second": 0.432, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.8521850899742931, | |
| "eval_xsum-pairs_loss": 0.006901867687702179, | |
| "eval_xsum-pairs_runtime": 3.212, | |
| "eval_xsum-pairs_samples_per_second": 39.851, | |
| "eval_xsum-pairs_steps_per_second": 0.311, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.8521850899742931, | |
| "eval_sciq_pairs_loss": 0.0706043541431427, | |
| "eval_sciq_pairs_runtime": 4.8786, | |
| "eval_sciq_pairs_samples_per_second": 26.237, | |
| "eval_sciq_pairs_steps_per_second": 0.205, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.8521850899742931, | |
| "eval_qasc_pairs_loss": 0.08733583986759186, | |
| "eval_qasc_pairs_runtime": 2.1645, | |
| "eval_qasc_pairs_samples_per_second": 59.137, | |
| "eval_qasc_pairs_steps_per_second": 0.462, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.8521850899742931, | |
| "eval_openbookqa_pairs_loss": 0.3373841345310211, | |
| "eval_openbookqa_pairs_runtime": 2.2848, | |
| "eval_openbookqa_pairs_samples_per_second": 56.022, | |
| "eval_openbookqa_pairs_steps_per_second": 0.438, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.8521850899742931, | |
| "eval_msmarco_pairs_loss": 0.01373821310698986, | |
| "eval_msmarco_pairs_runtime": 2.3098, | |
| "eval_msmarco_pairs_samples_per_second": 55.417, | |
| "eval_msmarco_pairs_steps_per_second": 0.433, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.8521850899742931, | |
| "eval_nq_pairs_loss": 0.04490604251623154, | |
| "eval_nq_pairs_runtime": 3.7868, | |
| "eval_nq_pairs_samples_per_second": 33.802, | |
| "eval_nq_pairs_steps_per_second": 0.264, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.8521850899742931, | |
| "eval_trivia_pairs_loss": 0.01734539307653904, | |
| "eval_trivia_pairs_runtime": 3.1386, | |
| "eval_trivia_pairs_samples_per_second": 40.783, | |
| "eval_trivia_pairs_steps_per_second": 0.319, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.8521850899742931, | |
| "eval_gooaq_pairs_loss": 0.06459866464138031, | |
| "eval_gooaq_pairs_runtime": 2.2089, | |
| "eval_gooaq_pairs_samples_per_second": 57.948, | |
| "eval_gooaq_pairs_steps_per_second": 0.453, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.8521850899742931, | |
| "eval_paws-pos_loss": 0.025955231860280037, | |
| "eval_paws-pos_runtime": 2.2617, | |
| "eval_paws-pos_samples_per_second": 56.594, | |
| "eval_paws-pos_steps_per_second": 0.442, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.8521850899742931, | |
| "eval_global_dataset_loss": 0.145456463098526, | |
| "eval_global_dataset_runtime": 8.4881, | |
| "eval_global_dataset_samples_per_second": 38.289, | |
| "eval_global_dataset_steps_per_second": 0.353, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 0.8560411311053985, | |
| "grad_norm": 6.109774589538574, | |
| "learning_rate": 1.3379637173923707e-05, | |
| "loss": 0.114, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 0.8618251928020566, | |
| "grad_norm": 3.212393283843994, | |
| "learning_rate": 1.3326043210472238e-05, | |
| "loss": 0.0713, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 0.8676092544987146, | |
| "grad_norm": 3.1123602390289307, | |
| "learning_rate": 1.3271702419690156e-05, | |
| "loss": 0.1336, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.8733933161953727, | |
| "grad_norm": 2.0358967781066895, | |
| "learning_rate": 1.3216622796621244e-05, | |
| "loss": 0.0612, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 0.8791773778920309, | |
| "grad_norm": 16.21796226501465, | |
| "learning_rate": 1.3160812445012099e-05, | |
| "loss": 0.1083, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 0.8849614395886889, | |
| "grad_norm": 2.596705675125122, | |
| "learning_rate": 1.310427957611983e-05, | |
| "loss": 0.0584, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 0.890745501285347, | |
| "grad_norm": 7.7204670906066895, | |
| "learning_rate": 1.3047032507503966e-05, | |
| "loss": 0.0929, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 0.8965295629820051, | |
| "grad_norm": 1.744662880897522, | |
| "learning_rate": 1.2989079661802703e-05, | |
| "loss": 0.066, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.9023136246786633, | |
| "grad_norm": 3.123241424560547, | |
| "learning_rate": 1.2930429565493705e-05, | |
| "loss": 0.0423, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 0.9023136246786633, | |
| "eval_Qnli-dev_cosine_accuracy": 0.712890625, | |
| "eval_Qnli-dev_cosine_accuracy_threshold": 0.6461004018783569, | |
| "eval_Qnli-dev_cosine_ap": 0.7429638364715181, | |
| "eval_Qnli-dev_cosine_f1": 0.7142857142857143, | |
| "eval_Qnli-dev_cosine_f1_threshold": 0.5756168365478516, | |
| "eval_Qnli-dev_cosine_precision": 0.5965909090909091, | |
| "eval_Qnli-dev_cosine_recall": 0.8898305084745762, | |
| "eval_Qnli-dev_dot_accuracy": 0.689453125, | |
| "eval_Qnli-dev_dot_accuracy_threshold": 560.8919067382812, | |
| "eval_Qnli-dev_dot_ap": 0.7158571398614493, | |
| "eval_Qnli-dev_dot_f1": 0.698292220113852, | |
| "eval_Qnli-dev_dot_f1_threshold": 560.8919067382812, | |
| "eval_Qnli-dev_dot_precision": 0.6323024054982818, | |
| "eval_Qnli-dev_dot_recall": 0.7796610169491526, | |
| "eval_Qnli-dev_euclidean_accuracy": 0.716796875, | |
| "eval_Qnli-dev_euclidean_accuracy_threshold": 23.726337432861328, | |
| "eval_Qnli-dev_euclidean_ap": 0.7468184659568129, | |
| "eval_Qnli-dev_euclidean_f1": 0.7103448275862068, | |
| "eval_Qnli-dev_euclidean_f1_threshold": 27.682994842529297, | |
| "eval_Qnli-dev_euclidean_precision": 0.5988372093023255, | |
| "eval_Qnli-dev_euclidean_recall": 0.8728813559322034, | |
| "eval_Qnli-dev_manhattan_accuracy": 0.71875, | |
| "eval_Qnli-dev_manhattan_accuracy_threshold": 729.7203369140625, | |
| "eval_Qnli-dev_manhattan_ap": 0.7468995752921103, | |
| "eval_Qnli-dev_manhattan_f1": 0.7122807017543861, | |
| "eval_Qnli-dev_manhattan_f1_threshold": 845.4459228515625, | |
| "eval_Qnli-dev_manhattan_precision": 0.6077844311377245, | |
| "eval_Qnli-dev_manhattan_recall": 0.8601694915254238, | |
| "eval_Qnli-dev_max_accuracy": 0.71875, | |
| "eval_Qnli-dev_max_accuracy_threshold": 729.7203369140625, | |
| "eval_Qnli-dev_max_ap": 0.7468995752921103, | |
| "eval_Qnli-dev_max_f1": 0.7142857142857143, | |
| "eval_Qnli-dev_max_f1_threshold": 845.4459228515625, | |
| "eval_Qnli-dev_max_precision": 0.6323024054982818, | |
| "eval_Qnli-dev_max_recall": 0.8898305084745762, | |
| "eval_allNLI-dev_cosine_accuracy": 0.74609375, | |
| "eval_allNLI-dev_cosine_accuracy_threshold": 0.7913638353347778, | |
| "eval_allNLI-dev_cosine_ap": 0.6147794488224912, | |
| "eval_allNLI-dev_cosine_f1": 0.6318289786223278, | |
| "eval_allNLI-dev_cosine_f1_threshold": 0.6930974721908569, | |
| "eval_allNLI-dev_cosine_precision": 0.5362903225806451, | |
| "eval_allNLI-dev_cosine_recall": 0.7687861271676301, | |
| "eval_allNLI-dev_dot_accuracy": 0.75390625, | |
| "eval_allNLI-dev_dot_accuracy_threshold": 700.6183471679688, | |
| "eval_allNLI-dev_dot_ap": 0.6091421586452705, | |
| "eval_allNLI-dev_dot_f1": 0.6189473684210527, | |
| "eval_allNLI-dev_dot_f1_threshold": 583.320556640625, | |
| "eval_allNLI-dev_dot_precision": 0.4867549668874172, | |
| "eval_allNLI-dev_dot_recall": 0.8497109826589595, | |
| "eval_allNLI-dev_euclidean_accuracy": 0.740234375, | |
| "eval_allNLI-dev_euclidean_accuracy_threshold": 19.20952606201172, | |
| "eval_allNLI-dev_euclidean_ap": 0.6122761202106319, | |
| "eval_allNLI-dev_euclidean_f1": 0.6340326340326341, | |
| "eval_allNLI-dev_euclidean_f1_threshold": 23.720947265625, | |
| "eval_allNLI-dev_euclidean_precision": 0.53125, | |
| "eval_allNLI-dev_euclidean_recall": 0.7861271676300579, | |
| "eval_allNLI-dev_manhattan_accuracy": 0.734375, | |
| "eval_allNLI-dev_manhattan_accuracy_threshold": 600.646240234375, | |
| "eval_allNLI-dev_manhattan_ap": 0.6131254272679774, | |
| "eval_allNLI-dev_manhattan_f1": 0.6327944572748269, | |
| "eval_allNLI-dev_manhattan_f1_threshold": 743.770263671875, | |
| "eval_allNLI-dev_manhattan_precision": 0.5269230769230769, | |
| "eval_allNLI-dev_manhattan_recall": 0.791907514450867, | |
| "eval_allNLI-dev_max_accuracy": 0.75390625, | |
| "eval_allNLI-dev_max_accuracy_threshold": 700.6183471679688, | |
| "eval_allNLI-dev_max_ap": 0.6147794488224912, | |
| "eval_allNLI-dev_max_f1": 0.6340326340326341, | |
| "eval_allNLI-dev_max_f1_threshold": 743.770263671875, | |
| "eval_allNLI-dev_max_precision": 0.5362903225806451, | |
| "eval_allNLI-dev_max_recall": 0.8497109826589595, | |
| "eval_sequential_score": 0.7468995752921103, | |
| "eval_sts-test_pearson_cosine": 0.9159220241890097, | |
| "eval_sts-test_pearson_dot": 0.9052308545844396, | |
| "eval_sts-test_pearson_euclidean": 0.9313844833923889, | |
| "eval_sts-test_pearson_manhattan": 0.9306550150689094, | |
| "eval_sts-test_pearson_max": 0.9313844833923889, | |
| "eval_sts-test_spearman_cosine": 0.9273589092085042, | |
| "eval_sts-test_spearman_dot": 0.9088352880883306, | |
| "eval_sts-test_spearman_euclidean": 0.9273073110041199, | |
| "eval_sts-test_spearman_manhattan": 0.9268169266817929, | |
| "eval_sts-test_spearman_max": 0.9273589092085042, | |
| "eval_vitaminc-pairs_loss": 2.987377166748047, | |
| "eval_vitaminc-pairs_runtime": 4.4371, | |
| "eval_vitaminc-pairs_samples_per_second": 28.848, | |
| "eval_vitaminc-pairs_steps_per_second": 0.225, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 0.9023136246786633, | |
| "eval_negation-triplets_loss": 0.5298277735710144, | |
| "eval_negation-triplets_runtime": 3.2754, | |
| "eval_negation-triplets_samples_per_second": 39.079, | |
| "eval_negation-triplets_steps_per_second": 0.305, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 0.9023136246786633, | |
| "eval_scitail-pairs-pos_loss": 0.008193486370146275, | |
| "eval_scitail-pairs-pos_runtime": 2.707, | |
| "eval_scitail-pairs-pos_samples_per_second": 47.285, | |
| "eval_scitail-pairs-pos_steps_per_second": 0.369, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 0.9023136246786633, | |
| "eval_scitail-pairs-qa_loss": 0.003911494743078947, | |
| "eval_scitail-pairs-qa_runtime": 2.2325, | |
| "eval_scitail-pairs-qa_samples_per_second": 57.334, | |
| "eval_scitail-pairs-qa_steps_per_second": 0.448, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 0.9023136246786633, | |
| "eval_xsum-pairs_loss": 0.003330780193209648, | |
| "eval_xsum-pairs_runtime": 3.0861, | |
| "eval_xsum-pairs_samples_per_second": 41.476, | |
| "eval_xsum-pairs_steps_per_second": 0.324, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 0.9023136246786633, | |
| "eval_sciq_pairs_loss": 0.06425436586141586, | |
| "eval_sciq_pairs_runtime": 4.8256, | |
| "eval_sciq_pairs_samples_per_second": 26.525, | |
| "eval_sciq_pairs_steps_per_second": 0.207, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 0.9023136246786633, | |
| "eval_qasc_pairs_loss": 0.06486833840608597, | |
| "eval_qasc_pairs_runtime": 2.1633, | |
| "eval_qasc_pairs_samples_per_second": 59.17, | |
| "eval_qasc_pairs_steps_per_second": 0.462, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 0.9023136246786633, | |
| "eval_openbookqa_pairs_loss": 0.34340742230415344, | |
| "eval_openbookqa_pairs_runtime": 2.225, | |
| "eval_openbookqa_pairs_samples_per_second": 57.527, | |
| "eval_openbookqa_pairs_steps_per_second": 0.449, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 0.9023136246786633, | |
| "eval_msmarco_pairs_loss": 0.00849138479679823, | |
| "eval_msmarco_pairs_runtime": 2.2296, | |
| "eval_msmarco_pairs_samples_per_second": 57.409, | |
| "eval_msmarco_pairs_steps_per_second": 0.449, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 0.9023136246786633, | |
| "eval_nq_pairs_loss": 0.03956576809287071, | |
| "eval_nq_pairs_runtime": 3.6403, | |
| "eval_nq_pairs_samples_per_second": 35.162, | |
| "eval_nq_pairs_steps_per_second": 0.275, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 0.9023136246786633, | |
| "eval_trivia_pairs_loss": 0.018078835681080818, | |
| "eval_trivia_pairs_runtime": 3.0876, | |
| "eval_trivia_pairs_samples_per_second": 41.456, | |
| "eval_trivia_pairs_steps_per_second": 0.324, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 0.9023136246786633, | |
| "eval_gooaq_pairs_loss": 0.07264826446771622, | |
| "eval_gooaq_pairs_runtime": 2.1431, | |
| "eval_gooaq_pairs_samples_per_second": 59.727, | |
| "eval_gooaq_pairs_steps_per_second": 0.467, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 0.9023136246786633, | |
| "eval_paws-pos_loss": 0.026289092376828194, | |
| "eval_paws-pos_runtime": 2.2416, | |
| "eval_paws-pos_samples_per_second": 57.102, | |
| "eval_paws-pos_steps_per_second": 0.446, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 0.9023136246786633, | |
| "eval_global_dataset_loss": 0.1634972095489502, | |
| "eval_global_dataset_runtime": 8.3728, | |
| "eval_global_dataset_samples_per_second": 38.816, | |
| "eval_global_dataset_steps_per_second": 0.358, | |
| "step": 468 | |
| } | |
| ], | |
| "logging_steps": 3, | |
| "max_steps": 1036, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 2, | |
| "save_steps": 52, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 0.0, | |
| "train_batch_size": 64, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |