Spaces:
Running
Running
update nlg results
Browse files
results/zero-shot/CerebrumTech__cere-llama-3-8b-tr.json
CHANGED
|
@@ -13,36 +13,42 @@
|
|
| 13 |
{
|
| 14 |
"name": "xlsum_tr",
|
| 15 |
"task": "summarization",
|
| 16 |
-
"rouge1": 0.
|
| 17 |
-
"rouge2": 0.
|
| 18 |
-
"rougeL": 0.
|
| 19 |
},
|
| 20 |
{
|
| 21 |
"name": "wmt-tr-en-prompt",
|
| 22 |
"task": "machine_translation",
|
| 23 |
-
"wer": 0.
|
| 24 |
-
"bleu": 0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 25 |
},
|
| 26 |
{
|
| 27 |
"name": "wiki_lingua_tr",
|
| 28 |
"task": "summarization",
|
| 29 |
-
"rouge1": 0.
|
| 30 |
-
"rouge2": 0.
|
| 31 |
-
"rougeL": 0.
|
| 32 |
},
|
| 33 |
{
|
| 34 |
"name": "tr-wikihow-summ",
|
| 35 |
"task": "summarization",
|
| 36 |
-
"rouge1": 0.
|
| 37 |
-
"rouge2": 0.
|
| 38 |
-
"rougeL": 0.
|
| 39 |
},
|
| 40 |
{
|
| 41 |
"name": "mlsum_tr",
|
| 42 |
"task": "summarization",
|
| 43 |
-
"rouge1": 0.
|
| 44 |
-
"rouge2": 0.
|
| 45 |
-
"rougeL": 0.
|
| 46 |
},
|
| 47 |
{
|
| 48 |
"name": "gecturk_generation",
|
|
@@ -192,12 +198,6 @@
|
|
| 192 |
"task": "multiple_choice",
|
| 193 |
"acc": 0.5857142857142857,
|
| 194 |
"acc_norm": 0.5857142857142857
|
| 195 |
-
},
|
| 196 |
-
{
|
| 197 |
-
"name": "wmt-en-tr-prompt",
|
| 198 |
-
"task": "machine_translation",
|
| 199 |
-
"wer": 0.9521851657325472,
|
| 200 |
-
"bleu": 0.03042164195826046
|
| 201 |
}
|
| 202 |
]
|
| 203 |
}
|
|
|
|
| 13 |
{
|
| 14 |
"name": "xlsum_tr",
|
| 15 |
"task": "summarization",
|
| 16 |
+
"rouge1": 0.2583902819569687,
|
| 17 |
+
"rouge2": 0.10868375368178293,
|
| 18 |
+
"rougeL": 0.1952205382650655
|
| 19 |
},
|
| 20 |
{
|
| 21 |
"name": "wmt-tr-en-prompt",
|
| 22 |
"task": "machine_translation",
|
| 23 |
+
"wer": 0.7999790984392787,
|
| 24 |
+
"bleu": 0.11905815617598979
|
| 25 |
+
},
|
| 26 |
+
{
|
| 27 |
+
"name": "wmt-en-tr-prompt",
|
| 28 |
+
"task": "machine_translation",
|
| 29 |
+
"wer": 0.933992772871535,
|
| 30 |
+
"bleu": 0.044204591279262746
|
| 31 |
},
|
| 32 |
{
|
| 33 |
"name": "wiki_lingua_tr",
|
| 34 |
"task": "summarization",
|
| 35 |
+
"rouge1": 0.20581622347355663,
|
| 36 |
+
"rouge2": 0.0641413363983985,
|
| 37 |
+
"rougeL": 0.14917636822465685
|
| 38 |
},
|
| 39 |
{
|
| 40 |
"name": "tr-wikihow-summ",
|
| 41 |
"task": "summarization",
|
| 42 |
+
"rouge1": 0.2246335258862419,
|
| 43 |
+
"rouge2": 0.06557441812173484,
|
| 44 |
+
"rougeL": 0.1559504924053904
|
| 45 |
},
|
| 46 |
{
|
| 47 |
"name": "mlsum_tr",
|
| 48 |
"task": "summarization",
|
| 49 |
+
"rouge1": 0.38085271760138706,
|
| 50 |
+
"rouge2": 0.24779093184456852,
|
| 51 |
+
"rougeL": 0.31787475025364564
|
| 52 |
},
|
| 53 |
{
|
| 54 |
"name": "gecturk_generation",
|
|
|
|
| 198 |
"task": "multiple_choice",
|
| 199 |
"acc": 0.5857142857142857,
|
| 200 |
"acc_norm": 0.5857142857142857
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 201 |
}
|
| 202 |
]
|
| 203 |
}
|
results/zero-shot/aya101.json
CHANGED
|
@@ -18,9 +18,9 @@
|
|
| 18 |
{
|
| 19 |
"name": "xlsum_tr",
|
| 20 |
"task": "summarization",
|
| 21 |
-
"rouge1": 0.
|
| 22 |
-
"rouge2": 0.
|
| 23 |
-
"rougeL": 0.
|
| 24 |
},
|
| 25 |
{
|
| 26 |
"name": "xcopa_tr",
|
|
@@ -31,15 +31,15 @@
|
|
| 31 |
{
|
| 32 |
"name": "wmt-tr-en-prompt",
|
| 33 |
"task": "machine_translation",
|
| 34 |
-
"wer": 0.
|
| 35 |
-
"bleu": 0.
|
| 36 |
},
|
| 37 |
{
|
| 38 |
"name": "wiki_lingua_tr",
|
| 39 |
"task": "summarization",
|
| 40 |
-
"rouge1": 0.
|
| 41 |
-
"rouge2": 0.
|
| 42 |
-
"rougeL": 0.
|
| 43 |
},
|
| 44 |
{
|
| 45 |
"name": "turkish_plu",
|
|
@@ -86,9 +86,9 @@
|
|
| 86 |
{
|
| 87 |
"name": "tr-wikihow-summ",
|
| 88 |
"task": "summarization",
|
| 89 |
-
"rouge1": 0.
|
| 90 |
-
"rouge2": 0.
|
| 91 |
-
"rougeL": 0.
|
| 92 |
},
|
| 93 |
{
|
| 94 |
"name": "tquad",
|
|
@@ -135,9 +135,9 @@
|
|
| 135 |
{
|
| 136 |
"name": "mlsum_tr",
|
| 137 |
"task": "summarization",
|
| 138 |
-
"rouge1": 0.
|
| 139 |
-
"rouge2": 0.
|
| 140 |
-
"rougeL": 0.
|
| 141 |
},
|
| 142 |
{
|
| 143 |
"name": "mkqa_tr",
|
|
@@ -195,7 +195,7 @@
|
|
| 195 |
{
|
| 196 |
"name": "wmt-en-tr-prompt",
|
| 197 |
"task": "machine_translation",
|
| 198 |
-
"wer": 0.
|
| 199 |
"bleu": 0.21521710925537071
|
| 200 |
}
|
| 201 |
]
|
|
|
|
| 18 |
{
|
| 19 |
"name": "xlsum_tr",
|
| 20 |
"task": "summarization",
|
| 21 |
+
"rouge1": 0.40329166826890306,
|
| 22 |
+
"rouge2": 0.24618000685069955,
|
| 23 |
+
"rougeL": 0.3540445580839572
|
| 24 |
},
|
| 25 |
{
|
| 26 |
"name": "xcopa_tr",
|
|
|
|
| 31 |
{
|
| 32 |
"name": "wmt-tr-en-prompt",
|
| 33 |
"task": "machine_translation",
|
| 34 |
+
"wer": 0.7665278686420781,
|
| 35 |
+
"bleu": 0.17158412519070387
|
| 36 |
},
|
| 37 |
{
|
| 38 |
"name": "wiki_lingua_tr",
|
| 39 |
"task": "summarization",
|
| 40 |
+
"rouge1": 0.39998774165132267,
|
| 41 |
+
"rouge2": 0.2218282180595083,
|
| 42 |
+
"rougeL": 0.35347919895816715
|
| 43 |
},
|
| 44 |
{
|
| 45 |
"name": "turkish_plu",
|
|
|
|
| 86 |
{
|
| 87 |
"name": "tr-wikihow-summ",
|
| 88 |
"task": "summarization",
|
| 89 |
+
"rouge1": 0.3465851896339094,
|
| 90 |
+
"rouge2": 0.16864200623272294,
|
| 91 |
+
"rougeL": 0.2930576004057389
|
| 92 |
},
|
| 93 |
{
|
| 94 |
"name": "tquad",
|
|
|
|
| 135 |
{
|
| 136 |
"name": "mlsum_tr",
|
| 137 |
"task": "summarization",
|
| 138 |
+
"rouge1": 0.45270949786483405,
|
| 139 |
+
"rouge2": 0.3184853277364837,
|
| 140 |
+
"rougeL": 0.39689938533216157
|
| 141 |
},
|
| 142 |
{
|
| 143 |
"name": "mkqa_tr",
|
|
|
|
| 195 |
{
|
| 196 |
"name": "wmt-en-tr-prompt",
|
| 197 |
"task": "machine_translation",
|
| 198 |
+
"wer": 0.7025061825593685,
|
| 199 |
"bleu": 0.21521710925537071
|
| 200 |
}
|
| 201 |
]
|
results/zero-shot/llama-3.2-1b.json
CHANGED
|
@@ -44,13 +44,6 @@
|
|
| 44 |
"exact_match": 0.007694584196507843,
|
| 45 |
"f1": 0.03304091036050505
|
| 46 |
},
|
| 47 |
-
{
|
| 48 |
-
"name": "mlsum_tr",
|
| 49 |
-
"task": "summarization",
|
| 50 |
-
"rouge1": 0.23283491254211872,
|
| 51 |
-
"rouge2": 0.13426790568610214,
|
| 52 |
-
"rougeL": 0.18915548037371513
|
| 53 |
-
},
|
| 54 |
{
|
| 55 |
"name": "news_cat",
|
| 56 |
"task": "text_classification",
|
|
@@ -81,43 +74,18 @@
|
|
| 81 |
"exact_match": 0.06278026905829596,
|
| 82 |
"f1": 0.21486130318406463
|
| 83 |
},
|
| 84 |
-
{
|
| 85 |
-
"name": "wiki_lingua_tr",
|
| 86 |
-
"task": "summarization",
|
| 87 |
-
"rouge1": 0.10861529436199803,
|
| 88 |
-
"rouge2": 0.034862923521078545,
|
| 89 |
-
"rougeL": 0.08692160533533941
|
| 90 |
-
},
|
| 91 |
-
{
|
| 92 |
-
"name": "wmt-tr-en-prompt",
|
| 93 |
-
"task": "machine_translation",
|
| 94 |
-
"wer": 3.910683208136067,
|
| 95 |
-
"bleu": 0.012043288243775466
|
| 96 |
-
},
|
| 97 |
{
|
| 98 |
"name": "xcopa_tr",
|
| 99 |
"task": "multiple_choice",
|
| 100 |
"acc": 0.556,
|
| 101 |
"acc_norm": 0.556
|
| 102 |
},
|
| 103 |
-
{
|
| 104 |
-
"name": "xlsum_tr",
|
| 105 |
-
"task": "summarization",
|
| 106 |
-
"rouge1": 0.16924699150407269,
|
| 107 |
-
"rouge2": 0.07190935921365724,
|
| 108 |
-
"rougeL": 0.13255123335488528
|
| 109 |
-
},
|
| 110 |
{
|
| 111 |
"name": "xquad_tr",
|
| 112 |
"task": "extractive_question_answering",
|
| 113 |
"exact_match": 0.04873949579831932,
|
| 114 |
"f1": 0.11156636293859905
|
| 115 |
},
|
| 116 |
-
{
|
| 117 |
-
"name": "gecturk_generation",
|
| 118 |
-
"task": "grammatical_error_correction",
|
| 119 |
-
"exact_match": 0.0073185998362944775
|
| 120 |
-
},
|
| 121 |
{
|
| 122 |
"name": "mlsum_tr",
|
| 123 |
"task": "summarization",
|
|
|
|
| 44 |
"exact_match": 0.007694584196507843,
|
| 45 |
"f1": 0.03304091036050505
|
| 46 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 47 |
{
|
| 48 |
"name": "news_cat",
|
| 49 |
"task": "text_classification",
|
|
|
|
| 74 |
"exact_match": 0.06278026905829596,
|
| 75 |
"f1": 0.21486130318406463
|
| 76 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 77 |
{
|
| 78 |
"name": "xcopa_tr",
|
| 79 |
"task": "multiple_choice",
|
| 80 |
"acc": 0.556,
|
| 81 |
"acc_norm": 0.556
|
| 82 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 83 |
{
|
| 84 |
"name": "xquad_tr",
|
| 85 |
"task": "extractive_question_answering",
|
| 86 |
"exact_match": 0.04873949579831932,
|
| 87 |
"f1": 0.11156636293859905
|
| 88 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 89 |
{
|
| 90 |
"name": "mlsum_tr",
|
| 91 |
"task": "summarization",
|
results/zero-shot/llama-3.2-3b-instruct.json
CHANGED
|
@@ -44,13 +44,6 @@
|
|
| 44 |
"exact_match": 0.04675939627108612,
|
| 45 |
"f1": 0.08114473798410345
|
| 46 |
},
|
| 47 |
-
{
|
| 48 |
-
"name": "mlsum_tr",
|
| 49 |
-
"task": "summarization",
|
| 50 |
-
"rouge1": 0.2669056212126977,
|
| 51 |
-
"rouge2": 0.1480446780314802,
|
| 52 |
-
"rougeL": 0.2106440565987865
|
| 53 |
-
},
|
| 54 |
{
|
| 55 |
"name": "news_cat",
|
| 56 |
"task": "text_classification",
|
|
@@ -81,43 +74,18 @@
|
|
| 81 |
"exact_match": 0.18721973094170405,
|
| 82 |
"f1": 0.5109898180473623
|
| 83 |
},
|
| 84 |
-
{
|
| 85 |
-
"name": "wiki_lingua_tr",
|
| 86 |
-
"task": "summarization",
|
| 87 |
-
"rouge1": 0.1342879173103036,
|
| 88 |
-
"rouge2": 0.041489300068460175,
|
| 89 |
-
"rougeL": 0.10482785510181569
|
| 90 |
-
},
|
| 91 |
-
{
|
| 92 |
-
"name": "wmt-tr-en-prompt",
|
| 93 |
-
"task": "machine_translation",
|
| 94 |
-
"wer": 1.7706536060519733,
|
| 95 |
-
"bleu": 0.048843165627950165
|
| 96 |
-
},
|
| 97 |
{
|
| 98 |
"name": "xcopa_tr",
|
| 99 |
"task": "multiple_choice",
|
| 100 |
"acc": 0.546,
|
| 101 |
"acc_norm": 0.546
|
| 102 |
},
|
| 103 |
-
{
|
| 104 |
-
"name": "xlsum_tr",
|
| 105 |
-
"task": "summarization",
|
| 106 |
-
"rouge1": 0.17224405229987672,
|
| 107 |
-
"rouge2": 0.06736413357191079,
|
| 108 |
-
"rougeL": 0.12750762702828333
|
| 109 |
-
},
|
| 110 |
{
|
| 111 |
"name": "xquad_tr",
|
| 112 |
"task": "extractive_question_answering",
|
| 113 |
"exact_match": 0.23025210084033615,
|
| 114 |
"f1": 0.4335914561273987
|
| 115 |
},
|
| 116 |
-
{
|
| 117 |
-
"name": "gecturk_generation",
|
| 118 |
-
"task": "grammatical_error_correction",
|
| 119 |
-
"exact_match": 0.009726033992970293
|
| 120 |
-
},
|
| 121 |
{
|
| 122 |
"name": "mlsum_tr",
|
| 123 |
"task": "summarization",
|
|
|
|
| 44 |
"exact_match": 0.04675939627108612,
|
| 45 |
"f1": 0.08114473798410345
|
| 46 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 47 |
{
|
| 48 |
"name": "news_cat",
|
| 49 |
"task": "text_classification",
|
|
|
|
| 74 |
"exact_match": 0.18721973094170405,
|
| 75 |
"f1": 0.5109898180473623
|
| 76 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 77 |
{
|
| 78 |
"name": "xcopa_tr",
|
| 79 |
"task": "multiple_choice",
|
| 80 |
"acc": 0.546,
|
| 81 |
"acc_norm": 0.546
|
| 82 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 83 |
{
|
| 84 |
"name": "xquad_tr",
|
| 85 |
"task": "extractive_question_answering",
|
| 86 |
"exact_match": 0.23025210084033615,
|
| 87 |
"f1": 0.4335914561273987
|
| 88 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 89 |
{
|
| 90 |
"name": "mlsum_tr",
|
| 91 |
"task": "summarization",
|