abrek commited on
Commit
ec0db45
·
1 Parent(s): f8c770b

update nlg results

Browse files
results/zero-shot/CerebrumTech__cere-llama-3-8b-tr.json CHANGED
@@ -13,36 +13,42 @@
13
  {
14
  "name": "xlsum_tr",
15
  "task": "summarization",
16
- "rouge1": 0.0148042679964553,
17
- "rouge2": 0.006337012269480576,
18
- "rougeL": 0.011798434065329946
19
  },
20
  {
21
  "name": "wmt-tr-en-prompt",
22
  "task": "machine_translation",
23
- "wer": 0.9177312951556903,
24
- "bleu": 0.0010336244771491927
 
 
 
 
 
 
25
  },
26
  {
27
  "name": "wiki_lingua_tr",
28
  "task": "summarization",
29
- "rouge1": 0.07039446024412022,
30
- "rouge2": 0.02125810875804613,
31
- "rougeL": 0.05181069185122056
32
  },
33
  {
34
  "name": "tr-wikihow-summ",
35
  "task": "summarization",
36
- "rouge1": 0.17657152314685107,
37
- "rouge2": 0.05191014365298107,
38
- "rougeL": 0.1243539526593285
39
  },
40
  {
41
  "name": "mlsum_tr",
42
  "task": "summarization",
43
- "rouge1": 0.09056852256508315,
44
- "rouge2": 0.05971047138214301,
45
- "rougeL": 0.07758457056947823
46
  },
47
  {
48
  "name": "gecturk_generation",
@@ -192,12 +198,6 @@
192
  "task": "multiple_choice",
193
  "acc": 0.5857142857142857,
194
  "acc_norm": 0.5857142857142857
195
- },
196
- {
197
- "name": "wmt-en-tr-prompt",
198
- "task": "machine_translation",
199
- "wer": 0.9521851657325472,
200
- "bleu": 0.03042164195826046
201
  }
202
  ]
203
  }
 
13
  {
14
  "name": "xlsum_tr",
15
  "task": "summarization",
16
+ "rouge1": 0.2583902819569687,
17
+ "rouge2": 0.10868375368178293,
18
+ "rougeL": 0.1952205382650655
19
  },
20
  {
21
  "name": "wmt-tr-en-prompt",
22
  "task": "machine_translation",
23
+ "wer": 0.7999790984392787,
24
+ "bleu": 0.11905815617598979
25
+ },
26
+ {
27
+ "name": "wmt-en-tr-prompt",
28
+ "task": "machine_translation",
29
+ "wer": 0.933992772871535,
30
+ "bleu": 0.044204591279262746
31
  },
32
  {
33
  "name": "wiki_lingua_tr",
34
  "task": "summarization",
35
+ "rouge1": 0.20581622347355663,
36
+ "rouge2": 0.0641413363983985,
37
+ "rougeL": 0.14917636822465685
38
  },
39
  {
40
  "name": "tr-wikihow-summ",
41
  "task": "summarization",
42
+ "rouge1": 0.2246335258862419,
43
+ "rouge2": 0.06557441812173484,
44
+ "rougeL": 0.1559504924053904
45
  },
46
  {
47
  "name": "mlsum_tr",
48
  "task": "summarization",
49
+ "rouge1": 0.38085271760138706,
50
+ "rouge2": 0.24779093184456852,
51
+ "rougeL": 0.31787475025364564
52
  },
53
  {
54
  "name": "gecturk_generation",
 
198
  "task": "multiple_choice",
199
  "acc": 0.5857142857142857,
200
  "acc_norm": 0.5857142857142857
 
 
 
 
 
 
201
  }
202
  ]
203
  }
results/zero-shot/aya101.json CHANGED
@@ -18,9 +18,9 @@
18
  {
19
  "name": "xlsum_tr",
20
  "task": "summarization",
21
- "rouge1": 0.02416422194769531,
22
- "rouge2": 0.00149839274458772,
23
- "rougeL": 0.02416422194769531
24
  },
25
  {
26
  "name": "xcopa_tr",
@@ -31,15 +31,15 @@
31
  {
32
  "name": "wmt-tr-en-prompt",
33
  "task": "machine_translation",
34
- "wer": 0.9853633715998092,
35
- "bleu": 0.0
36
  },
37
  {
38
  "name": "wiki_lingua_tr",
39
  "task": "summarization",
40
- "rouge1": 0.029006633700390562,
41
- "rouge2": 0.0004998910319276452,
42
- "rougeL": 0.028967197984657227
43
  },
44
  {
45
  "name": "turkish_plu",
@@ -86,9 +86,9 @@
86
  {
87
  "name": "tr-wikihow-summ",
88
  "task": "summarization",
89
- "rouge1": 0.02053796966151103,
90
- "rouge2": 0.00029270301029826366,
91
- "rougeL": 0.020495031370814234
92
  },
93
  {
94
  "name": "tquad",
@@ -135,9 +135,9 @@
135
  {
136
  "name": "mlsum_tr",
137
  "task": "summarization",
138
- "rouge1": 0.021746360547255133,
139
- "rouge2": 0.003113110667892852,
140
- "rougeL": 0.021727065059735186
141
  },
142
  {
143
  "name": "mkqa_tr",
@@ -195,7 +195,7 @@
195
  {
196
  "name": "wmt-en-tr-prompt",
197
  "task": "machine_translation",
198
- "wer": 0.7025061825593694,
199
  "bleu": 0.21521710925537071
200
  }
201
  ]
 
18
  {
19
  "name": "xlsum_tr",
20
  "task": "summarization",
21
+ "rouge1": 0.40329166826890306,
22
+ "rouge2": 0.24618000685069955,
23
+ "rougeL": 0.3540445580839572
24
  },
25
  {
26
  "name": "xcopa_tr",
 
31
  {
32
  "name": "wmt-tr-en-prompt",
33
  "task": "machine_translation",
34
+ "wer": 0.7665278686420781,
35
+ "bleu": 0.17158412519070387
36
  },
37
  {
38
  "name": "wiki_lingua_tr",
39
  "task": "summarization",
40
+ "rouge1": 0.39998774165132267,
41
+ "rouge2": 0.2218282180595083,
42
+ "rougeL": 0.35347919895816715
43
  },
44
  {
45
  "name": "turkish_plu",
 
86
  {
87
  "name": "tr-wikihow-summ",
88
  "task": "summarization",
89
+ "rouge1": 0.3465851896339094,
90
+ "rouge2": 0.16864200623272294,
91
+ "rougeL": 0.2930576004057389
92
  },
93
  {
94
  "name": "tquad",
 
135
  {
136
  "name": "mlsum_tr",
137
  "task": "summarization",
138
+ "rouge1": 0.45270949786483405,
139
+ "rouge2": 0.3184853277364837,
140
+ "rougeL": 0.39689938533216157
141
  },
142
  {
143
  "name": "mkqa_tr",
 
195
  {
196
  "name": "wmt-en-tr-prompt",
197
  "task": "machine_translation",
198
+ "wer": 0.7025061825593685,
199
  "bleu": 0.21521710925537071
200
  }
201
  ]
results/zero-shot/llama-3.2-1b.json CHANGED
@@ -44,13 +44,6 @@
44
  "exact_match": 0.007694584196507843,
45
  "f1": 0.03304091036050505
46
  },
47
- {
48
- "name": "mlsum_tr",
49
- "task": "summarization",
50
- "rouge1": 0.23283491254211872,
51
- "rouge2": 0.13426790568610214,
52
- "rougeL": 0.18915548037371513
53
- },
54
  {
55
  "name": "news_cat",
56
  "task": "text_classification",
@@ -81,43 +74,18 @@
81
  "exact_match": 0.06278026905829596,
82
  "f1": 0.21486130318406463
83
  },
84
- {
85
- "name": "wiki_lingua_tr",
86
- "task": "summarization",
87
- "rouge1": 0.10861529436199803,
88
- "rouge2": 0.034862923521078545,
89
- "rougeL": 0.08692160533533941
90
- },
91
- {
92
- "name": "wmt-tr-en-prompt",
93
- "task": "machine_translation",
94
- "wer": 3.910683208136067,
95
- "bleu": 0.012043288243775466
96
- },
97
  {
98
  "name": "xcopa_tr",
99
  "task": "multiple_choice",
100
  "acc": 0.556,
101
  "acc_norm": 0.556
102
  },
103
- {
104
- "name": "xlsum_tr",
105
- "task": "summarization",
106
- "rouge1": 0.16924699150407269,
107
- "rouge2": 0.07190935921365724,
108
- "rougeL": 0.13255123335488528
109
- },
110
  {
111
  "name": "xquad_tr",
112
  "task": "extractive_question_answering",
113
  "exact_match": 0.04873949579831932,
114
  "f1": 0.11156636293859905
115
  },
116
- {
117
- "name": "gecturk_generation",
118
- "task": "grammatical_error_correction",
119
- "exact_match": 0.0073185998362944775
120
- },
121
  {
122
  "name": "mlsum_tr",
123
  "task": "summarization",
 
44
  "exact_match": 0.007694584196507843,
45
  "f1": 0.03304091036050505
46
  },
 
 
 
 
 
 
 
47
  {
48
  "name": "news_cat",
49
  "task": "text_classification",
 
74
  "exact_match": 0.06278026905829596,
75
  "f1": 0.21486130318406463
76
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
77
  {
78
  "name": "xcopa_tr",
79
  "task": "multiple_choice",
80
  "acc": 0.556,
81
  "acc_norm": 0.556
82
  },
 
 
 
 
 
 
 
83
  {
84
  "name": "xquad_tr",
85
  "task": "extractive_question_answering",
86
  "exact_match": 0.04873949579831932,
87
  "f1": 0.11156636293859905
88
  },
 
 
 
 
 
89
  {
90
  "name": "mlsum_tr",
91
  "task": "summarization",
results/zero-shot/llama-3.2-3b-instruct.json CHANGED
@@ -44,13 +44,6 @@
44
  "exact_match": 0.04675939627108612,
45
  "f1": 0.08114473798410345
46
  },
47
- {
48
- "name": "mlsum_tr",
49
- "task": "summarization",
50
- "rouge1": 0.2669056212126977,
51
- "rouge2": 0.1480446780314802,
52
- "rougeL": 0.2106440565987865
53
- },
54
  {
55
  "name": "news_cat",
56
  "task": "text_classification",
@@ -81,43 +74,18 @@
81
  "exact_match": 0.18721973094170405,
82
  "f1": 0.5109898180473623
83
  },
84
- {
85
- "name": "wiki_lingua_tr",
86
- "task": "summarization",
87
- "rouge1": 0.1342879173103036,
88
- "rouge2": 0.041489300068460175,
89
- "rougeL": 0.10482785510181569
90
- },
91
- {
92
- "name": "wmt-tr-en-prompt",
93
- "task": "machine_translation",
94
- "wer": 1.7706536060519733,
95
- "bleu": 0.048843165627950165
96
- },
97
  {
98
  "name": "xcopa_tr",
99
  "task": "multiple_choice",
100
  "acc": 0.546,
101
  "acc_norm": 0.546
102
  },
103
- {
104
- "name": "xlsum_tr",
105
- "task": "summarization",
106
- "rouge1": 0.17224405229987672,
107
- "rouge2": 0.06736413357191079,
108
- "rougeL": 0.12750762702828333
109
- },
110
  {
111
  "name": "xquad_tr",
112
  "task": "extractive_question_answering",
113
  "exact_match": 0.23025210084033615,
114
  "f1": 0.4335914561273987
115
  },
116
- {
117
- "name": "gecturk_generation",
118
- "task": "grammatical_error_correction",
119
- "exact_match": 0.009726033992970293
120
- },
121
  {
122
  "name": "mlsum_tr",
123
  "task": "summarization",
 
44
  "exact_match": 0.04675939627108612,
45
  "f1": 0.08114473798410345
46
  },
 
 
 
 
 
 
 
47
  {
48
  "name": "news_cat",
49
  "task": "text_classification",
 
74
  "exact_match": 0.18721973094170405,
75
  "f1": 0.5109898180473623
76
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
77
  {
78
  "name": "xcopa_tr",
79
  "task": "multiple_choice",
80
  "acc": 0.546,
81
  "acc_norm": 0.546
82
  },
 
 
 
 
 
 
 
83
  {
84
  "name": "xquad_tr",
85
  "task": "extractive_question_answering",
86
  "exact_match": 0.23025210084033615,
87
  "f1": 0.4335914561273987
88
  },
 
 
 
 
 
89
  {
90
  "name": "mlsum_tr",
91
  "task": "summarization",