Fixing some errors of the leaderboard evaluation results in the ModelCard yaml
Browse filesThe name of a few benchmarks are incorrect on the model metadata.
This commit fixes some minor errors of the [last PR](9) on the ModelCard YAML metadata.
README.md
CHANGED
|
@@ -167,6 +167,19 @@ model-index:
|
|
| 167 |
- type: f1_macro
|
| 168 |
value: 43.01
|
| 169 |
name: f1-macro
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 170 |
- type: pearson
|
| 171 |
value: 8.92
|
| 172 |
name: pearson
|
|
@@ -194,7 +207,7 @@ model-index:
|
|
| 194 |
name: Text Generation
|
| 195 |
dataset:
|
| 196 |
name: HateBR Binary
|
| 197 |
-
type:
|
| 198 |
split: test
|
| 199 |
args:
|
| 200 |
num_few_shot: 25
|
|
@@ -202,6 +215,19 @@ model-index:
|
|
| 202 |
- type: f1_macro
|
| 203 |
value: 50.46
|
| 204 |
name: f1-macro
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 205 |
- type: f1_macro
|
| 206 |
value: 41.19
|
| 207 |
name: f1-macro
|
|
|
|
| 167 |
- type: f1_macro
|
| 168 |
value: 43.01
|
| 169 |
name: f1-macro
|
| 170 |
+
source:
|
| 171 |
+
url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard?query=22h/open-cabrita3b
|
| 172 |
+
name: Open Portuguese LLM Leaderboard
|
| 173 |
+
- task:
|
| 174 |
+
type: text-generation
|
| 175 |
+
name: Text Generation
|
| 176 |
+
dataset:
|
| 177 |
+
name: Assin2 STS
|
| 178 |
+
type: eduagarcia/portuguese_benchmark
|
| 179 |
+
split: test
|
| 180 |
+
args:
|
| 181 |
+
num_few_shot: 15
|
| 182 |
+
metrics:
|
| 183 |
- type: pearson
|
| 184 |
value: 8.92
|
| 185 |
name: pearson
|
|
|
|
| 207 |
name: Text Generation
|
| 208 |
dataset:
|
| 209 |
name: HateBR Binary
|
| 210 |
+
type: ruanchaves/hatebr
|
| 211 |
split: test
|
| 212 |
args:
|
| 213 |
num_few_shot: 25
|
|
|
|
| 215 |
- type: f1_macro
|
| 216 |
value: 50.46
|
| 217 |
name: f1-macro
|
| 218 |
+
source:
|
| 219 |
+
url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard?query=22h/open-cabrita3b
|
| 220 |
+
name: Open Portuguese LLM Leaderboard
|
| 221 |
+
- task:
|
| 222 |
+
type: text-generation
|
| 223 |
+
name: Text Generation
|
| 224 |
+
dataset:
|
| 225 |
+
name: PT Hate Speech Binary
|
| 226 |
+
type: hate_speech_portuguese
|
| 227 |
+
split: test
|
| 228 |
+
args:
|
| 229 |
+
num_few_shot: 25
|
| 230 |
+
metrics:
|
| 231 |
- type: f1_macro
|
| 232 |
value: 41.19
|
| 233 |
name: f1-macro
|