Librarian Bot: Update Hugging Face dataset ID
#8
by
librarian-bot
- opened
README.md
CHANGED
|
@@ -1,6 +1,7 @@
|
|
| 1 |
---
|
| 2 |
-
license: apache-2.0
|
| 3 |
language: en
|
|
|
|
|
|
|
| 4 |
tags:
|
| 5 |
- deberta-v3-base
|
| 6 |
- deberta-v3
|
|
@@ -16,32 +17,6 @@ tags:
|
|
| 16 |
- tasksource
|
| 17 |
- zero-shot
|
| 18 |
- rlhf
|
| 19 |
-
model-index:
|
| 20 |
-
- name: deberta-v3-base-tasksource-nli
|
| 21 |
-
results:
|
| 22 |
-
- task:
|
| 23 |
-
type: text-classification
|
| 24 |
-
name: Text Classification
|
| 25 |
-
dataset:
|
| 26 |
-
name: glue
|
| 27 |
-
type: glue
|
| 28 |
-
config: rte
|
| 29 |
-
split: validation
|
| 30 |
-
metrics:
|
| 31 |
-
- type: accuracy
|
| 32 |
-
value: 0.89
|
| 33 |
-
- task:
|
| 34 |
-
type: natural-language-inference
|
| 35 |
-
name: Natural Language Inference
|
| 36 |
-
dataset:
|
| 37 |
-
name: anli-r3
|
| 38 |
-
type: anli
|
| 39 |
-
config: plain_text
|
| 40 |
-
split: validation
|
| 41 |
-
metrics:
|
| 42 |
-
- type: accuracy
|
| 43 |
-
value: 0.52
|
| 44 |
-
name: Accuracy
|
| 45 |
datasets:
|
| 46 |
- glue
|
| 47 |
- super_glue
|
|
@@ -109,7 +84,7 @@ datasets:
|
|
| 109 |
- tasksource/mmlu
|
| 110 |
- winogrande
|
| 111 |
- codah
|
| 112 |
-
- ai2_arc
|
| 113 |
- definite_pronoun_resolution
|
| 114 |
- swag
|
| 115 |
- math_qa
|
|
@@ -310,8 +285,33 @@ datasets:
|
|
| 310 |
- lmsys/chatbot_arena_conversations
|
| 311 |
metrics:
|
| 312 |
- accuracy
|
| 313 |
-
library_name: transformers
|
| 314 |
pipeline_tag: zero-shot-classification
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 315 |
---
|
| 316 |
|
| 317 |
# Model Card for DeBERTa-v3-base-tasksource-nli
|
|
|
|
| 1 |
---
|
|
|
|
| 2 |
language: en
|
| 3 |
+
license: apache-2.0
|
| 4 |
+
library_name: transformers
|
| 5 |
tags:
|
| 6 |
- deberta-v3-base
|
| 7 |
- deberta-v3
|
|
|
|
| 17 |
- tasksource
|
| 18 |
- zero-shot
|
| 19 |
- rlhf
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 20 |
datasets:
|
| 21 |
- glue
|
| 22 |
- super_glue
|
|
|
|
| 84 |
- tasksource/mmlu
|
| 85 |
- winogrande
|
| 86 |
- codah
|
| 87 |
+
- allenai/ai2_arc
|
| 88 |
- definite_pronoun_resolution
|
| 89 |
- swag
|
| 90 |
- math_qa
|
|
|
|
| 285 |
- lmsys/chatbot_arena_conversations
|
| 286 |
metrics:
|
| 287 |
- accuracy
|
|
|
|
| 288 |
pipeline_tag: zero-shot-classification
|
| 289 |
+
model-index:
|
| 290 |
+
- name: deberta-v3-base-tasksource-nli
|
| 291 |
+
results:
|
| 292 |
+
- task:
|
| 293 |
+
type: text-classification
|
| 294 |
+
name: Text Classification
|
| 295 |
+
dataset:
|
| 296 |
+
name: glue
|
| 297 |
+
type: glue
|
| 298 |
+
config: rte
|
| 299 |
+
split: validation
|
| 300 |
+
metrics:
|
| 301 |
+
- type: accuracy
|
| 302 |
+
value: 0.89
|
| 303 |
+
- task:
|
| 304 |
+
type: natural-language-inference
|
| 305 |
+
name: Natural Language Inference
|
| 306 |
+
dataset:
|
| 307 |
+
name: anli-r3
|
| 308 |
+
type: anli
|
| 309 |
+
config: plain_text
|
| 310 |
+
split: validation
|
| 311 |
+
metrics:
|
| 312 |
+
- type: accuracy
|
| 313 |
+
value: 0.52
|
| 314 |
+
name: Accuracy
|
| 315 |
---
|
| 316 |
|
| 317 |
# Model Card for DeBERTa-v3-base-tasksource-nli
|