Update README.md
Browse files
README.md
CHANGED
|
@@ -118,20 +118,22 @@ As an example, the distilled version of BERT has 40% fewer parameters and runs 6
|
|
| 118 |
This model has been fine-tuned on the downstream tasks of the Catalan Language Understanding Evaluation benchmark (CLUB).
|
| 119 |
|
| 120 |
Here are the train/dev/test splits of each dataset:
|
| 121 |
-
|
| 122 |
-
|
| 123 |
-
|
| 124 |
-
| Ancora
|
| 125 |
-
|
|
| 126 |
-
|
|
| 127 |
-
|
|
| 128 |
-
|
|
| 129 |
-
|
|
| 130 |
-
|
|
|
|
|
| 131 |
|
| 132 |
### Evaluation results
|
| 133 |
|
| 134 |
This is how it compares to the teacher model when fine-tuned on the same downstream tasks:
|
|
|
|
| 135 |
| Model \ Task| NER (F1) | POS (F1) | STS-ca (Comb) | TeCla (Acc.) | TEca (Acc.) | VilaQuAD (F1/EM)| ViquiQuAD (F1/EM) | CatalanQA (F1/EM) | XQuAD-ca <sup>1</sup> (F1/EM) |
|
| 136 |
| ------------|:-------------:| -----:|:------|:------|:-------|:------|:----|:----|:----|
|
| 137 |
| RoBERTa-large-ca-v2 | 89.82 | 99.02 | 83.41 | 75.46 | 83.61 | 89.34/75.50 | 89.20/75.77 | 90.72/79.06 | 73.79/55.34 |
|
|
|
|
| 118 |
This model has been fine-tuned on the downstream tasks of the Catalan Language Understanding Evaluation benchmark (CLUB).
|
| 119 |
|
| 120 |
Here are the train/dev/test splits of each dataset:
|
| 121 |
+
|
| 122 |
+
| Dataset | Task| Total | Train | Dev | Test |
|
| 123 |
+
|:----------|:----|:--------|:-------|:------|:------|
|
| 124 |
+
| Ancora | NER | 13,581 | 10,628 | 1,427 | 1,526 |
|
| 125 |
+
| Ancora | POS | 16,678 | 13,123 | 1,709 | 1,846 |
|
| 126 |
+
| STS-ca | STS | 3,073 | 2,073 | 500 | 500 |
|
| 127 |
+
| TeCla | TC | 137,775 | 110,203| 13,786| 13,786|
|
| 128 |
+
| TE-ca | TE | 21,163 | 16,930 | 2,116 | 2,117 |
|
| 129 |
+
| VilaQuAD | QA | 6,282 | 3,882 | 1,200 | 1,200 |
|
| 130 |
+
| ViquiQuAD | QA | 14,239 | 11,255 | 1,492 | 1,429 |
|
| 131 |
+
| CatalanQA | QA | 21,427 | 17,135 | 2,157 | 2,135 |
|
| 132 |
|
| 133 |
### Evaluation results
|
| 134 |
|
| 135 |
This is how it compares to the teacher model when fine-tuned on the same downstream tasks:
|
| 136 |
+
|
| 137 |
| Model \ Task| NER (F1) | POS (F1) | STS-ca (Comb) | TeCla (Acc.) | TEca (Acc.) | VilaQuAD (F1/EM)| ViquiQuAD (F1/EM) | CatalanQA (F1/EM) | XQuAD-ca <sup>1</sup> (F1/EM) |
|
| 138 |
| ------------|:-------------:| -----:|:------|:------|:-------|:------|:----|:----|:----|
|
| 139 |
| RoBERTa-large-ca-v2 | 89.82 | 99.02 | 83.41 | 75.46 | 83.61 | 89.34/75.50 | 89.20/75.77 | 90.72/79.06 | 73.79/55.34 |
|