mapama247 commited on
Commit
d987a1e
·
1 Parent(s): be87ee2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -10
README.md CHANGED
@@ -118,20 +118,22 @@ As an example, the distilled version of BERT has 40% fewer parameters and runs 6
118
  This model has been fine-tuned on the downstream tasks of the Catalan Language Understanding Evaluation benchmark (CLUB).
119
 
120
  Here are the train/dev/test splits of each dataset:
121
- | Dataset | Task | Total | Train | Dev | Test |
122
- |:--|:--|:--|:--|:--|:--|
123
- | Ancora | NER |13,581 | 10,628 | 1,427 | 1,526 |
124
- | Ancora | POS | 16,678 | 13,123 | 1,709 | 1,846 |
125
- | STS-ca | STS | 3,073 | 2,073 | 500 | 500 |
126
- | TeCla | TC | 137,775 | 110,203 | 13,786 | 13,786|
127
- | TE-ca | TE | 21,163 | 16,930 | 2,116 | 2,117
128
- | VilaQuAD | QA | 6,282 | 3,882 | 1,200 | 1,200 |
129
- | ViquiQuAD | QA | 14,239 | 11,255 | 1,492 | 1,429 |
130
- | CatalanQA | QA | 21,427 | 17,135 | 2,157 | 2,135 |
 
131
 
132
  ### Evaluation results
133
 
134
  This is how it compares to the teacher model when fine-tuned on the same downstream tasks:
 
135
  | Model \ Task| NER (F1) | POS (F1) | STS-ca (Comb) | TeCla (Acc.) | TEca (Acc.) | VilaQuAD (F1/EM)| ViquiQuAD (F1/EM) | CatalanQA (F1/EM) | XQuAD-ca <sup>1</sup> (F1/EM) |
136
  | ------------|:-------------:| -----:|:------|:------|:-------|:------|:----|:----|:----|
137
  | RoBERTa-large-ca-v2 | 89.82 | 99.02 | 83.41 | 75.46 | 83.61 | 89.34/75.50 | 89.20/75.77 | 90.72/79.06 | 73.79/55.34 |
 
118
  This model has been fine-tuned on the downstream tasks of the Catalan Language Understanding Evaluation benchmark (CLUB).
119
 
120
  Here are the train/dev/test splits of each dataset:
121
+
122
+ | Dataset | Task| Total | Train | Dev | Test |
123
+ |:----------|:----|:--------|:-------|:------|:------|
124
+ | Ancora | NER | 13,581 | 10,628 | 1,427 | 1,526 |
125
+ | Ancora | POS | 16,678 | 13,123 | 1,709 | 1,846 |
126
+ | STS-ca | STS | 3,073 | 2,073 | 500 | 500 |
127
+ | TeCla | TC | 137,775 | 110,203| 13,786| 13,786|
128
+ | TE-ca | TE | 21,163 | 16,930 | 2,116 | 2,117 |
129
+ | VilaQuAD | QA | 6,282 | 3,882 | 1,200 | 1,200 |
130
+ | ViquiQuAD | QA | 14,239 | 11,255 | 1,492 | 1,429 |
131
+ | CatalanQA | QA | 21,427 | 17,135 | 2,157 | 2,135 |
132
 
133
  ### Evaluation results
134
 
135
  This is how it compares to the teacher model when fine-tuned on the same downstream tasks:
136
+
137
  | Model \ Task| NER (F1) | POS (F1) | STS-ca (Comb) | TeCla (Acc.) | TEca (Acc.) | VilaQuAD (F1/EM)| ViquiQuAD (F1/EM) | CatalanQA (F1/EM) | XQuAD-ca <sup>1</sup> (F1/EM) |
138
  | ------------|:-------------:| -----:|:------|:------|:-------|:------|:----|:----|:----|
139
  | RoBERTa-large-ca-v2 | 89.82 | 99.02 | 83.41 | 75.46 | 83.61 | 89.34/75.50 | 89.20/75.77 | 90.72/79.06 | 73.79/55.34 |