mathiascreutz
commited on
Commit
·
d28b703
1
Parent(s):
5eb35be
Minor modifications
Browse files
README.md
CHANGED
|
@@ -159,29 +159,44 @@ TODO: Add comment about larger and noisier sets being better for training.
|
|
| 159 |
|
| 160 |
### Data Instances
|
| 161 |
|
|
|
|
|
|
|
|
|
|
| 162 |
```
|
|
|
|
|
|
|
|
|
|
| 163 |
DatasetDict({
|
| 164 |
test: Dataset({
|
| 165 |
features: ['lang', 'sent1', 'sent2', 'annot_score', 'gem_id'],
|
| 166 |
-
num_rows:
|
| 167 |
})
|
| 168 |
validation: Dataset({
|
| 169 |
features: ['lang', 'sent1', 'sent2', 'annot_score', 'gem_id'],
|
| 170 |
-
num_rows:
|
| 171 |
})
|
| 172 |
test.full: Dataset({
|
| 173 |
features: ['lang', 'sent1', 'sent2', 'annot_score', 'gem_id'],
|
| 174 |
-
num_rows:
|
| 175 |
})
|
| 176 |
validation.full: Dataset({
|
| 177 |
features: ['lang', 'sent1', 'sent2', 'annot_score', 'gem_id'],
|
| 178 |
-
num_rows:
|
| 179 |
})
|
| 180 |
train: Dataset({
|
| 181 |
features: ['lang', 'sent1', 'sent2', 'annot_score', 'gem_id'],
|
| 182 |
-
num_rows:
|
| 183 |
})
|
| 184 |
})
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 185 |
```
|
| 186 |
|
| 187 |
### Data Fields
|
|
|
|
| 159 |
|
| 160 |
### Data Instances
|
| 161 |
|
| 162 |
+
As a concrete example, loading the English data with a 95% quality of
|
| 163 |
+
the train split produces the following:
|
| 164 |
+
|
| 165 |
```
|
| 166 |
+
>>> data = load_dataset("GEM/opusparcus", lang="en", quality=95)
|
| 167 |
+
|
| 168 |
+
>>> data
|
| 169 |
DatasetDict({
|
| 170 |
test: Dataset({
|
| 171 |
features: ['lang', 'sent1', 'sent2', 'annot_score', 'gem_id'],
|
| 172 |
+
num_rows: 982
|
| 173 |
})
|
| 174 |
validation: Dataset({
|
| 175 |
features: ['lang', 'sent1', 'sent2', 'annot_score', 'gem_id'],
|
| 176 |
+
num_rows: 1015
|
| 177 |
})
|
| 178 |
test.full: Dataset({
|
| 179 |
features: ['lang', 'sent1', 'sent2', 'annot_score', 'gem_id'],
|
| 180 |
+
num_rows: 1445
|
| 181 |
})
|
| 182 |
validation.full: Dataset({
|
| 183 |
features: ['lang', 'sent1', 'sent2', 'annot_score', 'gem_id'],
|
| 184 |
+
num_rows: 1455
|
| 185 |
})
|
| 186 |
train: Dataset({
|
| 187 |
features: ['lang', 'sent1', 'sent2', 'annot_score', 'gem_id'],
|
| 188 |
+
num_rows: 1000000
|
| 189 |
})
|
| 190 |
})
|
| 191 |
+
|
| 192 |
+
>>> data["test"][0]
|
| 193 |
+
{'annot_score': 4.0, 'gem_id': 'gem-opusparcus-test-1587', 'lang': 'en', 'sent1': "I haven 't been contacted by anybody .", 'sent2': "Nobody 's contacted me ."}
|
| 194 |
+
|
| 195 |
+
>>> data["validation"][2]
|
| 196 |
+
{'annot_score': 3.0, 'gem_id': 'gem-opusparcus-validation-1586', 'lang': 'en', 'sent1': 'No promises , okay ?', 'sent2': "I 'm not promising anything ."}
|
| 197 |
+
|
| 198 |
+
>>> data["train"][1000]
|
| 199 |
+
{'annot_score': 0.0, 'gem_id': 'gem-opusparcus-train-12501001', 'lang': 'en', 'sent1': 'Am I beautiful ?', 'sent2': 'Am I pretty ?'}
|
| 200 |
```
|
| 201 |
|
| 202 |
### Data Fields
|