Datasets:
Tasks:
Question Answering
Modalities:
Text
Formats:
parquet
Sub-tasks:
extractive-qa
Languages:
Catalan
Size:
10K - 100K
ArXiv:
License:
Commit
·
a876915
1
Parent(s):
f0861bf
Update README.md
Browse files
README.md
CHANGED
|
@@ -52,9 +52,9 @@ task_ids:
|
|
| 52 |
|
| 53 |
### Dataset Summary
|
| 54 |
|
| 55 |
-
This dataset can be used to build extractive-QA and Language Models. It is an aggregation and balancing of 2 previous datasets:
|
| 56 |
|
| 57 |
-
Splits have been balanced by kind of question, and unlike other datasets like SQuAD, it only contains, per record, one question and one answer for each context, although the contexts can repeat multiple times.
|
| 58 |
|
| 59 |
### Supported Tasks and Leaderboards
|
| 60 |
Extractive-QA, Language Model.
|
|
@@ -88,7 +88,7 @@ The dataset is in Catalan (`ca-CA`).
|
|
| 88 |
```
|
| 89 |
|
| 90 |
### Data Fields
|
| 91 |
-
Follows [Rajpurkar, Pranav et al., 2016](http://arxiv.org/abs/1606.05250) for
|
| 92 |
|
| 93 |
- `id` (str): Unique ID assigned to the question.
|
| 94 |
- `title` (str): Title of the article.
|
|
@@ -120,7 +120,7 @@ Volunteers from [Catalan Wikipedia](ca.wikipedia.org) and professional journalis
|
|
| 120 |
|
| 121 |
### Annotations
|
| 122 |
#### Annotation process
|
| 123 |
-
We commissioned the creation of 1 to 5 questions for each context, following an adaptation of the guidelines from
|
| 124 |
|
| 125 |
#### Who are the annotators?
|
| 126 |
Annotation was commissioned by a specialized company that hired a team of native language speakers.
|
|
|
|
| 52 |
|
| 53 |
### Dataset Summary
|
| 54 |
|
| 55 |
+
This dataset can be used to build extractive-QA and Language Models. It is an aggregation and balancing of 2 previous datasets: [VilaQuAD](https://huggingface.co/datasets/projecte-aina/vilaquad) and [ViquiQuAD](https://huggingface.co/datasets/projecte-aina/viquiquad), which were described in [The Catalan Language CLUB](https://arxiv.org/abs/2112.01894).
|
| 56 |
|
| 57 |
+
Splits have been balanced by kind of question, and unlike other datasets like [SQuAD](http://arxiv.org/abs/1606.05250), it only contains, per record, one question and one answer for each context, although the contexts can repeat multiple times.
|
| 58 |
|
| 59 |
### Supported Tasks and Leaderboards
|
| 60 |
Extractive-QA, Language Model.
|
|
|
|
| 88 |
```
|
| 89 |
|
| 90 |
### Data Fields
|
| 91 |
+
Follows [Rajpurkar, Pranav et al., 2016](http://arxiv.org/abs/1606.05250) for SQuAD v1 datasets:
|
| 92 |
|
| 93 |
- `id` (str): Unique ID assigned to the question.
|
| 94 |
- `title` (str): Title of the article.
|
|
|
|
| 120 |
|
| 121 |
### Annotations
|
| 122 |
#### Annotation process
|
| 123 |
+
We commissioned the creation of 1 to 5 questions for each context, following an adaptation of the guidelines from SQuAD 1.0 ([Rajpurkar, Pranav et al. “SQuAD: 100, 000+ Questions for Machine Comprehension of Text.” EMNLP (2016)](http://arxiv.org/abs/1606.05250)).
|
| 124 |
|
| 125 |
#### Who are the annotators?
|
| 126 |
Annotation was commissioned by a specialized company that hired a team of native language speakers.
|