Datasets:
Update README.md
Browse files
README.md
CHANGED
|
@@ -47,7 +47,7 @@ tags:
|
|
| 47 |
size_categories:
|
| 48 |
- 100K<n<1M
|
| 49 |
---
|
| 50 |
-
# GenDocVQA
|
| 51 |
|
| 52 |
This dataset provides a broad set of documents with questions related to their contents.
|
| 53 |
These questions are non-extractive, meaning that the model, which solves our task should be
|
|
@@ -62,7 +62,7 @@ generative and compute the answers by itself.
|
|
| 62 |
In order to load dataset using following code:
|
| 63 |
|
| 64 |
```python
|
| 65 |
-
ds = datasets.load_dataset('lenagibee/
|
| 66 |
```
|
| 67 |
|
| 68 |
ds is a dict consisting from two splits `train` and `validation`.
|
|
@@ -74,15 +74,15 @@ im = Image.open(ds['train'][0]['image_path'])
|
|
| 74 |
```
|
| 75 |
|
| 76 |
Dataset generator:
|
| 77 |
-
https://huggingface.co/datasets/lenagibee/
|
| 78 |
|
| 79 |
## Dataset Structure
|
| 80 |
|
| 81 |
All the necessary data is stored in the following archives:
|
| 82 |
|
| 83 |
-
* Images: https://huggingface.co/datasets/lenagibee/
|
| 84 |
-
* OCR: https://huggingface.co/datasets/lenagibee/
|
| 85 |
-
* Annotations: https://huggingface.co/datasets/lenagibee/
|
| 86 |
|
| 87 |
Data parsing is already implemented in the attached dataset generator.
|
| 88 |
Images should be processed by the user himself.
|
|
@@ -182,6 +182,15 @@ The questions from each dataset were filtered by the types of the questions,
|
|
| 182 |
leaving only non-extractive questions, related to one page. After that the questions
|
| 183 |
were paraphrased.
|
| 184 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 185 |
## Dataset Card Contact
|
| 186 |
Please feel free to contact in the community page of this dataset of via
|
| 187 |
the Telegram chat of the challenge:
|
|
|
|
| 47 |
size_categories:
|
| 48 |
- 100K<n<1M
|
| 49 |
---
|
| 50 |
+
# GenDocVQA
|
| 51 |
|
| 52 |
This dataset provides a broad set of documents with questions related to their contents.
|
| 53 |
These questions are non-extractive, meaning that the model, which solves our task should be
|
|
|
|
| 62 |
In order to load dataset using following code:
|
| 63 |
|
| 64 |
```python
|
| 65 |
+
ds = datasets.load_dataset('lenagibee/GenDocVQA')
|
| 66 |
```
|
| 67 |
|
| 68 |
ds is a dict consisting from two splits `train` and `validation`.
|
|
|
|
| 74 |
```
|
| 75 |
|
| 76 |
Dataset generator:
|
| 77 |
+
https://huggingface.co/datasets/lenagibee/GenDocVQA/resolve/main/GenDocVQA.py?download=true
|
| 78 |
|
| 79 |
## Dataset Structure
|
| 80 |
|
| 81 |
All the necessary data is stored in the following archives:
|
| 82 |
|
| 83 |
+
* Images: https://huggingface.co/datasets/lenagibee/GenDocVQA/resolve/main/archives/gendocvqa2024_imgs.tar.gz?download=true
|
| 84 |
+
* OCR: https://huggingface.co/datasets/lenagibee/GenDocVQA/resolve/main/archives/gendocvqa2024_ocr.tar.gz?download=true
|
| 85 |
+
* Annotations: https://huggingface.co/datasets/lenagibee/GenDocVQA/resolve/main/archives/gendocvqa2024_annotations.tar.gz?download=true
|
| 86 |
|
| 87 |
Data parsing is already implemented in the attached dataset generator.
|
| 88 |
Images should be processed by the user himself.
|
|
|
|
| 182 |
leaving only non-extractive questions, related to one page. After that the questions
|
| 183 |
were paraphrased.
|
| 184 |
|
| 185 |
+
### Source Data Licenses
|
| 186 |
+
|
| 187 |
+
The dataset adheres to the licenses of its constituents.
|
| 188 |
+
1. SlideVQA: https://github.com/nttmdlab-nlp/SlideVQA/blob/main/LICENSE
|
| 189 |
+
2. PDFVQA: https://github.com/adlnlp/pdfvqa (Unknown)
|
| 190 |
+
3. InfographicsVQA: https://www.docvqa.org/datasets/infographicvqa (Unknown)
|
| 191 |
+
4. TAT-DQA: https://nextplusplus.github.io/TAT-DQA/ (CC BY 4.0)
|
| 192 |
+
5. DUDE: https://github.com/duchallenge-team/dude/blob/main/LICENSE (GPL 3.0)
|
| 193 |
+
|
| 194 |
## Dataset Card Contact
|
| 195 |
Please feel free to contact in the community page of this dataset of via
|
| 196 |
the Telegram chat of the challenge:
|