Deva8 commited on
Commit
c95d978
·
verified ·
1 Parent(s): e61b9ed

updated readme

Browse files
Files changed (1) hide show
  1. README.md +83 -3
README.md CHANGED
@@ -1,4 +1,85 @@
1
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  license: mit
3
  task_categories:
4
  - visual-question-answering
@@ -7,5 +88,4 @@ tags:
7
  - code
8
  pretty_name: Vqa-Dataset for generative tasks
9
  size_categories:
10
- - 10K<n<100K
11
- ---
 
1
+
2
+
3
+ # Generative-VQA-V2-Curated
4
+
5
+ A curated, balanced, and cleaned version of the VQA v2 dataset specifically optimized for **Generative Visual Question Answering**.
6
+
7
+ This dataset transforms the standard VQA task into a generative challenge by removing "yes/no" shortcuts and balancing answer distributions to prevent model over-fitting on dominant classes.
8
+
9
+ ## Dataset Summary
10
+
11
+ The primary goal of this curated set is to provide a "clean" signal for training multimodal models by:
12
+
13
+ 1. **Eliminating Binary Biases:** Removed all "yes/no" and "unknown" style answers.
14
+ 2. **Balancing Classes:** Capped samples at **600 per answer** to ensure the model learns a diverse vocabulary.
15
+ 3. **Filtering Ambiguity:** Removed generic questions (e.g., "What is this?") to focus on specific visual grounding.
16
+
17
+ ## Dataset Statistics
18
+
19
+ * **Total QA Pairs:** 135,268
20
+ * **Unique Answer Classes:** 1,251
21
+ * **Source Images:** COCO Train 2014
22
+ * **Minimum Frequency per Answer:** 20
23
+ * **Maximum Samples per Answer:** 600
24
+
25
+ ## 🛠 Curation Logic
26
+
27
+ The dataset was generated using the following filtering pipeline:
28
+
29
+ * **Consensus-Based:** Only the majority-vote answer from the 10 human annotators is used.
30
+ * **Exclusion List:** * Boolean: `yes`, `no`
31
+ * Uncertainty: `unknown`, `none`, `n/a`, `cant tell`, `not sure`
32
+
33
+
34
+ * **Ambiguity Filter:** Removed questions containing "what is in the image", "what is this", "what is that", or "what do you see".
35
+ * **Conciseness:** Answers are restricted to **5 words** and **30 characters** or fewer.
36
+
37
+ ## Structure
38
+
39
+ ```text
40
+ .
41
+ ├── images/ # Curated COCO images (JPG)
42
+ ├── qa_pairs.json # Full metadata (JSON)
43
+ ├── metadata.csv # Metadata (CSV) for easy loading
44
+ └── README.md
45
+
46
+ ```
47
+
48
+ ### Metadata Fields
49
+
50
+ * `image_id`: Original COCO Image ID.
51
+ * `question_id`: Original VQA v2 Question ID.
52
+ * `question`: The natural language question.
53
+ * `answer`: The curated ground-truth answer.
54
+ * `image_path`: Path relative to the dataset root.
55
+
56
+ ## Usage
57
+
58
+ ### Loading with Hugging Face `datasets`
59
+
60
+ ```python
61
+ from datasets import load_dataset
62
+
63
+ dataset = load_dataset("Deva8/Generative-VQA-V2-Curated")
64
+
65
+ ```
66
+
67
+ ### Quick Look (Pandas)
68
+
69
+ ```python
70
+ import pandas as pd
71
+ df = pd.read_csv("metadata.csv")
72
+ print(df['answer'].value_counts().head(10)) # Check balancing
73
+
74
+ ```
75
+
76
+ ## License & Attribution
77
+
78
+ This dataset is a derivative work of the **VQA v2 Dataset** and the **COCO Dataset**.
79
+
80
+ * **Images:** [COCO Consortium (CC BY 4.0)](https://www.google.com/search?q=https://cocodataset.org/%23termsofuse)
81
+ * **Annotations:** [VQA v2 (CC BY 4.0)](https://visualqa.org/download.html)
82
+
83
  license: mit
84
  task_categories:
85
  - visual-question-answering
 
88
  - code
89
  pretty_name: Vqa-Dataset for generative tasks
90
  size_categories:
91
+ - 10K<n<100K