SarraChab's picture
Update README.md
0f8c70e verified
---
dataset_info:
features:
- name: id
dtype: string
- name: dataset
dtype: string
- name: question
dtype: string
- name: options
sequence: string
- name: answer
dtype: string
splits:
- name: train
num_bytes: 2268293
num_examples: 10687
download_size: 1254741
dataset_size: 2268293
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
---
This dataset was constructed as part of the EPFL Modern NLP (MNLP) course project to train and evaluate large language models on **multiple-choice question answering (MCQA)** tasks focused on scientific reasoning.
It aggregates and reformats **10,687 unique examples** from five high-quality academic and biomedical QA datasets, applying consistent structure, question normalization, and cross-source deduplication.
### 📊 Dataset Composition
| Source Dataset | Link | Questions Used | Description |
|----------------|------|----------------|-------------|
| ARC-Challenge | [ai2_arc](https://huggingface.co/datasets/ai2_arc) | 1,119 | Harder science exam questions requiring multi-step reasoning |
| ARC-Easy | [ai2_arc](https://huggingface.co/datasets/ai2_arc) | 2,251 | Simpler science questions at the elementary/middle school level |
| QASC | [qasc](https://huggingface.co/datasets/qasc) | 3,000 (subset) | A filtered and deduplicated subset of the QASC dataset, which was originally larger (~8,000+ examples). Only 3,000 unique and diverse questions were selected for balance |
| OpenBookQA | [openbookqa](https://huggingface.co/datasets/openbookqa) | 3,317 | 4-option science questions, filtered to keep `humanScore ≥ 1` |
| PubMedQA | [pubmed_qa](https://huggingface.co/datasets/pubmed_qa) | 1,000 | Biomedical questions with Yes/No/Maybe answers based on PubMed abstracts |
### 🧪 Preprocessing Pipeline
- **Normalization**: All questions were lowercased and stripped of whitespace for consistency.
- **Deduplication**: Each question was hashed (`md5(lowercase question)`) to detect and eliminate duplicates across datasets.
- **Filtering**:
- OpenBookQA was filtered to retain only questions with `humanScore ≥ 1`.
- PubMedQA was filtered to retain only labeled questions with answers in {yes, no, maybe}.
- QASC was **sampled and capped** at 3,000 unique questions to ensure dataset balance.
- **Unified formatting**: All entries follow the same JSON schema across sources.
### 📦 Format
Each sample follows this structure:
```json
{
"id": "qasc_481",
"dataset": "qasc",
"question": "What do bees use to make honey?",
"options": ["nectar", "pollen", "water", "leaves"],
"answer": "A"
}