Datasets:
Tasks:
Text Retrieval
Modalities:
Text
Formats:
json
Sub-tasks:
document-retrieval
Languages:
English
Size:
10K - 100K
License:
metadata
annotations_creators:
- derived
language:
- eng
license: cc-by-4.0
multilinguality: monolingual
task_categories:
- text-retrieval
task_ids:
- document-retrieval
tags:
- table-retrieval
- text
pretty_name: OTT-QA
config_names:
- default
- queries
- corpus_linearized
- corpus_md
- corpus_structure
dataset_info:
- config_name: default
features:
- name: qid
dtype: string
- name: did
dtype: string
- name: score
dtype: int32
splits:
- name: dev
num_bytes: 185688
num_examples: 2214
- config_name: queries
features:
- name: _id
dtype: string
- name: text
dtype: string
splits:
- name: dev_queries
num_bytes: 336275
num_examples: 2214
- config_name: corpus_linearized
features:
- name: _id
dtype: string
- name: title
dtype: string
- name: text
dtype: string
splits:
- name: corpus_linearized
num_bytes: 19907385
num_examples: 8891
- config_name: corpus_md
features:
- name: _id
dtype: string
- name: title
dtype: string
- name: text
dtype: string
splits:
- name: corpus_md
num_bytes: 22363801
num_examples: 8891
- config_name: corpus_structure
features:
- name: _id
dtype: string
- name: title
dtype: string
- name: text
dtype: string
- name: meta_data
dtype: string
- name: headers
sequence: string
- name: cells
sequence: string
splits:
- name: corpus_structure
num_bytes: 41759044
num_examples: 8891
configs:
- config_name: default
data_files:
- split: dev
path: dev_qrels.jsonl
- config_name: queries
data_files:
- split: dev_queries
path: dev_queries.jsonl
- config_name: corpus_linearized
data_files:
- split: corpus_linearized
path: corpus_linearized.jsonl
- config_name: corpus_md
data_files:
- split: corpus_md
path: corpus_md.jsonl
- config_name: corpus_structure
data_files:
- split: corpus_structure
path: corpus_structure.jsonl
OTT-QA Retrieval
This dataset is part of a Table + Text retrieval benchmark. Includes queries and relevance judgments across dev split(s), with corpus in 3 format(s): corpus_linearized, corpus_md, corpus_structure.
Configs
| Config | Description | Split(s) |
|---|---|---|
default |
Relevance judgments (qrels): qid, did, score |
dev |
queries |
Query IDs and text | dev_queries |
corpus_linearized |
Linearized table representation | corpus_linearized |
corpus_md |
Markdown table representation | corpus_md |
corpus_structure |
Structured corpus with headers, cells, meta_data. text field corresponds to linearized Text + Table. |
corpus_structure |
corpus_structure additional fields
| Field | Type | Description |
|---|---|---|
meta_data |
string | Table metadata / caption |
headers |
list[string] | Column headers |
cells |
list[string] | Flattened cell values |
TableIR Benchmark Statistics
| Dataset | Structured | #Train | #Dev | #Test | #Corpus |
|---|---|---|---|---|---|
| OpenWikiTables | ✓ | 53.8k | 6.6k | 6.6k | 24.7k |
| NQTables | ✓ | 9.6k | 1.1k | 1k | 170k |
| FeTaQA | ✓ | 7.3k | 1k | 2k | 10.3k |
| OTT-QA (small) | ✓ | 41.5k | 2.2k | -- | 8.8k |
| MultiHierTT | ✗ | -- | 929 | -- | 9.9k |
| AIT-QA | ✗ | -- | -- | 515 | 1.9k |
| StatcanRetrieval | ✗ | -- | -- | 870 | 5.9k |
| watsonxDocsQA | ✗ | -- | -- | 30 | 1.1k |
Citation
If you use TableIR Eval: Table-Text IR Evaluation Collection, please cite:
@misc{doshi2026tableir,
title = {TableIR Eval: Table-Text IR Evaluation Collection},
author = {Doshi, Meet and Boni, Odellia and Kumar, Vishwajeet and Sen, Jaydeep and Joshi, Sachindra},
year = {2026},
institution = {IBM Research},
howpublished = {https://huggingface.co/collections/ibm-research/table-text-ir-evaluation},
note = {Hugging Face dataset collection}
}
All credit goes to original authors. Please cite their work:
@article{chen2021ottqa,
title={Open Question Answering over Tables and Text},
author={Wenhu Chen, Ming-wei Chang, Eva Schlinger, William Wang, William Cohen},
journal={Proceedings of ICLR 2021},
year={2021}
}