Datasets:
Tasks:
Text Retrieval
Modalities:
Text
Formats:
json
Sub-tasks:
document-retrieval
Languages:
English
Size:
10K - 100K
License:
| annotations_creators: | |
| - derived | |
| language: | |
| - eng | |
| license: cc-by-4.0 | |
| multilinguality: monolingual | |
| task_categories: | |
| - text-retrieval | |
| task_ids: | |
| - document-retrieval | |
| tags: | |
| - table-retrieval | |
| - text | |
| pretty_name: OTT-QA | |
| config_names: | |
| - default | |
| - queries | |
| - corpus_linearized | |
| - corpus_md | |
| - corpus_structure | |
| dataset_info: | |
| - config_name: default | |
| features: | |
| - name: qid | |
| dtype: string | |
| - name: did | |
| dtype: string | |
| - name: score | |
| dtype: int32 | |
| splits: | |
| - name: dev | |
| num_bytes: 185688 | |
| num_examples: 2214 | |
| - config_name: queries | |
| features: | |
| - name: _id | |
| dtype: string | |
| - name: text | |
| dtype: string | |
| splits: | |
| - name: dev_queries | |
| num_bytes: 336275 | |
| num_examples: 2214 | |
| - config_name: corpus_linearized | |
| features: | |
| - name: _id | |
| dtype: string | |
| - name: title | |
| dtype: string | |
| - name: text | |
| dtype: string | |
| splits: | |
| - name: corpus_linearized | |
| num_bytes: 19907385 | |
| num_examples: 8891 | |
| - config_name: corpus_md | |
| features: | |
| - name: _id | |
| dtype: string | |
| - name: title | |
| dtype: string | |
| - name: text | |
| dtype: string | |
| splits: | |
| - name: corpus_md | |
| num_bytes: 22363801 | |
| num_examples: 8891 | |
| - config_name: corpus_structure | |
| features: | |
| - name: _id | |
| dtype: string | |
| - name: title | |
| dtype: string | |
| - name: text | |
| dtype: string | |
| - name: meta_data | |
| dtype: string | |
| - name: headers | |
| sequence: string | |
| - name: cells | |
| sequence: string | |
| splits: | |
| - name: corpus_structure | |
| num_bytes: 41759044 | |
| num_examples: 8891 | |
| configs: | |
| - config_name: default | |
| data_files: | |
| - split: dev | |
| path: dev_qrels.jsonl | |
| - config_name: queries | |
| data_files: | |
| - split: dev_queries | |
| path: dev_queries.jsonl | |
| - config_name: corpus_linearized | |
| data_files: | |
| - split: corpus_linearized | |
| path: corpus_linearized.jsonl | |
| - config_name: corpus_md | |
| data_files: | |
| - split: corpus_md | |
| path: corpus_md.jsonl | |
| - config_name: corpus_structure | |
| data_files: | |
| - split: corpus_structure | |
| path: corpus_structure.jsonl | |
| # OTT-QA Retrieval | |
| This dataset is part of a Table + Text retrieval benchmark. Includes queries and relevance judgments across dev split(s), with corpus in 3 format(s): `corpus_linearized`, `corpus_md`, `corpus_structure`. | |
| ## Configs | |
| | Config | Description | Split(s) | | |
| |---|---|---| | |
| | `default` | Relevance judgments (qrels): `qid`, `did`, `score` | `dev` | | |
| | `queries` | Query IDs and text | `dev_queries` | | |
| | `corpus_linearized` | Linearized table representation | `corpus_linearized` | | |
| | `corpus_md` | Markdown table representation | `corpus_md` | | |
| | `corpus_structure` | Structured corpus with `headers`, `cells`, `meta_data`. `text` field corresponds to linearized Text + Table. | `corpus_structure` | | |
| ## `corpus_structure` additional fields | |
| | Field | Type | Description | | |
| |---|---|---| | |
| | `meta_data` | string | Table metadata / caption | | |
| | `headers` | list[string] | Column headers | | |
| | `cells` | list[string] | Flattened cell values | | |
| ## TableIR Benchmark Statistics | |
| | Dataset | Structured | #Train | #Dev | #Test | #Corpus | | |
| |---|:---:|---:|---:|---:|---:| | |
| | OpenWikiTables | ✓ | 53.8k | 6.6k | 6.6k | 24.7k | | |
| | NQTables | ✓ | 9.6k | 1.1k | 1k | 170k | | |
| | FeTaQA | ✓ | 7.3k | 1k | 2k | 10.3k | | |
| | OTT-QA (small) | ✓ | 41.5k | 2.2k | -- | 8.8k | | |
| | MultiHierTT | ✗ | -- | 929 | -- | 9.9k | | |
| | AIT-QA | ✗ | -- | -- | 515 | 1.9k | | |
| | StatcanRetrieval | ✗ | -- | -- | 870 | 5.9k | | |
| | watsonxDocsQA | ✗ | -- | -- | 30 | 1.1k | | |
| ## Citation | |
| If you use **TableIR Eval: Table-Text IR Evaluation Collection**, please cite: | |
| ```bibtex | |
| @misc{doshi2026tableir, | |
| title = {TableIR Eval: Table-Text IR Evaluation Collection}, | |
| author = {Doshi, Meet and Boni, Odellia and Kumar, Vishwajeet and Sen, Jaydeep and Joshi, Sachindra}, | |
| year = {2026}, | |
| institution = {IBM Research}, | |
| howpublished = {https://huggingface.co/collections/ibm-research/table-text-ir-evaluation}, | |
| note = {Hugging Face dataset collection} | |
| } | |
| ``` | |
| All credit goes to original authors. Please cite their work: | |
| ```bibtex | |
| @article{chen2021ottqa, | |
| title={Open Question Answering over Tables and Text}, | |
| author={Wenhu Chen, Ming-wei Chang, Eva Schlinger, William Wang, William Cohen}, | |
| journal={Proceedings of ICLR 2021}, | |
| year={2021} | |
| } | |
| ``` | |