--- annotations_creators: - derived language: - eng license: cc-by-4.0 multilinguality: monolingual task_categories: - text-retrieval task_ids: - document-retrieval tags: - table-retrieval - text pretty_name: OTT-QA config_names: - default - queries - corpus_linearized - corpus_md - corpus_structure dataset_info: - config_name: default features: - name: qid dtype: string - name: did dtype: string - name: score dtype: int32 splits: - name: dev num_bytes: 185688 num_examples: 2214 - config_name: queries features: - name: _id dtype: string - name: text dtype: string splits: - name: dev_queries num_bytes: 336275 num_examples: 2214 - config_name: corpus_linearized features: - name: _id dtype: string - name: title dtype: string - name: text dtype: string splits: - name: corpus_linearized num_bytes: 19907385 num_examples: 8891 - config_name: corpus_md features: - name: _id dtype: string - name: title dtype: string - name: text dtype: string splits: - name: corpus_md num_bytes: 22363801 num_examples: 8891 - config_name: corpus_structure features: - name: _id dtype: string - name: title dtype: string - name: text dtype: string - name: meta_data dtype: string - name: headers sequence: string - name: cells sequence: string splits: - name: corpus_structure num_bytes: 41759044 num_examples: 8891 configs: - config_name: default data_files: - split: dev path: dev_qrels.jsonl - config_name: queries data_files: - split: dev_queries path: dev_queries.jsonl - config_name: corpus_linearized data_files: - split: corpus_linearized path: corpus_linearized.jsonl - config_name: corpus_md data_files: - split: corpus_md path: corpus_md.jsonl - config_name: corpus_structure data_files: - split: corpus_structure path: corpus_structure.jsonl --- # OTT-QA Retrieval This dataset is part of a Table + Text retrieval benchmark. Includes queries and relevance judgments across dev split(s), with corpus in 3 format(s): `corpus_linearized`, `corpus_md`, `corpus_structure`. ## Configs | Config | Description | Split(s) | |---|---|---| | `default` | Relevance judgments (qrels): `qid`, `did`, `score` | `dev` | | `queries` | Query IDs and text | `dev_queries` | | `corpus_linearized` | Linearized table representation | `corpus_linearized` | | `corpus_md` | Markdown table representation | `corpus_md` | | `corpus_structure` | Structured corpus with `headers`, `cells`, `meta_data`. `text` field corresponds to linearized Text + Table. | `corpus_structure` | ## `corpus_structure` additional fields | Field | Type | Description | |---|---|---| | `meta_data` | string | Table metadata / caption | | `headers` | list[string] | Column headers | | `cells` | list[string] | Flattened cell values | ## TableIR Benchmark Statistics | Dataset | Structured | #Train | #Dev | #Test | #Corpus | |---|:---:|---:|---:|---:|---:| | OpenWikiTables | ✓ | 53.8k | 6.6k | 6.6k | 24.7k | | NQTables | ✓ | 9.6k | 1.1k | 1k | 170k | | FeTaQA | ✓ | 7.3k | 1k | 2k | 10.3k | | OTT-QA (small) | ✓ | 41.5k | 2.2k | -- | 8.8k | | MultiHierTT | ✗ | -- | 929 | -- | 9.9k | | AIT-QA | ✗ | -- | -- | 515 | 1.9k | | StatcanRetrieval | ✗ | -- | -- | 870 | 5.9k | | watsonxDocsQA | ✗ | -- | -- | 30 | 1.1k | ## Citation If you use **TableIR Eval: Table-Text IR Evaluation Collection**, please cite: ```bibtex @misc{doshi2026tableir, title = {TableIR Eval: Table-Text IR Evaluation Collection}, author = {Doshi, Meet and Boni, Odellia and Kumar, Vishwajeet and Sen, Jaydeep and Joshi, Sachindra}, year = {2026}, institution = {IBM Research}, howpublished = {https://huggingface.co/collections/ibm-research/table-text-ir-evaluation}, note = {Hugging Face dataset collection} } ``` All credit goes to original authors. Please cite their work: ```bibtex @article{chen2021ottqa, title={Open Question Answering over Tables and Text}, author={Wenhu Chen, Ming-wei Chang, Eva Schlinger, William Wang, William Cohen}, journal={Proceedings of ICLR 2021}, year={2021} } ```