Table + Text IR Evaluation
Collection
An evaluation suite created for benchmarking of retrieval models on Table+Text retrieval datasets.
•
8 items
•
Updated
qid
string | did
string | score
int32 |
|---|---|---|
wikisql-00022
|
9
| 1
|
wikisql-00023
|
9
| 1
|
wikisql-00024
|
9
| 1
|
wikisql-00025
|
9
| 1
|
wikisql-00144
|
75
| 1
|
wikisql-00145
|
75
| 1
|
wikisql-00146
|
75
| 1
|
wikisql-00147
|
75
| 1
|
wikisql-00148
|
75
| 1
|
wikisql-00149
|
75
| 1
|
wikisql-00291
|
167
| 1
|
wikisql-00292
|
167
| 1
|
wikisql-00293
|
167
| 1
|
wikisql-00294
|
167
| 1
|
wikisql-00295
|
167
| 1
|
wikisql-00324
|
181
| 1
|
wikisql-00324
|
182
| 1
|
wikisql-00324
|
183
| 1
|
wikisql-00324
|
184
| 1
|
wikisql-00325
|
181
| 1
|
wikisql-00326
|
183
| 1
|
wikisql-00327
|
181
| 1
|
wikisql-00327
|
182
| 1
|
wikisql-00327
|
184
| 1
|
wikisql-00426
|
220
| 1
|
wikisql-00427
|
219
| 1
|
wikisql-00429
|
219
| 1
|
wikisql-00445
|
229
| 1
|
wikisql-00447
|
230
| 1
|
wikisql-00508
|
252
| 1
|
wikisql-00512
|
259
| 1
|
wikisql-00513
|
259
| 1
|
wikisql-00514
|
258
| 1
|
wikisql-00515
|
259
| 1
|
wikisql-00516
|
259
| 1
|
wikisql-00517
|
258
| 1
|
wikisql-00518
|
256
| 1
|
wikisql-00518
|
257
| 1
|
wikisql-00519
|
256
| 1
|
wikisql-00521
|
256
| 1
|
wikisql-00525
|
261
| 1
|
wikisql-00526
|
262
| 1
|
wikisql-00526
|
263
| 1
|
wikisql-00526
|
261
| 1
|
wikisql-00527
|
261
| 1
|
wikisql-00528
|
263
| 1
|
wikisql-00529
|
263
| 1
|
wikisql-00529
|
261
| 1
|
wikisql-00529
|
262
| 1
|
wikisql-00550
|
276
| 1
|
wikisql-00552
|
275
| 1
|
wikisql-00553
|
276
| 1
|
wikisql-00554
|
275
| 1
|
wikisql-00555
|
275
| 1
|
wikisql-00610
|
309
| 1
|
wikisql-00740
|
383
| 1
|
wikisql-00741
|
383
| 1
|
wikisql-00742
|
383
| 1
|
wikisql-00743
|
383
| 1
|
wikisql-00744
|
383
| 1
|
wikisql-00807
|
418
| 1
|
wikisql-00808
|
418
| 1
|
wikisql-00809
|
417
| 1
|
wikisql-00810
|
417
| 1
|
wikisql-00959
|
494
| 1
|
wikisql-00960
|
495
| 1
|
wikisql-00961
|
494
| 1
|
wikisql-00962
|
495
| 1
|
wikisql-01043
|
559
| 1
|
wikisql-01044
|
558
| 1
|
wikisql-01045
|
557
| 1
|
wikisql-01046
|
556
| 1
|
wikisql-01047
|
556
| 1
|
wikisql-01048
|
556
| 1
|
wikisql-01191
|
619
| 1
|
wikisql-01357
|
734
| 1
|
wikisql-01358
|
733
| 1
|
wikisql-01359
|
733
| 1
|
wikisql-01360
|
733
| 1
|
wikisql-01423
|
763
| 1
|
wikisql-01485
|
798
| 1
|
wikisql-01486
|
799
| 1
|
wikisql-01491
|
804
| 1
|
wikisql-01492
|
804
| 1
|
wikisql-01493
|
804
| 1
|
wikisql-01514
|
808
| 1
|
wikisql-01517
|
808
| 1
|
wikisql-01518
|
808
| 1
|
wikisql-01562
|
834
| 1
|
wikisql-01565
|
841
| 1
|
wikisql-01567
|
841
| 1
|
wikisql-01567
|
842
| 1
|
wikisql-01602
|
867
| 1
|
wikisql-01603
|
867
| 1
|
wikisql-01604
|
867
| 1
|
wikisql-01613
|
874
| 1
|
wikisql-01614
|
874
| 1
|
wikisql-01621
|
875
| 1
|
wikisql-01622
|
875
| 1
|
wikisql-01623
|
876
| 1
|
This dataset is part of a Table + Text retrieval benchmark. Includes queries and relevance judgments across test split(s), with corpus in 3 format(s): corpus_linearized, corpus_md, corpus_structure.
| Config | Description | Split(s) |
|---|---|---|
default |
Relevance judgments (qrels): qid, did, score |
test |
queries |
Query IDs and text | test_queries |
corpus_linearized |
Linearized table representation | corpus_linearized |
corpus_md |
Markdown table representation | corpus_md |
corpus_structure |
Structured corpus with headers, cells, meta_data. text field corresponds to linearized Text + Table. |
corpus_structure |
corpus_structure additional fields
| Field | Type | Description |
|---|---|---|
meta_data |
string | Table metadata / caption |
headers |
list[string] | Column headers |
cells |
list[string] | Flattened cell values |
| Dataset | Structured | #Train | #Dev | #Test | #Corpus |
|---|---|---|---|---|---|
| OpenWikiTables | ✓ | 53.8k | 6.6k | 6.6k | 24.7k |
| NQTables | ✓ | 9.6k | 1.1k | 1k | 170k |
| FeTaQA | ✓ | 7.3k | 1k | 2k | 10.3k |
| OTT-QA (small) | ✓ | 41.5k | 2.2k | -- | 8.8k |
| MultiHierTT | ✗ | -- | 929 | -- | 9.9k |
| AIT-QA | ✗ | -- | -- | 515 | 1.9k |
| StatcanRetrieval | ✗ | -- | -- | 870 | 5.9k |
| watsonxDocsQA | ✗ | -- | -- | 30 | 1.1k |
If you use TableIR Eval: Table-Text IR Evaluation Collection, please cite:
@misc{doshi2026tableir,
title = {TableIR Eval: Table-Text IR Evaluation Collection},
author = {Doshi, Meet and Boni, Odellia and Kumar, Vishwajeet and Sen, Jaydeep and Joshi, Sachindra},
year = {2026},
institution = {IBM Research},
howpublished = {https://huggingface.co/collections/ibm-research/table-text-ir-evaluation},
note = {Hugging Face dataset collection}
}
All credit goes to original authors. Please cite their work:
@inproceedings{kweon-etal-2023-open,
title = "Open-{W}iki{T}able : Dataset for Open Domain Question Answering with Complex Reasoning over Table",
author = "Kweon, Sunjun and
Kwon, Yeonsu and
Cho, Seonhee and
Jo, Yohan and
Choi, Edward",
editor = "Rogers, Anna and
Boyd-Graber, Jordan and
Okazaki, Naoaki",
booktitle = "Findings of the Association for Computational Linguistics: ACL 2023",
month = jul,
year = "2023",
address = "Toronto, Canada",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2023.findings-acl.526/",
doi = "10.18653/v1/2023.findings-acl.526",
pages = "8285--8297",
abstract = "Despite recent interest in open domain question answering (ODQA) over tables, many studies still rely on datasets that are not truly optimal for the task with respect to utilizing structural nature of table. These datasets assume answers reside as a single cell value and do not necessitate exploring over multiple cells such as aggregation, comparison, and sorting. Thus, we release Open-WikiTable, the first ODQA dataset that requires complex reasoning over tables. Open-WikiTable is built upon WikiSQL and WikiTableQuestions to be applicable in the open-domain setting. As each question is coupled with both textual answers and SQL queries, Open-WikiTable opens up a wide range of possibilities for future research, as both reader and parser methods can be applied. The dataset is publicly available."
}