Table + Text IR Evaluation
Collection
An evaluation suite created for benchmarking of retrieval models on Table+Text retrieval datasets.
•
8 items
•
Updated
qid
stringlengths 1
5
| did
stringlengths 36
42
| score
int32 1
1
|
|---|---|---|
2206
|
totto_source/dev_json/example-2205.json
| 1
|
15555
|
totto_source/train_json/example-7854.json
| 1
|
561
|
totto_source/dev_json/example-560.json
| 1
|
9835
|
totto_source/train_json/example-2134.json
| 1
|
21418
|
totto_source/train_json/example-13717.json
| 1
|
8032
|
totto_source/train_json/example-331.json
| 1
|
431
|
totto_source/dev_json/example-430.json
| 1
|
20853
|
totto_source/train_json/example-13152.json
| 1
|
767
|
totto_source/dev_json/example-766.json
| 1
|
1812
|
totto_source/dev_json/example-1811.json
| 1
|
15064
|
totto_source/train_json/example-7363.json
| 1
|
1834
|
totto_source/dev_json/example-1833.json
| 1
|
21624
|
totto_source/train_json/example-13923.json
| 1
|
16730
|
totto_source/train_json/example-9029.json
| 1
|
10064
|
totto_source/train_json/example-2363.json
| 1
|
21159
|
totto_source/train_json/example-13458.json
| 1
|
12617
|
totto_source/train_json/example-4916.json
| 1
|
21070
|
totto_source/train_json/example-13369.json
| 1
|
13839
|
totto_source/train_json/example-6138.json
| 1
|
20930
|
totto_source/train_json/example-13229.json
| 1
|
12516
|
totto_source/train_json/example-4815.json
| 1
|
17544
|
totto_source/train_json/example-9843.json
| 1
|
21045
|
totto_source/train_json/example-13344.json
| 1
|
10083
|
totto_source/train_json/example-2382.json
| 1
|
1789
|
totto_source/dev_json/example-1788.json
| 1
|
17067
|
totto_source/train_json/example-9366.json
| 1
|
11595
|
totto_source/train_json/example-3894.json
| 1
|
10549
|
totto_source/train_json/example-2848.json
| 1
|
2034
|
totto_source/dev_json/example-2033.json
| 1
|
1770
|
totto_source/dev_json/example-1769.json
| 1
|
8524
|
totto_source/train_json/example-823.json
| 1
|
33
|
totto_source/dev_json/example-32.json
| 1
|
7783
|
totto_source/train_json/example-82.json
| 1
|
9853
|
totto_source/train_json/example-2152.json
| 1
|
1262
|
totto_source/dev_json/example-1261.json
| 1
|
8567
|
totto_source/train_json/example-866.json
| 1
|
1804
|
totto_source/dev_json/example-1803.json
| 1
|
12122
|
totto_source/train_json/example-4421.json
| 1
|
8270
|
totto_source/train_json/example-569.json
| 1
|
10502
|
totto_source/train_json/example-2801.json
| 1
|
9243
|
totto_source/train_json/example-1542.json
| 1
|
2255
|
totto_source/dev_json/example-2254.json
| 1
|
11703
|
totto_source/train_json/example-4002.json
| 1
|
21257
|
totto_source/train_json/example-13556.json
| 1
|
9053
|
totto_source/train_json/example-1352.json
| 1
|
17677
|
totto_source/train_json/example-9976.json
| 1
|
2127
|
totto_source/dev_json/example-2126.json
| 1
|
15547
|
totto_source/train_json/example-7846.json
| 1
|
320
|
totto_source/dev_json/example-319.json
| 1
|
20746
|
totto_source/train_json/example-13045.json
| 1
|
21526
|
totto_source/train_json/example-13825.json
| 1
|
1431
|
totto_source/dev_json/example-1430.json
| 1
|
1516
|
totto_source/dev_json/example-1515.json
| 1
|
16773
|
totto_source/train_json/example-9072.json
| 1
|
12510
|
totto_source/train_json/example-4809.json
| 1
|
9796
|
totto_source/train_json/example-2095.json
| 1
|
20887
|
totto_source/train_json/example-13186.json
| 1
|
9080
|
totto_source/train_json/example-1379.json
| 1
|
717
|
totto_source/dev_json/example-716.json
| 1
|
21182
|
totto_source/train_json/example-13481.json
| 1
|
7921
|
totto_source/train_json/example-220.json
| 1
|
10226
|
totto_source/train_json/example-2525.json
| 1
|
17674
|
totto_source/train_json/example-9973.json
| 1
|
11222
|
totto_source/train_json/example-3521.json
| 1
|
1811
|
totto_source/dev_json/example-1810.json
| 1
|
21241
|
totto_source/train_json/example-13540.json
| 1
|
14571
|
totto_source/train_json/example-6870.json
| 1
|
21603
|
totto_source/train_json/example-13902.json
| 1
|
1507
|
totto_source/dev_json/example-1506.json
| 1
|
12107
|
totto_source/train_json/example-4406.json
| 1
|
21092
|
totto_source/train_json/example-13391.json
| 1
|
14990
|
totto_source/train_json/example-7289.json
| 1
|
11319
|
totto_source/train_json/example-3618.json
| 1
|
8404
|
totto_source/train_json/example-703.json
| 1
|
20761
|
totto_source/train_json/example-13060.json
| 1
|
11374
|
totto_source/train_json/example-3673.json
| 1
|
21057
|
totto_source/train_json/example-13356.json
| 1
|
10055
|
totto_source/train_json/example-2354.json
| 1
|
13051
|
totto_source/train_json/example-5350.json
| 1
|
1870
|
totto_source/dev_json/example-1869.json
| 1
|
1775
|
totto_source/dev_json/example-1774.json
| 1
|
2289
|
totto_source/dev_json/example-2288.json
| 1
|
706
|
totto_source/dev_json/example-705.json
| 1
|
20704
|
totto_source/train_json/example-13003.json
| 1
|
7867
|
totto_source/train_json/example-166.json
| 1
|
8564
|
totto_source/train_json/example-863.json
| 1
|
11699
|
totto_source/train_json/example-3998.json
| 1
|
1792
|
totto_source/dev_json/example-1791.json
| 1
|
21407
|
totto_source/train_json/example-13706.json
| 1
|
945
|
totto_source/dev_json/example-944.json
| 1
|
7932
|
totto_source/train_json/example-231.json
| 1
|
8302
|
totto_source/train_json/example-601.json
| 1
|
12409
|
totto_source/train_json/example-4708.json
| 1
|
9726
|
totto_source/train_json/example-2025.json
| 1
|
21213
|
totto_source/train_json/example-13512.json
| 1
|
7979
|
totto_source/train_json/example-278.json
| 1
|
12360
|
totto_source/train_json/example-4659.json
| 1
|
21594
|
totto_source/train_json/example-13893.json
| 1
|
1488
|
totto_source/dev_json/example-1487.json
| 1
|
11646
|
totto_source/train_json/example-3945.json
| 1
|
Table retrieval benchmark dataset. Includes queries and relevance judgments across test split(s), with corpus in 2 format(s): corpus_md, corpus_structure.
| Config | Description | Split(s) |
|---|---|---|
default |
Relevance judgments (qrels): qid, did, score |
test |
queries |
Query IDs and text | test_queries |
corpus_md |
Markdown table representation | corpus_md |
corpus_structure |
Structured corpus with headers, cells, meta_data. text field corresponds to linearized Text + Table. |
corpus_structure |
corpus_structure additional fields
| Field | Type | Description |
|---|---|---|
meta_data |
string | Table metadata / caption |
headers |
list[string] | Column headers |
cells |
list[string] | Flattened cell values |
| Dataset | Structured | #Train | #Dev | #Test | #Corpus |
|---|---|---|---|---|---|
| OpenWikiTables | ✓ | 53.8k | 6.6k | 6.6k | 24.7k |
| NQTables | ✓ | 9.6k | 1.1k | 1k | 170k |
| FeTaQA | ✓ | 7.3k | 1k | 2k | 10.3k |
| OTT-QA (small) | ✓ | 41.5k | 2.2k | -- | 8.8k |
| MultiHierTT | ✗ | -- | 929 | -- | 9.9k |
| AIT-QA | ✗ | -- | -- | 515 | 1.9k |
| StatcanRetrieval | ✗ | -- | -- | 870 | 5.9k |
| watsonxDocsQA | ✗ | -- | -- | 30 | 1.1k |
If you use TableIR Eval: Table-Text IR Evaluation Collection, please cite:
@misc{doshi2026tableir,
title = {TableIR Eval: Table-Text IR Evaluation Collection},
author = {Doshi, Meet and Boni, Odellia and Kumar, Vishwajeet and Sen, Jaydeep and Joshi, Sachindra},
year = {2026},
institution = {IBM Research},
howpublished = {https://huggingface.co/collections/ibm-research/table-text-ir-evaluation},
note = {Hugging Face dataset collection}
}
All credit goes to original authors. Please cite their work:
@article{Nan2021FeTaQAFT,
title={FeTaQA: Free-form Table Question Answering},
author={Nan, Linyong and Hsieh, Chiachun and Mao, Ziming and Lin, Xi Victoria and Verma, Neha and Zhang, Rui and Kryściński, Wojciech and Schoelkopf, Hailey and Kong, Riley and Tang, Xiangru and Mutuma, Mutethia and Rosand, Ben and Trindade, Isabel and Bandaru, Renusree and Cunningham, Jacob and Xiong, Caiming and Radev, Dragomir},
journal={Transactions of the Association for Computational Linguistics},
year={2022},
volume={10},
pages={35-49}
}