File size: 4,838 Bytes
be32629
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
---
annotations_creators:
  - derived
language:
  - eng
license: other
license_name: statcan-dialogue-retrieval-license
license_link: >-
  https://huggingface.co/datasets/McGill-NLP/statcan-dialogue-dataset-retrieval/blob/main/LICENSE.md
multilinguality: monolingual
task_categories:
  - text-retrieval
task_ids:
  - document-retrieval
tags:
  - table-retrieval
  - text
pretty_name: StatCan
config_names:
  - default
  - queries
  - corpus

dataset_info:
  - config_name: default
    features:
      - name: qid
        dtype: string
      - name: did
        dtype: string
      - name: score
        dtype: int32
    splits:
      - name: test
        num_bytes: 43500
        num_examples: 870
  - config_name: queries
    features:
      - name: _id
        dtype: string
      - name: text
        dtype: string
    splits:
      - name: test_queries
        num_bytes: 829463
        num_examples: 870
  - config_name: corpus
    features:
      - name: _id
        dtype: string
      - name: title
        dtype: string
      - name: text
        dtype: string
    splits:
      - name: corpus
        num_bytes: 40220365
        num_examples: 5907

configs:
  - config_name: default
    data_files:
      - split: test
        path: test_qrels.jsonl
  - config_name: queries
    data_files:
      - split: test_queries
        path: test_queries.jsonl
  - config_name: corpus
    data_files:
      - split: corpus
        path: corpus.jsonl
---

# StatCan Retrieval

This dataset is part of a Table + Text retrieval benchmark. Includes queries and relevance judgments across test split(s), with corpus in 1 format(s): `corpus`.

## Configs

| Config | Description | Split(s) |
|---|---|---|
| `default` | Relevance judgments (qrels): `qid`, `did`, `score` | `test` |
| `queries` | Query IDs and text | `test_queries` |
| `corpus` | Plain text corpus: `_id`, `title`, `text` | `corpus` |

## TableIR Benchmark Statistics

| Dataset | Structured | #Train | #Dev | #Test | #Corpus |
|---|:---:|---:|---:|---:|---:|
| OpenWikiTables | ✓ | 53.8k | 6.6k | 6.6k | 24.7k |
| NQTables | ✓ | 9.6k | 1.1k | 1k | 170k |
| FeTaQA | ✓ | 7.3k | 1k | 2k | 10.3k |
| OTT-QA (small) | ✓ | 41.5k | 2.2k | -- | 8.8k |
| MultiHierTT | ✗ | -- | 929 | -- | 9.9k |
| AIT-QA | ✗ | -- | -- | 515 | 1.9k |
| StatcanRetrieval | ✗ | -- | -- | 870 | 5.9k |
| watsonxDocsQA | ✗ | -- | -- | 30 | 1.1k |

## Citation

If you use **TableIR Eval: Table-Text IR Evaluation Collection**, please cite:

```bibtex
@misc{doshi2026tableir,
  title        = {TableIR Eval: Table-Text IR Evaluation Collection},
  author       = {Doshi, Meet and Boni, Odellia and Kumar, Vishwajeet and Sen, Jaydeep and Joshi, Sachindra},
  year         = {2026},
  institution  = {IBM Research},
  howpublished = {https://huggingface.co/collections/ibm-research/table-text-ir-evaluation},
  note         = {Hugging Face dataset collection}
}
```

All credit goes to original authors. Please cite their work:
```bibtex
@inproceedings{lu-etal-2023-statcan,
    title = "The {S}tat{C}an Dialogue Dataset: Retrieving Data Tables through Conversations with Genuine Intents",
    author = "Lu, Xing Han  and
      Reddy, Siva  and
      de Vries, Harm",
    editor = "Vlachos, Andreas  and
      Augenstein, Isabelle",
    booktitle = "Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics",
    month = may,
    year = "2023",
    address = "Dubrovnik, Croatia",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2023.eacl-main.206/",
    doi = "10.18653/v1/2023.eacl-main.206",
    pages = "2799--2829",
    abstract = "We introduce the StatCan Dialogue Dataset consisting of 19,379 conversation turns between agents working at Statistics Canada and online users looking for published data tables. The conversations stem from genuine intents, are held in English or French, and lead to agents retrieving one of over 5000 complex data tables. Based on this dataset, we propose two tasks: (1) automatic retrieval of relevant tables based on a on-going conversation, and (2) automatic generation of appropriate agent responses at each turn. We investigate the difficulty of each task by establishing strong baselines. Our experiments on a temporal data split reveal that all models struggle to generalize to future conversations, as we observe a significant drop in performance across both tasks when we move from the validation to the test set. In addition, we find that response generation models struggle to decide when to return a table. Considering that the tasks pose significant challenges to existing models, we encourage the community to develop models for our task, which can be directly used to help knowledge workers find relevant tables for live chat users."
}
```