Catkamakura commited on
Commit
cedb333
·
verified ·
1 Parent(s): f8079f1

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +115 -43
README.md CHANGED
@@ -1,45 +1,117 @@
1
  ---
2
- dataset_info:
3
- features:
4
- - name: id
5
- dtype: string
6
- - name: question
7
- dtype: string
8
- - name: context
9
- dtype: string
10
- - name: answers
11
- list: string
12
- - name: source
13
- dtype: string
14
- - name: question_timestamp
15
- dtype: string
16
- - name: evidence_timestamp
17
- dtype: string
18
- - name: dimensions
19
- dtype: string
20
- - name: is_hard_negative
21
- dtype: bool
22
- - name: original_question
23
- dtype: string
24
- splits:
25
- - name: train
26
- num_bytes: 87197610
27
- num_examples: 25041
28
- - name: validation
29
- num_bytes: 10292059
30
- num_examples: 2800
31
- - name: test
32
- num_bytes: 30863658
33
- num_examples: 9089
34
- download_size: 71628923
35
- dataset_size: 128353327
36
- configs:
37
- - config_name: default
38
- data_files:
39
- - split: train
40
- path: data/train-*
41
- - split: validation
42
- path: data/validation-*
43
- - split: test
44
- path: data/test-*
45
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ license: mit
3
+ task_categories:
4
+ - question-answering
5
+ - text-retrieval
6
+ language:
7
+ - en
8
+ tags:
9
+ - temporal-qa
10
+ - time-sensitive
11
+ - retrieval-augmented-generation
12
+ - timestamp
13
+ - streamingqa
14
+ - timeqa
15
+ size_categories:
16
+ - 10K<n<100K
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17
  ---
18
+
19
+ # Time-Sensitive QA Dataset (StreamingQA + TimeQA)
20
+
21
+ A time-sensitive question answering dataset combining **StreamingQA** and **TimeQA** for evaluating temporal reasoning in RAG systems.
22
+
23
+ ## Dataset Description
24
+
25
+ This dataset contains question-answer pairs where the correct answer depends on a specific timestamp.
26
+ Each question is prefixed with a date (e.g., "Today is Tuesday, September 24, 2013.") and the model
27
+ must use temporal context to provide the correct answer for that point in time.
28
+
29
+ ### Sources
30
+
31
+ - **StreamingQA** ([Liska et al., 2022](https://arxiv.org/abs/2205.02266)): Time-sensitive questions about news articles from the WMT corpus
32
+ - **TimeQA** ([Chen et al., 2021](https://arxiv.org/abs/2108.06314)): Temporal knowledge base questions derived from WikiData
33
+
34
+ ### Features
35
+
36
+ | Field | Type | Description |
37
+ |-------|------|-------------|
38
+ | `id` | string | Unique identifier |
39
+ | `question` | string | Full question with timestamp prefix (e.g., "Today is Monday, Jan 5, 2015. Who is...") |
40
+ | `context` | string | Evidence document containing the answer |
41
+ | `answers` | list[string] | List of acceptable answers |
42
+ | `source` | string | Original dataset (`streamingqa` or `timeqa`) |
43
+ | `question_timestamp` | string | When the question is asked (ISO format) |
44
+ | `evidence_timestamp` | string | When the evidence document was published |
45
+ | `dimensions` | string | Metadata about temporal dimensions (JSON) |
46
+ | `is_hard_negative` | bool | Whether this is a hard negative (temporal perturbation) |
47
+ | `original_question` | string | Question without timestamp prefix |
48
+
49
+ ### Splits
50
+
51
+ | Split | Examples | Description |
52
+ |-------|----------|-------------|
53
+ | train | 25,041 | Training set |
54
+ | validation | 2,800 | Validation set |
55
+ | test | 9,089 | Test/evaluation set |
56
+
57
+ ### Hard Negatives
58
+
59
+ Examples with `is_hard_negative=True` are temporal perturbations where the question timestamp
60
+ is shifted to make the context outdated or incorrect. These are useful for evaluating a model's
61
+ ability to detect temporal mismatches between query time and evidence time.
62
+
63
+ ## Usage
64
+
65
+ ```python
66
+ from datasets import load_dataset
67
+
68
+ # Load the dataset
69
+ dataset = load_dataset("Catkamakura/ts-qa")
70
+
71
+ # Access splits
72
+ train = dataset["train"]
73
+ validation = dataset["validation"]
74
+ test = dataset["test"]
75
+
76
+ # Example
77
+ example = train[0]
78
+ print(example["question"])
79
+ # "Today is Tuesday, September 24, 2013. Who was the Transport Minister of Thailand?"
80
+ print(example["answers"])
81
+ # ["Chadchart Sittipunt"]
82
+ print(example["source"])
83
+ # "streamingqa"
84
+
85
+ # Filter by source
86
+ streamingqa_examples = train.filter(lambda x: x["source"] == "streamingqa")
87
+ timeqa_examples = train.filter(lambda x: x["source"] == "timeqa")
88
+
89
+ # Filter hard negatives
90
+ hard_negatives = train.filter(lambda x: x["is_hard_negative"])
91
+ ```
92
+
93
+ ## Citation
94
+
95
+ If you use this dataset, please cite the original sources:
96
+
97
+ ```bibtex
98
+ @inproceedings{liska2022streamingqa,
99
+ title={StreamingQA: A Benchmark for Adaptation to New Knowledge over Time in Question Answering Models},
100
+ author={Liska, Adam and Kocisky, Tomas and Gribovskaya, Elena and Terber, Tayfun and Shutova, Ekaterina and Glaese, Amelia},
101
+ booktitle={International Conference on Machine Learning},
102
+ pages={13604--13622},
103
+ year={2022},
104
+ organization={PMLR}
105
+ }
106
+
107
+ @inproceedings{chen2021timeqa,
108
+ title={A Dataset for Answering Time-Sensitive Questions},
109
+ author={Chen, Wenhu and Wang, Xinyi and Wang, William Yang},
110
+ booktitle={NeurIPS 2021 Datasets and Benchmarks Track},
111
+ year={2021}
112
+ }
113
+ ```
114
+
115
+ ## License
116
+
117
+ MIT License