Dataset in parquet format, and README.MD

#1
Files changed (2) hide show
  1. FinRAG.parquet +3 -0
  2. README.md +119 -3
FinRAG.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e6106354d99c4ed109b886281ad5a838d011f184340fcd156cca1f26e3dd939
3
+ size 20332214
README.md CHANGED
@@ -1,3 +1,119 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ task_categories:
4
+ - question-answering
5
+ - multiple-choice
6
+ language:
7
+ - en
8
+ tags:
9
+ - finance
10
+ - numerical-reasoning
11
+ - table-qa
12
+ - financial-analysis
13
+ size_categories:
14
+ - 10K<n<100K
15
+ ---
16
+
17
+ # Financial Reasoning Dataset with Distractors
18
+
19
+ ## Dataset Description
20
+
21
+ This dataset contains 12,500 financial reasoning questions based on real-world financial documents, earnings reports, and financial tables. Each question is accompanied by a correct answer and four carefully crafted distractor answers, making it suitable for multiple-choice question answering tasks and assessing financial numerical reasoning capabilities.
22
+
23
+ ### Dataset Summary
24
+
25
+ - **Total Examples**: 12,500
26
+ - **Format**: Multiple-choice questions with 5 options (1 correct + 4 distractors)
27
+ - **Domain**: Financial documents, earnings reports, financial tables
28
+ - **Task**: Numerical reasoning over financial text and tables
29
+ - **Language**: English
30
+
31
+ ### Source Data
32
+
33
+ This dataset combines all splits (train, validation, and test) from:
34
+ - **FinQA**: Financial Question Answering dataset (7,750 questions, 62%)
35
+ - **TAT-QA**: Table-and-Text Question Answering dataset (4,750 questions, 38%)
36
+
37
+ ### Distractor Generation
38
+
39
+ Four distractor answers were algorithmically generated for each question using the following techniques:
40
+ - **Stop early**: Stopping calculation before completion
41
+ - **Negate operand**: Negating numbers in the calculation
42
+ - **Operand bleeding**: Using the wrong operands from the table
43
+ - **Replace operator**: Using the wrong mathematical operation (e.g., multiply instead of add)
44
+ - **Switch order**: Changing the order of operations
45
+ - **Percentage error**: Mistakes in percentage conversion
46
+ - **Unit error**: Mistakes in unit conversion (e.g., millions vs. thousands)
47
+ - **Append operation**: Adding extra unnecessary operations
48
+ - **Substitution error**: Substituting incorrect values from the table
49
+
50
+ These techniques create plausible but incorrect answers that test true understanding of the financial reasoning task.
51
+ ## Dataset Structure
52
+
53
+ ### Data Fields
54
+
55
+ Each example in the dataset contains:
56
+
57
+ - **`id`** (string): Unique identifier for each question
58
+ - **`pre_text`** (list of strings): Contextual text passages from the financial document that appear before the table
59
+ - **`post_text`** (list of strings): Additional contextual text passages that appear after the table (may be empty)
60
+ - **`table`** (list of lists): Financial table data in row-major format, where the first row typically contains headers
61
+ - **`question`** (string): The financial reasoning question to be answered
62
+ - **`choices`** (list of strings): List of 5 answer choices (1 correct + 4 distractors), randomly shuffled
63
+ - **`answer`** (integer): Index (0-4) pointing to the correct choice in the `choices` list
64
+ - **`metadata`** (dict): Additional information including:
65
+ - `instructions`: General instructions for the question type
66
+ - `date_created`: Date the entry was created
67
+ - `identifier`: Numeric identifier
68
+
69
+ ### Data Example
70
+
71
+ ```json
72
+ {
73
+ "id": "9bbb9fb3-3482-4d4d-be40-dd6ff47c23e9",
74
+ "pre_text": [
75
+ "Orders at Mobility grew to a record high on a sharp increase in volume...",
76
+ "Revenue grew slightly as double-digit growth in the customer services business..."
77
+ ],
78
+ "post_text": [],
79
+ "table": [
80
+ ["", "", "Fiscal year", "", "% Change"],
81
+ ["(in millions of €)", "2019", "2018", "Actual", "Comp."],
82
+ ["Orders", "12,894", "11,025", "17 %", "16 %"],
83
+ ["Revenue", "8,916", "8,821", "1 %", "0 %"]
84
+ ],
85
+ "question": "Analyse this data from a financial earnings document. What it the increase / (decrease) in revenue from 2018 to 2019?",
86
+ "choices": ["-3978", "17737", "94", "95", "1"],
87
+ "answer": 3,
88
+ "metadata": {
89
+ "instructions": "Analyse this data from a financial earnings document.",
90
+ "date_created": "2024-07-16",
91
+ "identifier": 1000
92
+ }
93
+ }
94
+ ```
95
+
96
+ ## Citation
97
+
98
+ **FinQA:**
99
+ ```bibtex
100
+ @inproceedings{chen-etal-2021-finqa,
101
+ title = "{F}in{QA}: A Dataset of Numerical Reasoning over Financial Data",
102
+ author = "Chen, Zhiyu and Chen, Wenhu and Smiley, Charese and Shah, Sameena and
103
+ Borova, Iana and Langdon, Dylan and Moussa, Reema and Beane, Matt and
104
+ Huang, Ting-Hao and Routledge, Bryan and Wang, William Yang",
105
+ booktitle = "Proceedings of EMNLP 2021",
106
+ year = "2021"
107
+ }
108
+ ```
109
+
110
+ **TAT-QA:**
111
+ ```bibtex
112
+ @inproceedings{zhu-etal-2021-tat,
113
+ title = "{TAT}-{QA}: A Question Answering Benchmark on a Hybrid of Tabular and Textual Content in Finance",
114
+ author = "Zhu, Fengbin and Lei, Wenqiang and Wang, Chao and Zheng, Jianming and
115
+ Poria, Soujanya and Chua, Tat-Seng",
116
+ booktitle = "Proceedings of ACL-IJCNLP 2021",
117
+ year = "2021"
118
+ }
119
+ ```