fkapsahili commited on
Commit
44529cd
·
verified ·
1 Parent(s): 342f9e6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +146 -36
README.md CHANGED
@@ -1,36 +1,146 @@
1
- ---
2
- license: cc-by-4.0
3
- configs:
4
- - config_name: default
5
- data_files:
6
- - split: train
7
- path: data/train-*
8
- dataset_info:
9
- features:
10
- - name: id
11
- dtype: string
12
- - name: domain
13
- dtype: string
14
- - name: question_type
15
- dtype: string
16
- - name: dynamism
17
- dtype: string
18
- - name: question
19
- dtype: string
20
- - name: reference_answer
21
- dtype: string
22
- - name: sources
23
- list:
24
- - name: filename
25
- dtype: string
26
- - name: id
27
- dtype: string
28
- - name: pages
29
- sequence: int64
30
- splits:
31
- - name: train
32
- num_bytes: 35785
33
- num_examples: 100
34
- download_size: 21165
35
- dataset_size: 35785
36
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-4.0
3
+ configs:
4
+ - config_name: default
5
+ data_files:
6
+ - split: train
7
+ path: data/train-*
8
+ dataset_info:
9
+ features:
10
+ - name: id
11
+ dtype: string
12
+ - name: domain
13
+ dtype: string
14
+ - name: question_type
15
+ dtype: string
16
+ - name: dynamism
17
+ dtype: string
18
+ - name: question
19
+ dtype: string
20
+ - name: reference_answer
21
+ dtype: string
22
+ - name: sources
23
+ list:
24
+ - name: filename
25
+ dtype: string
26
+ - name: id
27
+ dtype: string
28
+ - name: pages
29
+ sequence: int64
30
+ splits:
31
+ - name: train
32
+ num_bytes: 35785
33
+ num_examples: 100
34
+ download_size: 21165
35
+ dataset_size: 35785
36
+ ---
37
+
38
+
39
+ # EntRAG Benchmark: Question Answering Dataset
40
+
41
+ ## Description
42
+
43
+ EntRAG is a specialized benchmark dataset designed for evaluating Retrieval-Augmented Generation (RAG) systems in enterprise contexts.
44
+ The dataset addresses the unique challenges of business environments where information comes from heterogeneous sources including structured databases, documents, and dynamic mock APIs.
45
+
46
+ The dataset comprises 100 manually constructed question-answer pairs across six enterprise domains: Finance, Technical Documentation, Environment, Legal and Compliance, Human Resources, and Marketing and Sales.
47
+ Questions are designed to evaluate both static document retrieval and dynamic API integration scenarios, reflecting realistic enterprise information needs.
48
+
49
+ ## Dataset Structure
50
+
51
+ ### Columns
52
+
53
+ * `id`: Unique identifier for each question-answer pair
54
+ * `domain`: The subject area or field of knowledge the question pertains to (e.g., "Technical Documentation", "Finance", "Healthcare")
55
+ * `question_type`: The category of reasoning required (e.g., "comparison", "factual", "analytical", "procedural")
56
+ * `dynamism`: Indicates whether the answer content changes over time ("static" for timeless information, "dynamic" for evolving content)
57
+ * `question`: A natural language question that requires information retrieval and reasoning to answer accurately
58
+ * `reference_answer`: The correct, comprehensive answer that serves as the ground truth for evaluation
59
+ * `sources`: Array of source documents that contain the information needed to answer the question, including:
60
+ * `id`: Unique identifier for the source
61
+ * `filename`: Name of the source document or API endpoint
62
+ * `pages`: Array of specific page numbers where relevant information is found (empty for API sources)
63
+
64
+ ## Use Cases
65
+
66
+ This dataset is particularly valuable for:
67
+
68
+ * **RAG System Evaluation**: Testing RAG systems with realistic business scenarios and multi-source information integration
69
+ * **Hybrid System Assessment**: Evaluating systems that combine document retrieval with API-based data access
70
+ * **Domain-Specific Analysis**: Understanding RAG performance across different business domains
71
+ * **Dynamic Information Handling**: Assessing systems that work with both static documents and real-time data sources
72
+
73
+ ## Accessing the Dataset
74
+
75
+ You can load this dataset via the Hugging Face Datasets library using the following Python code:
76
+
77
+ ```python
78
+ from datasets import load_dataset
79
+
80
+ # Load the dataset
81
+ dataset = load_dataset("fkapsahili/EntRAG")
82
+
83
+ # Access the data
84
+ for example in dataset['train']:
85
+ print(f"Domain: {example['domain']}")
86
+ print(f"Question Type: {example['question_type']}")
87
+ print(f"Dynamism: {example['dynamism']}")
88
+ print(f"Question: {example['question']}")
89
+ print(f"Answer: {example['reference_answer']}")
90
+ print(f"Sources: {len(example['sources'])} documents")
91
+ print("---")
92
+ ```
93
+
94
+ ### Alternative Loading Methods
95
+
96
+ For direct integration with evaluation frameworks:
97
+
98
+ ```python
99
+ import json
100
+ from datasets import load_dataset
101
+
102
+ # Load and convert to list format
103
+ dataset = load_dataset("fkapsahili/EntRAG", split="train")
104
+ qa_pairs = [dict(item) for item in dataset]
105
+ ```
106
+
107
+ ## Integration with RAG Frameworks
108
+
109
+ This dataset supports evaluation of various RAG architectures and can be integrated with existing evaluation pipelines.
110
+ The format is compatible with standard RAG evaluation frameworks and supports both document-based and API-integrated systems.
111
+
112
+ ## Dataset Statistics
113
+
114
+ * **Total QA Pairs**: 100 manually constructed questions
115
+ * **Domains**: 6 domains (Finance, Technical Documentation, Environment, Legal and Compliance, Human Resources, Marketing and Sales)
116
+ * **Question Types**: 7 reasoning patterns (simple queries, comparison, aggregation, multi-hop reasoning, simple with conditions, factual contradiction, post-processing)
117
+ * **Dynamism Distribution**:
118
+ * Static questions: 28% (document-based retrieval)
119
+ * Dynamic questions: 72% (requiring real-time API integration)
120
+ * **Source Documents**: 11,000+ pages from authentic enterprise documents across 10 major companies
121
+ * **Company Sectors**: Technology, healthcare, e-commerce, retail, automotive, and energy
122
+ * **Mock APIs**: 4 domain-specific APIs (finance, SEC filings, HR statistics, web search)
123
+
124
+ ## Citation
125
+
126
+ If you use this dataset in your research, please cite:
127
+
128
+ ```bibtex
129
+ @dataset{entrag_2025,
130
+ title={EntRAG: Enterprise RAG Benchmark},
131
+ author={Fabio Kapsahili},
132
+ year={2025},
133
+ publisher={Hugging Face},
134
+ url={https://huggingface.co/datasets/fkapsahili/EntRAG}
135
+ }
136
+ ```
137
+
138
+ ## License
139
+
140
+ This dataset is released under Creative Commons Attribution 4.0. Please see the LICENSE file for full details.
141
+
142
+ ## Additional Resources
143
+
144
+ * **Evaluation Code**: https://github.com/fkapsahili/EntRAG
145
+
146
+ For questions, issues, please open an issue in the associated GitHub repository.