EntRAG / README.md
fkapsahili's picture
Update README.md
548523d verified
---
license: cc-by-4.0
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
dataset_info:
features:
- name: id
dtype: string
- name: domain
dtype: string
- name: question_type
dtype: string
- name: dynamism
dtype: string
- name: question
dtype: string
- name: reference_answer
dtype: string
- name: sources
list:
- name: filename
dtype: string
- name: id
dtype: string
- name: pages
sequence: int64
splits:
- name: train
num_bytes: 35785
num_examples: 100
download_size: 21165
dataset_size: 35785
---
# EntRAG Benchmark: Question Answering Dataset
## Description
EntRAG is a specialized benchmark dataset designed for evaluating Retrieval-Augmented Generation (RAG) systems in enterprise contexts.
The dataset addresses the unique challenges of business environments where information comes from heterogeneous sources including structured databases, documents, and dynamic mock APIs.
The dataset comprises 100 manually constructed question-answer pairs across six enterprise domains: Finance, Technical Documentation, Environment, Legal and Compliance, Human Resources, and Marketing and Sales.
Questions are designed to evaluate both static document retrieval and dynamic API integration scenarios, reflecting realistic enterprise information needs.
## Dataset Structure
### Columns
* `id`: Unique identifier for each question-answer pair
* `domain`: The subject area or field of knowledge the question pertains to (e.g., "Technical Documentation", "Finance", "Healthcare")
* `question_type`: The category of reasoning required (e.g., "comparison", "factual", "analytical", "procedural")
* `dynamism`: Indicates whether the answer content changes over time ("static" for timeless information, "dynamic" for evolving content)
* `question`: A natural language question that requires information retrieval and reasoning to answer accurately
* `reference_answer`: The correct, comprehensive answer that serves as the ground truth for evaluation
* `sources`: Array of source documents that contain the information needed to answer the question, including:
* `id`: Unique identifier for the source
* `filename`: Name of the source document or API endpoint
* `pages`: Array of specific page numbers where relevant information is found (empty for API sources)
## Use Cases
This dataset is particularly valuable for:
* **RAG System Evaluation**: Testing RAG systems with realistic business scenarios and multi-source information integration
* **Hybrid System Assessment**: Evaluating systems that combine document retrieval with API-based data access
* **Domain-Specific Analysis**: Understanding RAG performance across different business domains
* **Dynamic Information Handling**: Assessing systems that work with both static documents and real-time data sources
## Accessing the Dataset
You can load this dataset via the Hugging Face Datasets library using the following Python code:
```python
from datasets import load_dataset
# Load the dataset
dataset = load_dataset("fkapsahili/EntRAG")
# Access the data
for example in dataset['train']:
print(f"Domain: {example['domain']}")
print(f"Question Type: {example['question_type']}")
print(f"Dynamism: {example['dynamism']}")
print(f"Question: {example['question']}")
print(f"Answer: {example['reference_answer']}")
print(f"Sources: {len(example['sources'])} documents")
print("---")
```
### Alternative Loading Methods
For direct integration with evaluation frameworks:
```python
import json
from datasets import load_dataset
# Load and convert to list format
dataset = load_dataset("fkapsahili/EntRAG", split="train")
qa_pairs = [dict(item) for item in dataset]
```
## Integration with RAG Frameworks
This dataset supports evaluation of various RAG architectures and can be integrated with existing evaluation pipelines.
The format is compatible with standard RAG evaluation frameworks and supports both document-based and API-integrated systems.
## Dataset Statistics
* **Total QA Pairs**: 100 manually constructed questions
* **Domains**: 6 domains (Finance, Technical Documentation, Environment, Legal and Compliance, Human Resources, Marketing and Sales)
* **Question Types**: 7 reasoning patterns (simple queries, comparison, aggregation, multi-hop reasoning, simple with conditions, factual contradiction, post-processing)
* **Dynamism Distribution**:
* Static questions: 28% (document-based retrieval)
* Dynamic questions: 72% (requiring real-time API integration)
* **Source Documents**: 9,500+ pages from authentic enterprise documents across 10 major companies
* **Company Sectors**: Technology, healthcare, e-commerce, retail, automotive, and energy
* **Mock APIs**: 4 domain-specific APIs (finance, SEC filings, HR statistics, web search)
## Citation
If you use this dataset in your research, please cite:
```bibtex
@dataset{entrag_2025,
title={EntRAG: Enterprise RAG Benchmark},
author={Fabio Kapsahili},
year={2025},
publisher={Hugging Face},
url={https://huggingface.co/datasets/fkapsahili/EntRAG}
}
```
## License
This dataset is released under Creative Commons Attribution 4.0. Please see the LICENSE file for full details.
## Additional Resources
* **Evaluation Code**: https://github.com/fkapsahili/EntRAG
For questions, issues, please open an issue in the associated GitHub repository.