Raniahossam33's picture
Upload dataset
f7777ba verified
---
language:
- ar
license: apache-2.0
task_categories:
- multiple-choice
- question-answering
pretty_name: Arabic Accounting MCQ Evaluation Dataset
tags:
- accounting
- mcq
- arabic
- evaluation
- benchmark
- education
dataset_info:
features:
- name: id
dtype: string
- name: query
dtype: string
- name: answer
dtype: string
- name: text
dtype: string
- name: choices
list: string
- name: gold
dtype: int64
splits:
- name: test
num_bytes: 379104
num_examples: 167
download_size: 113685
dataset_size: 379104
configs:
- config_name: default
data_files:
- split: test
path: data/test-*
---
# Arabic Accounting MCQ Evaluation Dataset
Validation and test splits for Arabic accounting MCQ with English letter choices.
## Dataset Structure
- **Format**: Multiple choice questions (4 options)
- **Language**: Arabic questions with English letter choices
- **Domain**: Accounting and finance
- **Validation**: 10% of total dataset
- **Test**: 10% of total dataset
## Fields
- `id`: Unique identifier
- `query`: Full MCQ prompt
- `answer`: Correct answer letter (a, b, c, d)
- `text`: Question text
- `choices`: Options ['a', 'b', 'c', 'd']
- `gold`: Correct answer index (0-3)
## Answer Mapping
- a → gold: 0
- b → gold: 1
- c → gold: 2
- d → gold: 3
## Usage
```python
from datasets import load_dataset
dataset = load_dataset("SahmBenchmark/arabic-accounting-mcq_eval")
# Access splits
val_data = dataset['validation']
test_data = dataset['test']
# Evaluation
correct = 0
for example in test_data:
model_answer = model.generate(example['query'])
if model_answer == example['answer']:
correct += 1
accuracy = correct / len(test_data)
```
For training data, see: `SahmBenchmark/arabic-accounting-mcq_train`