assistant-bench / README.md
arcada-labs's picture
Upload README.md with huggingface_hub
7d8464c verified
---
language:
- en
license: mit
pretty_name: Assistant Bench
tags:
- audio
- benchmark
- speech-to-speech
- voice-ai
- multi-turn
- tool-use
- evaluation
- state-tracking
- function-calling
task_categories:
- automatic-speech-recognition
- text-generation
size_categories:
- n<1K
configs:
- config_name: default
data_files:
- split: train
path: metadata.jsonl
---
# Assistant Bench
**31-turn multi-turn speech-to-speech benchmark** for evaluating voice AI models as a personal assistant handling flights, email, calendar, and reminders.
Part of [Audio Arena](https://audioarena.ai), a suite of 6 benchmarks spanning 221 turns across different domains. Built by [Arcada Labs](https://arcada.dev).
[Leaderboard](https://audioarena.ai/leaderboard) | [GitHub](https://github.com/Design-Arena/audio-arena-bench) | [All Benchmarks](#part-of-audio-arena)
## Dataset Description
The model acts as a personal assistant managing flight bookings, email composition, calendar events, and reminders. Turns include dual requests packed into a single utterance, mid-conversation topic switching, late references back to early topics, and correction chains that the model must track across the full session.
## What This Benchmark Tests
- **Tool use**: 7 functions — flight booking, email, calendar, reminders, and more
- **Dual requests in single turns**: Two distinct tasks in one spoken utterance
- **Topic switching**: Abrupt mid-conversation jumps between domains (flights, email, calendar)
- **Late references to early topics**: Callbacks to details from 20+ turns earlier
- **Intent segmentation**: Parsing multi-intent speech into separate tool calls
- **Mid-sentence self-correction**: Speaker changes their mind partway through a request
- **Retroactive email correction**: Changing details of a previously composed email
- **Correction-chain recall**: Tracking a value through multiple sequential corrections
- **False memory traps**: 3 false memory traps plus a correction-chain trap
- **Audio traps**: Name spelling, airport codes, dates, and times that are ambiguous in speech
## Dataset Structure
```
assistant-bench/
├── audio/ # TTS-generated audio (1 WAV per turn)
│ ├── turn_000.wav
│ ├── turn_001.wav
│ └── ... (31 files)
├── real_audio/ # Human-recorded audio
│ ├── person1/
│ │ └── turn_000.wav ... turn_030.wav
│ └── person2/
│ └── turn_000.wav ... turn_030.wav
├── benchmark/
│ ├── turns.json # Turn definitions with golden answers
│ ├── hard_turns.json # Same as turns.json but input_text=null (audio-only)
│ ├── tool_schemas.json # Tool/function schemas (7 tools)
│ └── knowledge_base.txt # Assistant KB
└── metadata.jsonl # HF dataset viewer metadata
```
### Metadata Fields
| Field | Description |
|-------|-------------|
| `file_name` | Path to the audio file |
| `turn_id` | Turn index (0–30) |
| `speaker` | `tts`, `person1`, or `person2` |
| `input_text` | What the user says (text transcript) |
| `golden_text` | Expected assistant response |
| `required_function_call` | Tool call the model should make (JSON, nullable) |
| `function_call_response` | Scripted tool response (JSON, nullable) |
| `categories` | Evaluation categories for this turn |
| `subcategory` | Specific sub-skill being tested |
| `scoring_dimensions` | Which judge dimensions apply |
## Audio Format
- **Format**: WAV, 16-bit PCM, mono
- **TTS audio**: Generated via text-to-speech
- **Real audio**: Human-recorded by multiple speakers, same transcript content
## Usage
### With Audio Arena CLI
```bash
pip install audio-arena # or: git clone + uv sync
# Run with a text model
uv run audio-arena run assistant_bench --model claude-sonnet-4-5 --service anthropic
# Run with a speech-to-speech model
uv run audio-arena run assistant_bench --model gpt-realtime --service openai-realtime
# Judge the results
uv run audio-arena judge runs/assistant_bench/<run_dir>
```
### With Hugging Face Datasets
```python
from datasets import load_dataset
ds = load_dataset("arcada-labs/assistant-bench")
```
## Evaluation
Models are judged on up to 5 dimensions per turn:
| Dimension | Description |
|-----------|-------------|
| `tool_use_correct` | Correct function called with correct arguments |
| `instruction_following` | User's request was actually completed |
| `kb_grounding` | Claims are supported by the knowledge base or tool results |
| `state_tracking` | Consistency with earlier turns (scored on tagged turns only) |
| `ambiguity_handling` | Correct disambiguation (scored on tagged turns only) |
For speech-to-speech models, a 6th `turn_taking` dimension evaluates audio timing correctness.
See the [full methodology](https://github.com/Design-Arena/audio-arena-bench#methodology) for details on two-phase evaluation, penalty absorption, and category-aware scoring.
## Part of Audio Arena
| Benchmark | Turns | Scenario |
|-----------|-------|----------|
| [Conversation Bench](https://huggingface.co/datasets/arcada-labs/conversation-bench) | 75 | Conference assistant |
| [Appointment Bench](https://huggingface.co/datasets/arcada-labs/appointment-bench) | 25 | Dental office scheduling |
| **Assistant Bench** (this dataset) | 31 | Personal assistant |
| [Event Bench](https://huggingface.co/datasets/arcada-labs/event-bench) | 29 | Event planning |
| [Grocery Bench](https://huggingface.co/datasets/arcada-labs/grocery-bench) | 30 | Grocery ordering |
| [Product Bench](https://huggingface.co/datasets/arcada-labs/product-bench) | 31 | Laptop comparison shopping |
## Citation
```bibtex
@misc{audioarena2026,
title={Audio Arena: Multi-Turn Speech-to-Speech Evaluation Benchmarks},
author={Arcada Labs},
year={2026},
url={https://audioarena.ai}
}
```