--- license: cc-by-4.0 language: - de - en - es - cs - fr - hu - it - nl - pt - ru - sq - sv tags: - speech prompts - text prompts - instruction following - benchmark size_categories: - 1KNEW DOWIS now also contains spoken and written prompts in Albanian (sq), and for the tasks LIPREAD and SLU! > **TL;DR** — DOWIS is a multilingual dataset of human-recorded spoken and written instruction prompts, designed to enable realistic evaluation of Speech Large Language Models across 11 tasks and 12 languages. --- ## Dataset Summary Most Speech LLM benchmarks use text-based prompts, which does not reflect how users actually interact with these models in the real world. DOWIS fills this gap by providing human-recorded spoken prompts, paired with their written equivalents, across a wide range of tasks, languages, and prompt styles. Each prompt can be directly paired with any existing speech benchmark to evaluate how well Speech LLMs follow spoken instructions. The dataset contains **1,320 rows**, with up to 4 audio recordings per row (2 female, 2 male speakers where available), covering: - **12 languages**: cs, de, en, es, fr, hu, it, nl, pt, ru, sq, sv - **11 tasks**: ACHAP, ASR, MT, S2ST, SQA, SSUM, ST, TSUM, TTS, LIPREAD, SLU - **5 prompt styles**: basic, formal, informal, detailed, short - **10 prompt variants** per task-language pair Details can be found in the corresponding paper on [arXiv](https://arxiv.org/abs/2603.09881). Code for benchmarking Speech LLMs with different task benchmarks coupled with DOWIS can be found on [GitHub](https://github.com/MaikeZuefle/DOWIS/tree/main). --- ## Tasks | Task Code | Description | |-----------|-------------| | ACHAP | Audio Chaptering | | ASR | Automatic Speech Recognition | | MT | Machine Translation | | S2ST | Speech-to-Speech Translation | | SQA | Spoken Question Answering | | SSUM | Speech Summarization | | ST | Speech Translation | | TSUM | Text Summarization | | TTS | Text-to-Speech | | LIPREAD | Lip-Reading | | SLU | Spoken Language Understanding | ## Prompt Styles | Style | Description | |-------|-------------| | `basic` | Natural, everyday phrasing a researcher would use | | `formal` | Professional, polished language | | `informal` | Conversational and casual | | `detailed` | Explicit and precise instructions on how to perform the task | | `short` | Concise as possible while remaining unambiguous | --- ## Dataset Fields | Field | Type | Description | |-------|------|-------------| | `text_prompt` | `string` | Written version of the instruction prompt | | `audio_prompt_female_1` | `Audio` | Human-recorded female speaker (speaker 1), `null` if unavailable | | `audio_prompt_female_2` | `Audio` | Human-recorded female speaker (speaker 2), `null` if unavailable | | `audio_prompt_male_1` | `Audio` | Human-recorded male speaker (speaker 1), `null` if unavailable | | `audio_prompt_male_2` | `Audio` | Human-recorded male speaker (speaker 2), `null` if unavailable | | `language` | `string` | ISO 639-1 language code (e.g. `en`, `de`) | | `task` | `string` | Task code the prompt is designed for (e.g. `asr`, `mt`) | | `prompt_type` | `string` | Prompt style: `basic`, `formal`, `informal`, `detailed`, or `short` | --- ## Citation If you use this work, please cite: ```bibtex @misc{züfle2026isayspokenprompt, title={Do What I Say: A Spoken Prompt Dataset for Instruction-Following}, author={Maike Züfle and Sara Papi and Fabian Retkowski and Szymon Mazurek and Marek Kasztelnik and Alexander Waibel and Luisa Bentivogli and Jan Niehues}, year={2026}, eprint={2603.09881}, archivePrefix={arXiv}, primaryClass={cs.CL}, url={https://arxiv.org/abs/2603.09881}} ``` --- Dataset Contact: maike.zuefle@kit.edu