Do What I Say: A Spoken Prompt Dataset for Instruction-Following
Abstract
Speech Large Language Models are evaluated using a new multilingual dataset that pairs spoken and written prompts across multiple tasks and languages, revealing that text prompts generally outperform spoken prompts except in speech-output tasks.
Speech Large Language Models (SLLMs) have rapidly expanded, supporting a wide range of tasks. These models are typically evaluated using text prompts, which may not reflect real-world scenarios where users interact with speech. To address this gap, we introduce DoWhatISay (DOWIS), a multilingual dataset of human-recorded spoken and written prompts designed to pair with any existing benchmark for realistic evaluation of SLLMs under spoken instruction conditions. Spanning 9 tasks and 11 languages, it provides 10 prompt variants per task-language pair, across five styles. Using DOWIS, we benchmark state-of-the-art SLLMs, analyzing the interplay between prompt modality, style, language, and task type. Results show that text prompts consistently outperform spoken prompts, particularly for low-resource and cross-lingual settings. Only for tasks with speech output, spoken prompts do close the gap, highlighting the need for speech-based prompting in SLLM evaluation.
Community
DOWIS is a multilingual dataset of human-recorded spoken and written instruction prompts, designed to enable realistic evaluation of Speech Large Language Models across 11 tasks and 12 languages.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Language-Aware Distillation for Multilingual Instruction-Following Speech LLMs with ASR-Only Supervision (2026)
- Timbre-Aware LLM-based Direct Speech-to-Speech Translation Extendable to Multiple Language Pairs (2026)
- LongSpeech: A Scalable Benchmark for Transcription, Translation and Understanding in Long Speech (2026)
- Lost in Transcription: How Speech-to-Text Errors Derail Code Understanding (2026)
- Covo-Audio Technical Report (2026)
- StyleBench: Evaluating Speech Language Models on Conversational Speaking Style Control (2026)
- A Benchmark for Audio Reasoning Capabilities of Multimodal Large Language Models (2026)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 2
Spaces citing this paper 0
No Space linking this paper