Dataset Viewer
Auto-converted to Parquet Duplicate
Search is not available for this dataset
audio
audioduration (s)
4.64
136
End of preview. Expand in Data Studio

Speech-to-Intent Benchmark

License

Made in Vancouver, Canada by Picovoice

Twitter URL YouTube Channel Views

This framework benchmarks the accuracy of Picovoice's Speech-to-Intent engine, Rhino. It compares the accuracy of Rhino with:

Results

Command acceptance rate is the probability of an engine correctly understanding the spoken command. Below is the summary:

Command Acceptance Rate: Google Dialogflow - 77.6%, Amazon Lex - 84.5%, IBM Watson - 86.8%, Microsoft LUIS - 89.9%, Picovoice Rhino - 97.6%

The figure below depicts engines performance at each SNR:

Accuracy of NLU Engines

Data

The speech data are crowd-sourced from more than 50 unique speakers. Each speaker contributed about ten different utterances. Collectively there are 619 commands used in this benchmark. We test the engines in noisy conditions to simulate real-world situations. Noise is from Freesound.

How to Reproduce?

Clone the Picovoice/speech-to-intent-benchmark repository on GitHub:

git clone https://github.com/Picovoice/speech-to-intent-benchmark.git

Get the usage message:

python3 src/bench.py --help

Then run the script for each engine.

Downloads last month
31