Adde comment on deduplication
Browse files
README.md
CHANGED
|
@@ -30,13 +30,12 @@ size_categories:
|
|
| 30 |
|
| 31 |
## Dataset Summary
|
| 32 |
|
| 33 |
-
**Utterly** is a speech dataset derived from the *human_5_all* and *smart-turn-data-v3.1-train* by Pipecat-AI. It contains over **
|
| 34 |
|
| 35 |
* Verbatim Whisper-generated transcripts
|
| 36 |
* End-of-turn (EoT) markers
|
| 37 |
* Speaker identifiers (Coming soon)
|
| 38 |
|
| 39 |
-
|
| 40 |
The dataset is designed to support research and development of speech and dialogue systems that require joint modeling of **speech recognition** and **conversational turn-taking**, such as streaming ASR systems and real-time conversational agents.
|
| 41 |
|
| 42 |
---
|
|
@@ -47,10 +46,10 @@ The dataset is designed to support research and development of speech and dialog
|
|
| 47 |
* **Language(s)**: English
|
| 48 |
* **Modality**: Audio (speech; mono-channel; sampled at 16kHz), Text
|
| 49 |
* **Interaction type**: Human conversational speech
|
| 50 |
-
* **Utterances**:
|
| 51 |
* **Speakers**: >100
|
| 52 |
|
| 53 |
-
Dataset splits (e.g., train/validation/test) are not predefined and may be created by downstream users as needed.
|
| 54 |
|
| 55 |
Note that Utterly is a *derived dataset*. All audio originates from the base datasets, with additional annotations created by the dataset author.
|
| 56 |
|
|
|
|
| 30 |
|
| 31 |
## Dataset Summary
|
| 32 |
|
| 33 |
+
**Utterly** is a speech dataset derived from the *human_5_all* and *smart-turn-data-v3.1-train* by Pipecat-AI. It contains over **5.8k English utterances** by a variety of speakers, and augments each example with turn-level annotations, including:
|
| 34 |
|
| 35 |
* Verbatim Whisper-generated transcripts
|
| 36 |
* End-of-turn (EoT) markers
|
| 37 |
* Speaker identifiers (Coming soon)
|
| 38 |
|
|
|
|
| 39 |
The dataset is designed to support research and development of speech and dialogue systems that require joint modeling of **speech recognition** and **conversational turn-taking**, such as streaming ASR systems and real-time conversational agents.
|
| 40 |
|
| 41 |
---
|
|
|
|
| 46 |
* **Language(s)**: English
|
| 47 |
* **Modality**: Audio (speech; mono-channel; sampled at 16kHz), Text
|
| 48 |
* **Interaction type**: Human conversational speech
|
| 49 |
+
* **Utterances**: 5,860
|
| 50 |
* **Speakers**: >100
|
| 51 |
|
| 52 |
+
Dataset splits (e.g., train/validation/test) are not predefined and may be created by downstream users as needed. Care was taken to ensure examples are unique through deduplication of the underlying audio examples.
|
| 53 |
|
| 54 |
Note that Utterly is a *derived dataset*. All audio originates from the base datasets, with additional annotations created by the dataset author.
|
| 55 |
|