Establishing Data Card
Browse files
README.md
CHANGED
|
@@ -1,3 +1,37 @@
|
|
| 1 |
-
---
|
| 2 |
-
license: mit
|
| 3 |
-
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: mit
|
| 3 |
+
---
|
| 4 |
+
**OLMoASR-Pool** is a web-scale audio-text dataset collected from the public internet, consisting of approximately 3M hours of audio and 17M transcripts.
|
| 5 |
+
|
| 6 |
+
With OLMoASR-Pool, we trained **OLMoASR**, a series of English speech recognition models and observed strong generalization and robust capabilities!
|
| 7 |
+
|
| 8 |
+
# Content
|
| 9 |
+
- The dataset contains 18,761,823 unique IDs spanning approximately 3.4M hours of audio.
|
| 10 |
+
- It also spans across a variety speaking styles, accents and audio setups such as news segments, podcasts, outdoors, crowds, speeches, commentary, interviews and more!
|
| 11 |
+
- **OLMoASR-Pool** is multilingual as it can contain non-English audio/transcripts. To retrieve an English-only dataset, it is critical to perform audio-text language alignment.
|
| 12 |
+
- After downloading the collection for training, only 3M hours of audio and 17M transcripts remains.
|
| 13 |
+
|
| 14 |
+
# Usage
|
| 15 |
+
1. Download from HuggingFace
|
| 16 |
+
- Retrieve HF access token from [here](https://huggingface.co/settings/tokens) to gain access to the dataset.
|
| 17 |
+
- Run `pip install huggingface_hub[cli]`
|
| 18 |
+
- Run `huggingface-cli login` in your CLI and paste the HF access token to login
|
| 19 |
+
- Use the code below to access the IDs
|
| 20 |
+
```
|
| 21 |
+
from datasets import load_dataset
|
| 22 |
+
dataset = load_dataset("allenai/OLMoASR-Pool", streaming=True)
|
| 23 |
+
print(dataset) # features: ['id']
|
| 24 |
+
print(next(iter(dataset['train'])))
|
| 25 |
+
```
|
| 26 |
+
- If you're downloading all the IDs, you can run the code below
|
| 27 |
+
```
|
| 28 |
+
from datasets import load_dataset
|
| 29 |
+
dataset = load_dataset("allenai/OLMoASR-Pool", streaming=False, cache_dir=<where you want to download the IDs to>)
|
| 30 |
+
```
|
| 31 |
+
2. Download the audio and transcript files from ID information.
|
| 32 |
+
4. Preprocess the audio and transcript files. Follow the instructions at the [OLMoASR repo](https://github.com/allenai/OLMoASR_newest)
|
| 33 |
+
|
| 34 |
+
|
| 35 |
+
# Uses
|
| 36 |
+
The collection was used to train a speech recognition model, but it can also be used in research areas such as conversational data, audio understanding, speaker diarization, voice detection and more.
|
| 37 |
+
|