| | --- |
| | license: apache-2.0 |
| | tags: |
| | - Embeddings |
| | - ACAV100M |
| | - AE29H_float32 |
| | - nanowakeword |
| | - noice |
| | --- |
| | |
| | # AE29H_float32 |
| | |
| | **Audio Embeddings ~29 hours** dataset contains **precomputed audio embeddings** designed for **[Nanowakeword](https://github.com/arcosoph/nanowakeword)** framework. The embeddings are intended to be used as **general-purpose negative training data**, meaning the audio does **not contain the target wake word or phrase**. |
| | |
| | Unlike raw audio datasets, the files in this dataset contain **low-dimensional audio embeddings** extracted from audio clips using a pre-trained [speech embedding](https://www.kaggle.com/models/google/speech-embedding) model. These embeddings can be directly used as input features when training wake-word detection models with NanoWakeWord. |
| | |
| | The goal of this dataset is to provide **diverse background audio representations** (speech, environmental noise, music, etc.) that help wake-word models learn to avoid false activations. |
| | |
| | --- |
| | |
| | # Dataset Source |
| | |
| | The embeddings were generated from a subset of the **[ACAV100M](https://acav100m.github.io/)** dataset. |
| | |
| | ACAV100M is a very large automatically curated audio-visual dataset created from millions of internet videos and designed for large-scale audio-visual learning. It contains diverse real-world audio such as speech, environmental sounds, music, and background noise. |
| | |
| | For this dataset: |
| | |
| | * A **20K subset (~2 days of audio)** from ACAV100M. |
| | * Audio clips were processed and converted into embeddings suitable for wake-word training. |
| | |
| | --- |
| | |
| | # Dataset Statistics |
| | |
| | * **Shape:** |
| | |
| | ``` |
| | (21115, 16, 96) |
| | ``` |
| | |
| | * **Total samples:** |
| | **21,115** |
| | |
| | * **Feature dimensions:** |
| | |
| | * **Temporal steps:** 16 |
| | * **Embedding size:** 96 |
| | Each sample represents approximately **1.28 seconds of audio**, where each temporal step corresponds to **~80 ms**. |
| | |
| | --- |
| | |
| | # Data Type |
| | |
| | ``` |
| | dtype: float32 |
| | ``` |
| | |
| | **Value range** |
| | |
| | ``` |
| | min: -77.23914 |
| | max: 95.59355 |
| | ``` |
| | |
| | --- |
| | |
| | # Intended Use |
| | |
| | This dataset is intended for: |
| | |
| | * Training **NanoWakeWord wake-word detection models** |
| | * Providing **negative training examples** |
| | * Improving **false-positive robustness** |
| | * Training models that operate directly on **audio embeddings** |