Create README.md
Browse files
README.md
ADDED
|
@@ -0,0 +1,69 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# RAVE Models by Tangible Music Lab
|
| 2 |
+
|
| 3 |
+
This is a collection of RAVE (Realtime Audio Variational autoEncoder) models trained by the Tangible Music Lab for audio generation and transformation. The aim of this repository is to provide musicians with pre-trained models for building embedded RAVE models on the Raspberry Pi platform or similar, for physical hardware and tangible interface development for sound and music experimentation. These models enable real-time audio manipulation and generation while being optimized for resource-constrained environments, making them ideal for interactive musical instruments and sound installations.
|
| 4 |
+
|
| 5 |
+
## Model Details
|
| 6 |
+
|
| 7 |
+
### Model Description
|
| 8 |
+
|
| 9 |
+
- **Developed by:** Tangible Music Lab
|
| 10 |
+
- **Model type:** RAVE (Realtime Audio Variational autoEncoder)
|
| 11 |
+
- **License:** MIT
|
| 12 |
+
|
| 13 |
+
### Model Sources
|
| 14 |
+
|
| 15 |
+
- **Repository:** https://huggingface.co/TangibleMusicLab/rave-models
|
| 16 |
+
- **Training Code:** https://github.com/victor-shepardson/RAVE
|
| 17 |
+
|
| 18 |
+
## Uses
|
| 19 |
+
|
| 20 |
+
### Direct Use
|
| 21 |
+
|
| 22 |
+
These models are designed for real-time audio generation and transformation. They can be used with:
|
| 23 |
+
- nn~
|
| 24 |
+
- NN.ar
|
| 25 |
+
- rave-supercollider
|
| 26 |
+
|
| 27 |
+
### Models
|
| 28 |
+
|
| 29 |
+
#### tam_freesoundloop10k_default_z16.ts
|
| 30 |
+
- Dataset: Freesound Loop Dataset
|
| 31 |
+
- Model: RAVE v3 with default configuration
|
| 32 |
+
- Latent dimensions: 16
|
| 33 |
+
- Sample rate: 48kHz
|
| 34 |
+
|
| 35 |
+
#### tam_freesoundloop10k_raspi_z16.ts
|
| 36 |
+
- Dataset: Freesound Loop Dataset
|
| 37 |
+
- Model: Modified RAVE v3, optimized for Raspberry Pi 5
|
| 38 |
+
- Latent dimensions: 16
|
| 39 |
+
- Sample rate: 44.1kHz
|
| 40 |
+
- Special features: Scaled down for real-time performance on RPi 5
|
| 41 |
+
|
| 42 |
+
## Features
|
| 43 |
+
- All models are exported for streaming inference
|
| 44 |
+
- Compatible with nn~, NN.ar, and rave-supercollider
|
| 45 |
+
- Models focus on encoder-decoder architecture without prior networks
|
| 46 |
+
- Training checkpoints provided for transfer learning
|
| 47 |
+
- For training, use the Intelligent Instruments Lab RAVE fork: https://github.com/victor-shepardson/RAVE
|
| 48 |
+
|
| 49 |
+
## Training Details
|
| 50 |
+
|
| 51 |
+
### Training Data
|
| 52 |
+
|
| 53 |
+
The models were trained on the [Freesound Loop Dataset (FSL10K)](https://zenodo.org/records/3967852), a comprehensive collection of musical loops curated for machine learning applications. The dataset consists of 9,455 loops from Freesound.org.
|
| 54 |
+
All sounds in the dataset are licensed under various Creative Commons licenses.
|
| 55 |
+
|
| 56 |
+
### Training Procedure
|
| 57 |
+
|
| 58 |
+
Training checkpoints are provided for both models to enable transfer learning on custom datasets.
|
| 59 |
+
|
| 60 |
+
## Citation
|
| 61 |
+
|
| 62 |
+
```bibtex
|
| 63 |
+
@misc {tangible_music_lab_2025,
|
| 64 |
+
author = { {Tangible Music Lab} },
|
| 65 |
+
title = { RAVE Models },
|
| 66 |
+
year = 2025,
|
| 67 |
+
url = { https://huggingface.co/TangibleMusicLab/rave-models },
|
| 68 |
+
publisher = { Hugging Face }
|
| 69 |
+
}
|