Tangible-Music-Lab commited on
Commit
75abbd1
·
verified ·
1 Parent(s): 945db8c

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +69 -0
README.md ADDED
@@ -0,0 +1,69 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # RAVE Models by Tangible Music Lab
2
+
3
+ This is a collection of RAVE (Realtime Audio Variational autoEncoder) models trained by the Tangible Music Lab for audio generation and transformation. The aim of this repository is to provide musicians with pre-trained models for building embedded RAVE models on the Raspberry Pi platform or similar, for physical hardware and tangible interface development for sound and music experimentation. These models enable real-time audio manipulation and generation while being optimized for resource-constrained environments, making them ideal for interactive musical instruments and sound installations.
4
+
5
+ ## Model Details
6
+
7
+ ### Model Description
8
+
9
+ - **Developed by:** Tangible Music Lab
10
+ - **Model type:** RAVE (Realtime Audio Variational autoEncoder)
11
+ - **License:** MIT
12
+
13
+ ### Model Sources
14
+
15
+ - **Repository:** https://huggingface.co/TangibleMusicLab/rave-models
16
+ - **Training Code:** https://github.com/victor-shepardson/RAVE
17
+
18
+ ## Uses
19
+
20
+ ### Direct Use
21
+
22
+ These models are designed for real-time audio generation and transformation. They can be used with:
23
+ - nn~
24
+ - NN.ar
25
+ - rave-supercollider
26
+
27
+ ### Models
28
+
29
+ #### tam_freesoundloop10k_default_z16.ts
30
+ - Dataset: Freesound Loop Dataset
31
+ - Model: RAVE v3 with default configuration
32
+ - Latent dimensions: 16
33
+ - Sample rate: 48kHz
34
+
35
+ #### tam_freesoundloop10k_raspi_z16.ts
36
+ - Dataset: Freesound Loop Dataset
37
+ - Model: Modified RAVE v3, optimized for Raspberry Pi 5
38
+ - Latent dimensions: 16
39
+ - Sample rate: 44.1kHz
40
+ - Special features: Scaled down for real-time performance on RPi 5
41
+
42
+ ## Features
43
+ - All models are exported for streaming inference
44
+ - Compatible with nn~, NN.ar, and rave-supercollider
45
+ - Models focus on encoder-decoder architecture without prior networks
46
+ - Training checkpoints provided for transfer learning
47
+ - For training, use the Intelligent Instruments Lab RAVE fork: https://github.com/victor-shepardson/RAVE
48
+
49
+ ## Training Details
50
+
51
+ ### Training Data
52
+
53
+ The models were trained on the [Freesound Loop Dataset (FSL10K)](https://zenodo.org/records/3967852), a comprehensive collection of musical loops curated for machine learning applications. The dataset consists of 9,455 loops from Freesound.org.
54
+ All sounds in the dataset are licensed under various Creative Commons licenses.
55
+
56
+ ### Training Procedure
57
+
58
+ Training checkpoints are provided for both models to enable transfer learning on custom datasets.
59
+
60
+ ## Citation
61
+
62
+ ```bibtex
63
+ @misc {tangible_music_lab_2025,
64
+ author = { {Tangible Music Lab} },
65
+ title = { RAVE Models },
66
+ year = 2025,
67
+ url = { https://huggingface.co/TangibleMusicLab/rave-models },
68
+ publisher = { Hugging Face }
69
+ }