File size: 5,227 Bytes
cdf6676 ae449ae cdf6676 ae449ae |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 |
---
title: LEMM Training Data Repository
emoji: π΅
colorFrom: purple
colorTo: pink
sdk: static
pinned: false
license: mit
---
# π΅ LEMM Training Data Repository
Central storage for all LEMM training artifacts:
- **LoRA Adapters** - User-trained model adaptations
- **Training Datasets** - Prepared and curated audio datasets
Part of the [LEMM (Let Everyone Make Music)](https://huggingface.co/spaces/Gamahea/lemm-test-100) project.
---
## π Repository Structure
```
lemmdata/
βββ loras/ # LoRA Adapters (ZIP files)
β βββ jazz-v1.zip # Example: Jazz style adapter
β βββ metal-data1.zip # Example: Metal style adapter
β βββ ...
β
βββ datasets/ # Training Datasets (ZIP files)
βββ gtzan_prepared.zip # Example: GTZAN dataset
βββ user_dataset_123.zip # Example: User-uploaded dataset
βββ ...
```
---
## π¨ LoRA Adapters (`loras/`)
Each LoRA is packaged as a ZIP file containing:
- **`final_model.pt`** - Trained LoRA weights (PyTorch checkpoint)
- **`config.yaml`** - Training hyperparameters and settings
- **`metadata.json`** - Training statistics, timestamps, dataset info
- **`README.md`** - Documentation and usage instructions
### How to Use
**Download in LEMM:**
1. Go to [LEMM Space](https://huggingface.co/spaces/Gamahea/lemm-test-100)
2. Navigate to "LoRA Management" tab
3. Click "Sync from HuggingFace"
4. Select LoRA from dropdown
5. Use in generation or continue training
**Download via Code:**
```python
from huggingface_hub import hf_hub_download
import zipfile
# Download LoRA ZIP
zip_path = hf_hub_download(
repo_id="Gamahea/lemmdata",
repo_type="dataset",
filename="loras/jazz-v1.zip"
)
# Extract
with zipfile.ZipFile(zip_path, 'r') as zipf:
zipf.extractall("./my_loras/jazz-v1")
```
---
## π Training Datasets (`datasets/`)
Each dataset is packaged as a ZIP file containing:
- **`dataset_info.json`** - Metadata (size, format, split ratios)
- **`train/`** - Training audio files
- **`val/`** - Validation audio files
### Supported Formats
- **Audio**: WAV, MP3, FLAC, OGG
- **Sample Rate**: 44.1kHz or 48kHz recommended
- **Channels**: Mono or Stereo
### How to Use
**Download in LEMM:**
1. Go to [LEMM Space](https://huggingface.co/spaces/Gamahea/lemm-test-100)
2. Navigate to "Training" tab
3. Click "Import Dataset" β "From HuggingFace"
4. Select dataset
5. Use for training
**Download via Code:**
```python
from huggingface_hub import hf_hub_download
import zipfile
# Download dataset ZIP
zip_path = hf_hub_download(
repo_id="Gamahea/lemmdata",
repo_type="dataset",
filename="datasets/gtzan_prepared.zip"
)
# Extract
with zipfile.ZipFile(zip_path, 'r') as zipf:
zipf.extractall("./my_datasets/gtzan_prepared")
```
---
## π Contributing
### Upload LoRA
Train a LoRA in [LEMM Space](https://huggingface.co/spaces/Gamahea/lemm-test-100):
1. Prepare or select a dataset
2. Configure training parameters
3. Start training
4. LoRA automatically packaged as ZIP and uploaded to `loras/{your-lora-name}.zip`
### Upload Dataset
Prepare a dataset and export:
1. Upload audio files to LEMM
2. Use dataset preparation tools
3. Export as prepared dataset
4. Dataset packaged as ZIP and uploaded to `datasets/{your-dataset-name}.zip`
---
## π Naming Conventions
### LoRA Names
- **Format**: `{style}-{variant}_{version}`
- **Examples**:
- `jazz-bebop_v1`
- `rock-heavy_v2`
- `classical-piano_v1`
### Dataset Names
- **Format**: `{source}_{description}`
- **Examples**:
- `gtzan_prepared`
- `user_dataset_1702987654`
- `opensinger_vocals`
---
## π Authentication
**Read Access**: Public (no authentication required)
**Write Access**: Requires HuggingFace token
- Only LEMM Space can upload
- User-trained artifacts auto-upload
- Token managed via HF Space secrets
---
## π Related Resources
- **LEMM Space**: https://huggingface.co/spaces/Gamahea/lemm-test-100
- **GitHub Repo**: https://github.com/Gamahea/Angen
- **DiffRhythm2**: Music generation model with vocals
- **MuQ-MuLan**: Music style encoding
---
## π Training Best Practices
### LoRA Training
- **Dataset Size**: 100+ clips minimum
- **LoRA Rank**: 8-32 for most styles
- **Learning Rate**: 1e-4 to 1e-3
- **Epochs**: 20-50 depending on dataset
### Dataset Preparation
- **Clip Length**: 10-30 seconds
- **Audio Quality**: Clean, well-produced
- **Consistency**: Similar genre/style
- **Diversity**: Varied within target style
---
## π License
MIT License - Free to use, modify, and share.
All contributed LoRAs and datasets inherit this license unless otherwise specified.
---
## π·οΈ Tags
`music-generation` `lora` `diffrhythm2` `audio` `training-data` `datasets` `models` `lemm`
---
**Last Updated**: December 2025
**Repository**: https://huggingface.co/datasets/Gamahea/lemmdata
**LEMM Space**: https://huggingface.co/spaces/Gamahea/lemm-test-100
|