lemmdata / README.md
Gamahea's picture
Add comprehensive dataset card with folder structure documentation
ae449ae verified
---
title: LEMM Training Data Repository
emoji: 🎡
colorFrom: purple
colorTo: pink
sdk: static
pinned: false
license: mit
---
# 🎡 LEMM Training Data Repository
Central storage for all LEMM training artifacts:
- **LoRA Adapters** - User-trained model adaptations
- **Training Datasets** - Prepared and curated audio datasets
Part of the [LEMM (Let Everyone Make Music)](https://huggingface.co/spaces/Gamahea/lemm-test-100) project.
---
## πŸ“ Repository Structure
```
lemmdata/
β”œβ”€β”€ loras/ # LoRA Adapters (ZIP files)
β”‚ β”œβ”€β”€ jazz-v1.zip # Example: Jazz style adapter
β”‚ β”œβ”€β”€ metal-data1.zip # Example: Metal style adapter
β”‚ └── ...
β”‚
└── datasets/ # Training Datasets (ZIP files)
β”œβ”€β”€ gtzan_prepared.zip # Example: GTZAN dataset
β”œβ”€β”€ user_dataset_123.zip # Example: User-uploaded dataset
└── ...
```
---
## 🎨 LoRA Adapters (`loras/`)
Each LoRA is packaged as a ZIP file containing:
- **`final_model.pt`** - Trained LoRA weights (PyTorch checkpoint)
- **`config.yaml`** - Training hyperparameters and settings
- **`metadata.json`** - Training statistics, timestamps, dataset info
- **`README.md`** - Documentation and usage instructions
### How to Use
**Download in LEMM:**
1. Go to [LEMM Space](https://huggingface.co/spaces/Gamahea/lemm-test-100)
2. Navigate to "LoRA Management" tab
3. Click "Sync from HuggingFace"
4. Select LoRA from dropdown
5. Use in generation or continue training
**Download via Code:**
```python
from huggingface_hub import hf_hub_download
import zipfile
# Download LoRA ZIP
zip_path = hf_hub_download(
repo_id="Gamahea/lemmdata",
repo_type="dataset",
filename="loras/jazz-v1.zip"
)
# Extract
with zipfile.ZipFile(zip_path, 'r') as zipf:
zipf.extractall("./my_loras/jazz-v1")
```
---
## πŸ“Š Training Datasets (`datasets/`)
Each dataset is packaged as a ZIP file containing:
- **`dataset_info.json`** - Metadata (size, format, split ratios)
- **`train/`** - Training audio files
- **`val/`** - Validation audio files
### Supported Formats
- **Audio**: WAV, MP3, FLAC, OGG
- **Sample Rate**: 44.1kHz or 48kHz recommended
- **Channels**: Mono or Stereo
### How to Use
**Download in LEMM:**
1. Go to [LEMM Space](https://huggingface.co/spaces/Gamahea/lemm-test-100)
2. Navigate to "Training" tab
3. Click "Import Dataset" β†’ "From HuggingFace"
4. Select dataset
5. Use for training
**Download via Code:**
```python
from huggingface_hub import hf_hub_download
import zipfile
# Download dataset ZIP
zip_path = hf_hub_download(
repo_id="Gamahea/lemmdata",
repo_type="dataset",
filename="datasets/gtzan_prepared.zip"
)
# Extract
with zipfile.ZipFile(zip_path, 'r') as zipf:
zipf.extractall("./my_datasets/gtzan_prepared")
```
---
## πŸš€ Contributing
### Upload LoRA
Train a LoRA in [LEMM Space](https://huggingface.co/spaces/Gamahea/lemm-test-100):
1. Prepare or select a dataset
2. Configure training parameters
3. Start training
4. LoRA automatically packaged as ZIP and uploaded to `loras/{your-lora-name}.zip`
### Upload Dataset
Prepare a dataset and export:
1. Upload audio files to LEMM
2. Use dataset preparation tools
3. Export as prepared dataset
4. Dataset packaged as ZIP and uploaded to `datasets/{your-dataset-name}.zip`
---
## πŸ“ Naming Conventions
### LoRA Names
- **Format**: `{style}-{variant}_{version}`
- **Examples**:
- `jazz-bebop_v1`
- `rock-heavy_v2`
- `classical-piano_v1`
### Dataset Names
- **Format**: `{source}_{description}`
- **Examples**:
- `gtzan_prepared`
- `user_dataset_1702987654`
- `opensinger_vocals`
---
## πŸ” Authentication
**Read Access**: Public (no authentication required)
**Write Access**: Requires HuggingFace token
- Only LEMM Space can upload
- User-trained artifacts auto-upload
- Token managed via HF Space secrets
---
## πŸ“š Related Resources
- **LEMM Space**: https://huggingface.co/spaces/Gamahea/lemm-test-100
- **GitHub Repo**: https://github.com/Gamahea/Angen
- **DiffRhythm2**: Music generation model with vocals
- **MuQ-MuLan**: Music style encoding
---
## πŸŽ“ Training Best Practices
### LoRA Training
- **Dataset Size**: 100+ clips minimum
- **LoRA Rank**: 8-32 for most styles
- **Learning Rate**: 1e-4 to 1e-3
- **Epochs**: 20-50 depending on dataset
### Dataset Preparation
- **Clip Length**: 10-30 seconds
- **Audio Quality**: Clean, well-produced
- **Consistency**: Similar genre/style
- **Diversity**: Varied within target style
---
## πŸ“„ License
MIT License - Free to use, modify, and share.
All contributed LoRAs and datasets inherit this license unless otherwise specified.
---
## 🏷️ Tags
`music-generation` `lora` `diffrhythm2` `audio` `training-data` `datasets` `models` `lemm`
---
**Last Updated**: December 2025
**Repository**: https://huggingface.co/datasets/Gamahea/lemmdata
**LEMM Space**: https://huggingface.co/spaces/Gamahea/lemm-test-100