Gamahea commited on
Commit
ae449ae
Β·
verified Β·
1 Parent(s): cdf6676

Add comprehensive dataset card with folder structure documentation

Browse files
Files changed (1) hide show
  1. README.md +205 -1
README.md CHANGED
@@ -1,3 +1,207 @@
1
  ---
2
- license: apache-2.0
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ title: LEMM Training Data Repository
3
+ emoji: 🎡
4
+ colorFrom: purple
5
+ colorTo: pink
6
+ sdk: static
7
+ pinned: false
8
+ license: mit
9
  ---
10
+
11
+ # 🎡 LEMM Training Data Repository
12
+
13
+ Central storage for all LEMM training artifacts:
14
+ - **LoRA Adapters** - User-trained model adaptations
15
+ - **Training Datasets** - Prepared and curated audio datasets
16
+
17
+ Part of the [LEMM (Let Everyone Make Music)](https://huggingface.co/spaces/Gamahea/lemm-test-100) project.
18
+
19
+ ---
20
+
21
+ ## πŸ“ Repository Structure
22
+
23
+ ```
24
+ lemmdata/
25
+ β”œβ”€β”€ loras/ # LoRA Adapters (ZIP files)
26
+ β”‚ β”œβ”€β”€ jazz-v1.zip # Example: Jazz style adapter
27
+ β”‚ β”œβ”€β”€ metal-data1.zip # Example: Metal style adapter
28
+ β”‚ └── ...
29
+ β”‚
30
+ └── datasets/ # Training Datasets (ZIP files)
31
+ β”œβ”€β”€ gtzan_prepared.zip # Example: GTZAN dataset
32
+ β”œβ”€β”€ user_dataset_123.zip # Example: User-uploaded dataset
33
+ └── ...
34
+ ```
35
+
36
+ ---
37
+
38
+ ## 🎨 LoRA Adapters (`loras/`)
39
+
40
+ Each LoRA is packaged as a ZIP file containing:
41
+
42
+ - **`final_model.pt`** - Trained LoRA weights (PyTorch checkpoint)
43
+ - **`config.yaml`** - Training hyperparameters and settings
44
+ - **`metadata.json`** - Training statistics, timestamps, dataset info
45
+ - **`README.md`** - Documentation and usage instructions
46
+
47
+ ### How to Use
48
+
49
+ **Download in LEMM:**
50
+ 1. Go to [LEMM Space](https://huggingface.co/spaces/Gamahea/lemm-test-100)
51
+ 2. Navigate to "LoRA Management" tab
52
+ 3. Click "Sync from HuggingFace"
53
+ 4. Select LoRA from dropdown
54
+ 5. Use in generation or continue training
55
+
56
+ **Download via Code:**
57
+ ```python
58
+ from huggingface_hub import hf_hub_download
59
+ import zipfile
60
+
61
+ # Download LoRA ZIP
62
+ zip_path = hf_hub_download(
63
+ repo_id="Gamahea/lemmdata",
64
+ repo_type="dataset",
65
+ filename="loras/jazz-v1.zip"
66
+ )
67
+
68
+ # Extract
69
+ with zipfile.ZipFile(zip_path, 'r') as zipf:
70
+ zipf.extractall("./my_loras/jazz-v1")
71
+ ```
72
+
73
+ ---
74
+
75
+ ## πŸ“Š Training Datasets (`datasets/`)
76
+
77
+ Each dataset is packaged as a ZIP file containing:
78
+
79
+ - **`dataset_info.json`** - Metadata (size, format, split ratios)
80
+ - **`train/`** - Training audio files
81
+ - **`val/`** - Validation audio files
82
+
83
+ ### Supported Formats
84
+
85
+ - **Audio**: WAV, MP3, FLAC, OGG
86
+ - **Sample Rate**: 44.1kHz or 48kHz recommended
87
+ - **Channels**: Mono or Stereo
88
+
89
+ ### How to Use
90
+
91
+ **Download in LEMM:**
92
+ 1. Go to [LEMM Space](https://huggingface.co/spaces/Gamahea/lemm-test-100)
93
+ 2. Navigate to "Training" tab
94
+ 3. Click "Import Dataset" β†’ "From HuggingFace"
95
+ 4. Select dataset
96
+ 5. Use for training
97
+
98
+ **Download via Code:**
99
+ ```python
100
+ from huggingface_hub import hf_hub_download
101
+ import zipfile
102
+
103
+ # Download dataset ZIP
104
+ zip_path = hf_hub_download(
105
+ repo_id="Gamahea/lemmdata",
106
+ repo_type="dataset",
107
+ filename="datasets/gtzan_prepared.zip"
108
+ )
109
+
110
+ # Extract
111
+ with zipfile.ZipFile(zip_path, 'r') as zipf:
112
+ zipf.extractall("./my_datasets/gtzan_prepared")
113
+ ```
114
+
115
+ ---
116
+
117
+ ## πŸš€ Contributing
118
+
119
+ ### Upload LoRA
120
+
121
+ Train a LoRA in [LEMM Space](https://huggingface.co/spaces/Gamahea/lemm-test-100):
122
+ 1. Prepare or select a dataset
123
+ 2. Configure training parameters
124
+ 3. Start training
125
+ 4. LoRA automatically packaged as ZIP and uploaded to `loras/{your-lora-name}.zip`
126
+
127
+ ### Upload Dataset
128
+
129
+ Prepare a dataset and export:
130
+ 1. Upload audio files to LEMM
131
+ 2. Use dataset preparation tools
132
+ 3. Export as prepared dataset
133
+ 4. Dataset packaged as ZIP and uploaded to `datasets/{your-dataset-name}.zip`
134
+
135
+ ---
136
+
137
+ ## πŸ“ Naming Conventions
138
+
139
+ ### LoRA Names
140
+ - **Format**: `{style}-{variant}_{version}`
141
+ - **Examples**:
142
+ - `jazz-bebop_v1`
143
+ - `rock-heavy_v2`
144
+ - `classical-piano_v1`
145
+
146
+ ### Dataset Names
147
+ - **Format**: `{source}_{description}`
148
+ - **Examples**:
149
+ - `gtzan_prepared`
150
+ - `user_dataset_1702987654`
151
+ - `opensinger_vocals`
152
+
153
+ ---
154
+
155
+ ## πŸ” Authentication
156
+
157
+ **Read Access**: Public (no authentication required)
158
+
159
+ **Write Access**: Requires HuggingFace token
160
+ - Only LEMM Space can upload
161
+ - User-trained artifacts auto-upload
162
+ - Token managed via HF Space secrets
163
+
164
+ ---
165
+
166
+ ## πŸ“š Related Resources
167
+
168
+ - **LEMM Space**: https://huggingface.co/spaces/Gamahea/lemm-test-100
169
+ - **GitHub Repo**: https://github.com/Gamahea/Angen
170
+ - **DiffRhythm2**: Music generation model with vocals
171
+ - **MuQ-MuLan**: Music style encoding
172
+
173
+ ---
174
+
175
+ ## πŸŽ“ Training Best Practices
176
+
177
+ ### LoRA Training
178
+ - **Dataset Size**: 100+ clips minimum
179
+ - **LoRA Rank**: 8-32 for most styles
180
+ - **Learning Rate**: 1e-4 to 1e-3
181
+ - **Epochs**: 20-50 depending on dataset
182
+
183
+ ### Dataset Preparation
184
+ - **Clip Length**: 10-30 seconds
185
+ - **Audio Quality**: Clean, well-produced
186
+ - **Consistency**: Similar genre/style
187
+ - **Diversity**: Varied within target style
188
+
189
+ ---
190
+
191
+ ## πŸ“„ License
192
+
193
+ MIT License - Free to use, modify, and share.
194
+
195
+ All contributed LoRAs and datasets inherit this license unless otherwise specified.
196
+
197
+ ---
198
+
199
+ ## 🏷️ Tags
200
+
201
+ `music-generation` `lora` `diffrhythm2` `audio` `training-data` `datasets` `models` `lemm`
202
+
203
+ ---
204
+
205
+ **Last Updated**: December 2025
206
+ **Repository**: https://huggingface.co/datasets/Gamahea/lemmdata
207
+ **LEMM Space**: https://huggingface.co/spaces/Gamahea/lemm-test-100