Add task category, LeVo paper, project page, code, and sample usage, and remove redundant license
#2
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -1,23 +1,72 @@
|
|
| 1 |
---
|
| 2 |
-
license: apache-2.0
|
| 3 |
language:
|
| 4 |
- en
|
| 5 |
- zh
|
|
|
|
| 6 |
tags:
|
| 7 |
- music
|
|
|
|
|
|
|
| 8 |
---
|
| 9 |
|
| 10 |
# Song Structure and Lyric Dataset (SSLD-200)
|
| 11 |
|
| 12 |
-
|
|
|
|
|
|
|
|
|
|
| 13 |
|
| 14 |
-
|
|
|
|
|
|
|
| 15 |
- The structure is the label from StructureAnalysis for the segment.
|
| 16 |
- The start and end are the segment’s start and end times.
|
| 17 |
- The lyric is the recognized lyrics.
|
| 18 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 19 |
|
| 20 |
## Citation
|
|
|
|
|
|
|
| 21 |
```
|
| 22 |
@misc{tan2025songpreppreprocessingframeworkendtoend,
|
| 23 |
title={SongPrep: A Preprocessing Framework and End-to-end Model for Full-song Structure Parsing and Lyrics Transcription},
|
|
@@ -28,7 +77,4 @@ The **lyric_norm** in the format \[structure\]\[start:end\]lyric
|
|
| 28 |
primaryClass={eess.AS},
|
| 29 |
url={https://arxiv.org/abs/2509.17404},
|
| 30 |
}
|
| 31 |
-
```
|
| 32 |
-
|
| 33 |
-
## License
|
| 34 |
-
The code and weights in this repository is released in the [LICENSE](LICENSE) file.
|
|
|
|
| 1 |
---
|
|
|
|
| 2 |
language:
|
| 3 |
- en
|
| 4 |
- zh
|
| 5 |
+
license: apache-2.0
|
| 6 |
tags:
|
| 7 |
- music
|
| 8 |
+
task_categories:
|
| 9 |
+
- automatic-speech-recognition
|
| 10 |
---
|
| 11 |
|
| 12 |
# Song Structure and Lyric Dataset (SSLD-200)
|
| 13 |
|
| 14 |
+
This dataset is used as an evaluation benchmark in the paper [LeVo: High-Quality Song Generation with Multi-Preference Alignment](https://huggingface.co/papers/2506.07520).
|
| 15 |
+
|
| 16 |
+
Project page: [https://levo-demo.github.io](https://levo-demo.github.io)
|
| 17 |
+
Code: [https://github.com/tencent-ailab/songgeneration](https://github.com/tencent-ailab/songgeneration)
|
| 18 |
|
| 19 |
+
DataSet used to evaluate song structure parsing and lyrics transcription. SSLD-200 consists of 200 songs, 100 English and 100 Chinese, collected entirely from YouTube, with a total duration of 13.9 hours.
|
| 20 |
+
|
| 21 |
+
The **lyric_norm** in the format \\[structure\\]\\[start:end\\]lyric
|
| 22 |
- The structure is the label from StructureAnalysis for the segment.
|
| 23 |
- The start and end are the segment’s start and end times.
|
| 24 |
- The lyric is the recognized lyrics.
|
| 25 |
|
| 26 |
+
## Sample Usage
|
| 27 |
+
|
| 28 |
+
To use this dataset in conjunction with the associated `SongGeneration` models, follow these steps for inference.
|
| 29 |
+
|
| 30 |
+
First, download the required runtime components and a specific model checkpoint:
|
| 31 |
+
|
| 32 |
+
```bash
|
| 33 |
+
# Download runtime components
|
| 34 |
+
huggingface-cli download lglg666/SongGeneration-Runtime --local-dir ./runtime
|
| 35 |
+
mv runtime/ckpt ckpt
|
| 36 |
+
mv runtime/third_party third_party
|
| 37 |
+
|
| 38 |
+
# Download a specific model checkpoint (e.g., SongGeneration-base-new)
|
| 39 |
+
huggingface-cli download lglg666/SongGeneration-base-new --local-dir ./songgeneration_base_new
|
| 40 |
+
```
|
| 41 |
+
|
| 42 |
+
Once setup, you can run the inference script. You need to provide a `ckpt_path` (the directory where you downloaded the model checkpoint), an input `lyrics.jsonl` file, and an `output_path`.
|
| 43 |
+
|
| 44 |
+
```bash
|
| 45 |
+
sh generate.sh ckpt_path lyrics.jsonl output_path
|
| 46 |
+
```
|
| 47 |
+
|
| 48 |
+
**Input Format (`lyrics.jsonl`):**
|
| 49 |
+
|
| 50 |
+
Each line in the `.jsonl` file represents an individual song generation request and must contain the following fields:
|
| 51 |
+
|
| 52 |
+
- `idx`: A unique identifier for the output song.
|
| 53 |
+
- `gt_lyric`: The lyrics and song structure, following the format `[Structure] Text`. For example:
|
| 54 |
+
`[intro-short] ; [verse] These faded memories of us. I can't erase the tears you cried before. Unchained this heart to find its way. My peace won't beg you to stay ; [bridge] If ever your truth still remains. Turn around and see. Life rearranged its games. All these lessons in mistakes. Even years may never erase ; [inst-short] ; [chorus] Like a fool begs for supper. I find myself waiting for her. Only to find the broken pieces of my heart. That was needed for my soul to love again ; [outro-short]`
|
| 55 |
+
- `descriptions`: (Optional) Custom text prompt for generation attributes like gender, timbre, genre, emotion, instrument, and BPM (e.g., `"female, dark, pop, sad, piano and drums."`).
|
| 56 |
+
- `prompt_audio_path`: (Optional) Path to a 10-second reference audio file.
|
| 57 |
+
- `auto_prompt_audio_type`: (Optional) Used if `prompt_audio_path` is not provided, automatically selects a reference audio based on a given style (e.g., `'Pop'`, `'R&B'`, `'Dance'`, `'Jazz'`, etc.).
|
| 58 |
+
|
| 59 |
+
Example command:
|
| 60 |
+
|
| 61 |
+
```bash
|
| 62 |
+
sh generate.sh songgeneration_base_new sample/lyrics.jsonl sample/output
|
| 63 |
+
```
|
| 64 |
+
|
| 65 |
+
Additional flags can be used for specific inference scenarios, such as `--low_mem` for low-memory inference, `--not_use_flash_attn` to disable Flash Attention, or `--separate` to generate separated vocal and accompaniment tracks.
|
| 66 |
|
| 67 |
## Citation
|
| 68 |
+
|
| 69 |
+
The dataset itself is detailed in the following work:
|
| 70 |
```
|
| 71 |
@misc{tan2025songpreppreprocessingframeworkendtoend,
|
| 72 |
title={SongPrep: A Preprocessing Framework and End-to-end Model for Full-song Structure Parsing and Lyrics Transcription},
|
|
|
|
| 77 |
primaryClass={eess.AS},
|
| 78 |
url={https://arxiv.org/abs/2509.17404},
|
| 79 |
}
|
| 80 |
+
```
|
|
|
|
|
|
|
|
|