Add model card for SoulX-Singer
#1
by
nielsr HF Staff - opened
README.md
CHANGED
|
@@ -1,3 +1,44 @@
|
|
| 1 |
-
---
|
| 2 |
-
license: apache-2.0
|
| 3 |
-
--
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
pipeline_tag: audio-to-audio
|
| 4 |
+
---
|
| 5 |
+
|
| 6 |
+
# SoulX-Singer: Towards High-Quality Zero-Shot Singing Voice Synthesis
|
| 7 |
+
|
| 8 |
+
SoulX-Singer is a high-fidelity, zero-shot singing voice synthesis (SVS) model that enables the generation of realistic singing voices for unseen singers. It supports controllable singing generation conditioned on either symbolic musical scores (MIDI) or melodic representations (F0 contour), enabling flexible and expressive control in real-world production workflows.
|
| 9 |
+
|
| 10 |
+
[**Project Page**](https://soul-ailab.github.io/soulx-singer) | [**Paper (arXiv)**](https://arxiv.org/abs/2602.07803) | [**GitHub**](https://github.com/Soul-AILab/SoulX-Singer) | [**Online Demo**](https://huggingface.co/spaces/Soul-AILab/SoulX-Singer)
|
| 11 |
+
|
| 12 |
+
## Key Features
|
| 13 |
+
|
| 14 |
+
- **π€ Zero-Shot Singing**: Generate high-fidelity voices for unseen singers without any fine-tuning.
|
| 15 |
+
- **π΅ Flexible Control Modes**: Supports both Melody-conditioned (F0 contour) and Score-conditioned (MIDI notes) control.
|
| 16 |
+
- **π Large-Scale Dataset**: Trained on over 42,000 hours of aligned vocals, lyrics, and notes across Mandarin, English, and Cantonese.
|
| 17 |
+
- **π§βπ€ Timbre Cloning**: Preserves singer identity across different languages, styles, and edited lyrics.
|
| 18 |
+
- **βοΈ Singing Voice Editing**: Allows for modifying lyrics while maintaining natural prosody.
|
| 19 |
+
- **π Cross-Lingual Synthesis**: Disentangles timbre from content to ensure high-fidelity synthesis across supported languages.
|
| 20 |
+
|
| 21 |
+
## Usage
|
| 22 |
+
|
| 23 |
+
For detailed installation and inference instructions, please refer to the [official GitHub repository](https://github.com/Soul-AILab/SoulX-Singer).
|
| 24 |
+
|
| 25 |
+
The repository provides a `webui.py` for an interactive interface and example scripts for batch inference.
|
| 26 |
+
|
| 27 |
+
## Citation
|
| 28 |
+
|
| 29 |
+
If you use SoulX-Singer in your research, please cite the following:
|
| 30 |
+
|
| 31 |
+
```bibtex
|
| 32 |
+
@misc{soulxsinger,
|
| 33 |
+
title={SoulX-Singer: Towards High-Quality Zero-Shot Singing Voice Synthesis},
|
| 34 |
+
author={Jiale Qian and Hao Meng and Tian Zheng and Pengcheng Zhu and Haopeng Lin and Yuhang Dai and Hanke Xie and Wenxiao Cao and Ruixuan Shang and Jun Wu and Hongmei Liu and Hanlin Wen and Jian Zhao and Zhonglin Jiang and Yong Chen and Shunshun Yin and Ming Tao and Jianguo Wei and Lei Xie and Xinsheng Wang},
|
| 35 |
+
year={2026},
|
| 36 |
+
eprint={2602.07803},
|
| 37 |
+
archivePrefix={arXiv},
|
| 38 |
+
primaryClass={eess.AS},
|
| 39 |
+
url={https://arxiv.org/abs/2602.07803},
|
| 40 |
+
}
|
| 41 |
+
```
|
| 42 |
+
|
| 43 |
+
---
|
| 44 |
+
*SoulX-Singer is intended for academic research and legitimate applications. Please respect intellectual property and privacy when generating content.*
|