Update README.md
Browse files
README.md
CHANGED
|
@@ -77,4 +77,40 @@ Evaluated on `seed-tts-eval`. **GLM-TTS_RL** achieves the lowest Character Error
|
|
| 77 |
```bash
|
| 78 |
git clone [https://github.com/zai-org/GLM-TTS.git](https://github.com/zai-org/GLM-TTS.git)
|
| 79 |
cd GLM-TTS
|
| 80 |
-
pip install -r requirements.txt
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 77 |
```bash
|
| 78 |
git clone [https://github.com/zai-org/GLM-TTS.git](https://github.com/zai-org/GLM-TTS.git)
|
| 79 |
cd GLM-TTS
|
| 80 |
+
pip install -r requirements.txt
|
| 81 |
+
```
|
| 82 |
+
|
| 83 |
+
#### Command Line Inference
|
| 84 |
+
|
| 85 |
+
```bash
|
| 86 |
+
python glmtts_inference.py \
|
| 87 |
+
--data=example_zh \
|
| 88 |
+
--exp_name=_test \
|
| 89 |
+
--use_cache \
|
| 90 |
+
# --phoneme # Add this flag to enable phoneme capabilities.
|
| 91 |
+
```
|
| 92 |
+
|
| 93 |
+
#### Using Pre-trained Script
|
| 94 |
+
|
| 95 |
+
```bash
|
| 96 |
+
bash glmtts_inference.sh
|
| 97 |
+
```
|
| 98 |
+
|
| 99 |
+
## 🙏 Acknowledgments & Citation
|
| 100 |
+
|
| 101 |
+
We thank the following open-source projects for their support:
|
| 102 |
+
|
| 103 |
+
- [CosyVoice](https://github.com/FunAudioLLM/CosyVoice) - Providing frontend processing framework and high-quality vocoder
|
| 104 |
+
- [Llama](https://github.com/meta-llama/llama) - Providing basic language model architecture
|
| 105 |
+
- [Vocos](https://github.com/charactr-platform/vocos) - Providing high-quality vocoder
|
| 106 |
+
- [GRPO-Zero](https://github.com/policy-gradient/GRPO-Zero) - Reinforcement learning algorithm implementation inspiration
|
| 107 |
+
|
| 108 |
+
If you use GLM-TTS in your research, please cite:
|
| 109 |
+
|
| 110 |
+
```bibtex
|
| 111 |
+
@misc{glmtts2025,
|
| 112 |
+
title={GLM-TTS: Controllable & Emotion-Expressive Zero-shot TTS with Multi-Reward Reinforcement Learning},
|
| 113 |
+
author={CogAudio Group Members},
|
| 114 |
+
year={2025},
|
| 115 |
+
publisher={Zhipu AI Inc}
|
| 116 |
+
}
|