metadata
library_name: transformers
datasets:
- malaysia-ai/Multilingual-TTS
- Scicom-intl/Emilia-YODAS-Voice-Conversion
- Scicom-intl/Malaysian-Emilia
base_model:
- Qwen/Qwen3-4B-Base
language:
- en
- ms
- zh
- ta
Multilingual-TTS-4B-Base
Continue pretraining Qwen/Qwen3-4B-Base on Multilingual Voice Conversion and TTS.
- Use neucodec as speech detokenizer, 50 TPS, output in 24k sample rate.
- Multi-speaker multilingual Voice Conversion, up to 35.88B tokens.
- Multi-speaker multilingual TTS more than 150 languages, up to 14.64B tokens.
- Flash Attention 3 10k context length varlen multipacking.
- BF16 training.
- MuonAdamW optimizer.