BayanSynthTTS

Arabic Text-to-Speech powered by CosyVoice3 with LoRA fine-tuning.

Text in. Speech out. Inference only without training or preprocessing.

GitHub: Ramendan/BayanSynthTTS

Features

Feature Details
Arabic TTS Natural-sounding Modern Standard Arabic
Auto-Tashkeel Automatic diacritization via mishkal (always on by default)
Voice Cloning Clone any voice from a 5-15 s clip (WAV/MP3/OGG/M4A/FLAC)
Example voices Two reference voices (default.wav and muffled-talking.wav) are included; add your own to voices/
Speed control Slow down or speed up synthesis (0.5โ€“2.0ร—)
LoRA Swapping Change checkpoints via conf/models.yaml no code edits
Streaming Chunk-by-chunk audio generation
Gradio UI Simple web interface included
CLI One-liner inference from terminal
Multilingual base CosyVoice3 supports many languages; Arabic LoRA ships by default

Multilingual note: the underlying CosyVoice3 base model is trained for zero-shot synthesis across a wide range of languages. BayanSynthTTS currently defaults to an Arabic-conditioned LoRA checkpoint and delivers the best results in Modern Standard Arabic. You are free to plug in other LoRA files (not provided here) for additional languages, though quality may vary.


Audio Demos

All samples were generated with this library. No post-processing applied.

# Description Duration
1 Basic synthesis, pre-diacritized ~5 s
2 Pre-diacritized text, mishkal off ~4 s
3 Voice cloning from muffled reference ~10 s
4 Longer passage, AI topic, 3 sentences ~17 s
5 Slow speed (0.80x) ~10 s
6 Fast speed (1.20x) ~5 s
7 Phonetics test: halqiyyat, tanwin, shaddah ~7 s
8 Flow and rhythm, connected speech ~9 s
9 Challenge: identical root, different diacritics ~5 s
10 Phonetics, alternate seed (seed=17) ~9 s
11 Flow, alternate seed (seed=99) ~10 s
12 Instruct prompt: warm newsreader style ~8 s

1. Basic synthesis

ู…ูŽุฑู’ุญูŽุจู‹ุงุŒ ุฃูŽู†ูŽุง ุจูŽูŠูŽุงู†ู’ุณููŠู†ู’ุซุŒ ู†ูุธูŽุงู…ูŒ ู„ูุชูŽูˆู’ู„ููŠุฏู ุงู„ู’ูƒูŽู„ูŽุงู…ู ุงู„ู’ุนูŽุฑูŽุจููŠูู‘.

Hello, I am BayanSynth, a system for generating Arabic speech.

from bayansynthtts import BayanSynthTTS
tts = BayanSynthTTS()
audio = tts.synthesize(
    "ู…ูŽุฑู’ุญูŽุจู‹ุงุŒ ุฃูŽู†ูŽุง ุจูŽูŠูŽุงู†ู’ุณููŠู†ู’ุซุŒ ู†ูุธูŽุงู…ูŒ ู„ูุชูŽูˆู’ู„ููŠุฏู ุงู„ู’ูƒูŽู„ูŽุงู…ู ุงู„ู’ุนูŽุฑูŽุจููŠูู‘.",
    auto_tashkeel=False,
)
tts.save_wav(audio, "output.wav")


2. Pre-diacritized text (mishkal off)

ุฅูู†ูŽู‘ ุงู„ู„ูู‘ุบูŽุฉูŽ ุงู„ู’ุนูŽุฑูŽุจููŠูŽู‘ุฉูŽ ูƒูŽู†ู’ุฒูŒ ู…ูู†ูŽ ุงู„ุซูŽู‘ู‚ูŽุงููŽุฉู ูˆูŽุงู„ุชูู‘ุฑูŽุงุซู.

The Arabic language is a treasure of culture and heritage.

audio = tts.synthesize(
    "ุฅูู†ูŽู‘ ุงู„ู„ูู‘ุบูŽุฉูŽ ุงู„ู’ุนูŽุฑูŽุจููŠูŽู‘ุฉูŽ ูƒูŽู†ู’ุฒูŒ ู…ูู†ูŽ ุงู„ุซูŽู‘ู‚ูŽุงููŽุฉู ูˆูŽุงู„ุชูู‘ุฑูŽุงุซู.",
    auto_tashkeel=False,
)


3. Voice cloning

ู‡ูŽุฐูŽุง ุงู„ุตูŽู‘ูˆู’ุชู ู…ูุณู’ุชูŽู†ู’ุณูŽุฎูŒ ู…ูู†ู’ ู…ูŽู‚ู’ุทูŽุนู ุตูŽูˆู’ุชููŠูู‘ ู‚ูŽุตููŠุฑู. ูŠูู…ู’ูƒูู†ููƒูŽ ุงุณู’ุชูุฎู’ุฏูŽุงู…ู ุฃูŽูŠูู‘ ู…ูŽู‚ู’ุทูŽุนู ุจูู…ูุฏูŽู‘ุฉู ุฎูŽู…ู’ุณู ุฅูู„ูŽู‰ ุฎูŽู…ู’ุณูŽ ุนูŽุดูŽุฑูŽุฉูŽ ุซูŽุงู†ููŠูŽุฉู‹.

This voice is cloned from a short audio clip. You can use any clip between five and fifteen seconds.

audio = tts.synthesize(
    "ู‡ูŽุฐูŽุง ุงู„ุตูŽู‘ูˆู’ุชู ู…ูุณู’ุชูŽู†ู’ุณูŽุฎูŒ ู…ูู†ู’ ู…ูŽู‚ู’ุทูŽุนู ุตูŽูˆู’ุชููŠูู‘ ู‚ูŽุตููŠุฑู. "
    "ูŠูู…ู’ูƒูู†ููƒูŽ ุงุณู’ุชูุฎู’ุฏูŽุงู…ู ุฃูŽูŠูู‘ ู…ูŽู‚ู’ุทูŽุนู ุจูู…ูุฏูŽู‘ุฉู ุฎูŽู…ู’ุณู ุฅูู„ูŽู‰ ุฎูŽู…ู’ุณูŽ ุนูŽุดูŽุฑูŽุฉูŽ ุซูŽุงู†ููŠูŽุฉู‹.",
    ref_audio="voices/muffled_trim.wav",
    auto_tashkeel=False,
)

Reference clip:

Result:


4. Longer passage (auto-tashkeel, speed 0.88)

ุงู„ุฐูƒุงุก ุงู„ุงุตุทู†ุงุนูŠ ู‡ูˆ ุฃุญุฏ ุฃุจุฑุฒ ุงู„ุชุทูˆุฑุงุช ุงู„ุชูƒู†ูˆู„ูˆุฌูŠุฉ ููŠ ุนุตุฑู†ุง ุงู„ุญุฏูŠุซ. ูŠุนุชู…ุฏ ุนู„ู‰ ุชุญู„ูŠู„ ูƒู…ูŠุงุช ุถุฎู…ุฉ ู…ู† ุงู„ุจูŠุงู†ุงุช ู„ุงุณุชุฎู„ุงุต ุฃู†ู…ุงุท ู…ุนู‚ุฏุฉ. ูˆู…ู† ุฃุจุฑุฒ ุชุทุจูŠู‚ุงุชู‡ ู†ุธู… ุงู„ุชุนุฑู ุนู„ู‰ ุงู„ุตูˆุช ูˆุชุฑุฌู…ุฉ ุงู„ู„ุบุงุช ูˆุชูˆู„ูŠุฏ ุงู„ู†ุตูˆุต.

Artificial intelligence is one of the most prominent technological advances of our era. It relies on analyzing massive amounts of data to extract complex patterns. Among its most notable applications: speech recognition, language translation, and text generation.

audio = tts.synthesize(
    "ุงู„ุฐูƒุงุก ุงู„ุงุตุทู†ุงุนูŠ ู‡ูˆ ุฃุญุฏ ุฃุจุฑุฒ ุงู„ุชุทูˆุฑุงุช ุงู„ุชูƒู†ูˆู„ูˆุฌูŠุฉ ููŠ ุนุตุฑู†ุง ุงู„ุญุฏูŠุซ. "
    "ูŠุนุชู…ุฏ ุนู„ู‰ ุชุญู„ูŠู„ ูƒู…ูŠุงุช ุถุฎู…ุฉ ู…ู† ุงู„ุจูŠุงู†ุงุช ู„ุงุณุชุฎู„ุงุต ุฃู†ู…ุงุท ู…ุนู‚ุฏุฉ. "
    "ูˆู…ู† ุฃุจุฑุฒ ุชุทุจูŠู‚ุงุชู‡ ู†ุธู… ุงู„ุชุนุฑู ุนู„ู‰ ุงู„ุตูˆุช ูˆุชุฑุฌู…ุฉ ุงู„ู„ุบุงุช ูˆุชูˆู„ูŠุฏ ุงู„ู†ุตูˆุต.",
    auto_tashkeel=True,
    speed=0.88,
)


5. Speed control

ู…ูŽุฑู’ุญูŽุจุงู‹ ุจููƒูู…ู’ ูููŠ ุจูŽูŠูŽุงู†ู’ุณููŠู†ู’ุซู. ู‡ูŽุฐูŽุง ุชูŽูˆู’ู„ููŠุฏูŒ ุจูุณูุฑู’ุนูŽุฉู ู…ูุฎูŽููŽู‘ุถูŽุฉู ู„ูู„ุชูŽู‘ูˆู’ุถููŠุญู.

Welcome to BayanSynth. This is synthesis at reduced speed for demonstration.

TEXT = "ู…ูŽุฑู’ุญูŽุจุงู‹ ุจููƒูู…ู’ ูููŠ ุจูŽูŠูŽุงู†ู’ุณููŠู†ู’ุซู. ู‡ูŽุฐูŽุง ุชูŽูˆู’ู„ููŠุฏูŒ ุจูุณูุฑู’ุนูŽุฉู ู…ูุฎูŽููŽู‘ุถูŽุฉู ู„ูู„ุชูŽู‘ูˆู’ุถููŠุญู."
audio = tts.synthesize(TEXT, speed=0.80, auto_tashkeel=False)

Slow (0.80ร—):

Fast (1.20ร—):


6. Phonetics test: halqiyyat, tanwin, shaddah

Designed to exercise pharyngeal/velar consonants, gemination, and nunation at once:

ุงู„ู’ุฌูŽูˆู’ุฏูŽุฉู ุงู„ู’ุนูŽุงู„ููŠูŽุฉู ู„ูุชูŽู‚ู’ู†ููŠูŽู‘ุงุชู ุงู„ุฐูŽู‘ูƒูŽุงุกู ุงู„ุงุตู’ุทูู†ูŽุงุนููŠูู‘ ุชูุณูŽุงู‡ูู…ู ูููŠ ุจูู†ูŽุงุกู ู…ูุณู’ุชูŽู‚ู’ุจูŽู„ู ุจูŽุงู‡ูุฑู ู„ูู„ู’ุฃูŽุฌู’ูŠูŽุงู„ู.

The high quality of AI technologies contributes to building a brilliant future for generations to come.

audio = tts.synthesize(
    "ุงู„ู’ุฌูŽูˆู’ุฏูŽุฉู ุงู„ู’ุนูŽุงู„ููŠูŽุฉู ู„ูุชูŽู‚ู’ู†ููŠูŽู‘ุงุชู ุงู„ุฐูŽู‘ูƒูŽุงุกู ุงู„ุงุตู’ุทูู†ูŽุงุนููŠูู‘ "
    "ุชูุณูŽุงู‡ูู…ู ูููŠ ุจูู†ูŽุงุกู ู…ูุณู’ุชูŽู‚ู’ุจูŽู„ู ุจูŽุงู‡ูุฑู ู„ูู„ู’ุฃูŽุฌู’ูŠูŽุงู„ู.",
    auto_tashkeel=False,
)

seed=42:

seed=17 (different prosody):


7. Flow & rhythm test: connected speech

Tests natural sandhi, liaison, and intonation across a multi-clause sentence:

ุฅูู†ูŽู‘ ู†ูุธูŽุงู…ูŽ ุจูŽูŠูŽุงู†ูุณููŠู†ู’ุซ ูŠูŽู‡ู’ุฏููู ุฅูู„ูŽู‰ ุชูŽู‚ู’ุฏููŠู…ู ุชูŽุฌู’ุฑูุจูŽุฉู ุตูŽูˆู’ุชููŠูŽู‘ุฉู ููŽุฑููŠุฏูŽุฉูุŒ ุชูŽุฌู’ู…ูŽุนู ุจูŽูŠู’ู†ูŽ ุฏูู‚ูŽู‘ุฉู ุงู„ู†ูู‘ุทู’ู‚ู ูˆูŽุฌูŽู…ูŽุงู„ู ุงู„ู’ุฃูŽุฏูŽุงุกู.

BayanSynth aims to deliver a unique voice experience that combines precise pronunciation with beauty of delivery.

audio = tts.synthesize(
    "ุฅูู†ูŽู‘ ู†ูุธูŽุงู…ูŽ ุจูŽูŠูŽุงู†ูุณููŠู†ู’ุซ ูŠูŽู‡ู’ุฏููู ุฅูู„ูŽู‰ ุชูŽู‚ู’ุฏููŠู…ู ุชูŽุฌู’ุฑูุจูŽุฉู ุตูŽูˆู’ุชููŠูŽู‘ุฉู ููŽุฑููŠุฏูŽุฉูุŒ "
    "ุชูŽุฌู’ู…ูŽุนู ุจูŽูŠู’ู†ูŽ ุฏูู‚ูŽู‘ุฉู ุงู„ู†ูู‘ุทู’ู‚ู ูˆูŽุฌูŽู…ูŽุงู„ู ุงู„ู’ุฃูŽุฏูŽุงุกู.",
    auto_tashkeel=False,
)

seed=42:

seed=99 (different prosody):


8. Challenge: tashkeel disambiguation

All five ุน-rooted words differ only by their diacritics; correct rendering proves the model reads harakat accurately:

ุนูŽู„ูู…ูŽ ุงู„ู’ุนูŽุงู„ูู…ู ุฃูŽู†ูŽู‘ ุงู„ู’ุนูŽู„ูŽู…ูŽ ูŠูŽุนู’ู„ููˆ ุจูุงู„ู’ุนูู„ู’ู…ูุŒ ููŽุงุณู’ุชูŽุนู’ู„ูŽู…ูŽ ุนูŽู†ู’ ุนูู„ููˆู…ู ุงู„ู’ุฃูŽูˆูŽู‘ู„ููŠู†ูŽ.

The scholar knew that the flag rises with knowledge, so he inquired about the sciences of the ancients.

audio = tts.synthesize(
    "ุนูŽู„ูู…ูŽ ุงู„ู’ุนูŽุงู„ูู…ู ุฃูŽู†ูŽู‘ ุงู„ู’ุนูŽู„ูŽู…ูŽ ูŠูŽุนู’ู„ููˆ ุจูุงู„ู’ุนูู„ู’ู…ูุŒ "
    "ููŽุงุณู’ุชูŽุนู’ู„ูŽู…ูŽ ุนูŽู†ู’ ุนูู„ููˆู…ู ุงู„ู’ุฃูŽูˆูŽู‘ู„ููŠู†ูŽ.",
    auto_tashkeel=False,
)


9. Instruct prompt: warm newsreader style

Pass a free-text style directive alongside the synthesis text to steer the speaker's tone, register, or delivery:

ู…ูŽุฑู’ุญูŽุจุงู‹ ุจููƒูู…ู’. ู‡ูŽุฐูŽุง ู…ูุซูŽุงู„ูŒ ุนูŽู„ูŽู‰ ุงุณู’ุชูุฎู’ุฏูŽุงู…ู ุงู„ุชูŽู‘ูˆู’ุฌููŠู‡ู ู„ูุถูŽุจู’ุทู ุฃูุณู’ู„ููˆุจู ุงู„ุตูŽู‘ูˆู’ุชู.

Welcome. This is an example of using an instruct prompt to control voice style.

audio = tts.synthesize(
    "ู…ูŽุฑู’ุญูŽุจุงู‹ ุจููƒูู…ู’. ู‡ูŽุฐูŽุง ู…ูุซูŽุงู„ูŒ ุนูŽู„ูŽู‰ ุงุณู’ุชูุฎู’ุฏูŽุงู…ู ุงู„ุชูŽู‘ูˆู’ุฌููŠู‡ู ู„ูุถูŽุจู’ุทู ุฃูุณู’ู„ููˆุจู ุงู„ุตูŽู‘ูˆู’ุชู.",
    instruct="Speak in a warm, clear newsreader style with careful diction.",
    auto_tashkeel=False,
    seed=42,
)


Quick Start

1. Clone and install

git clone https://github.com/Ramendan/BayanSynthTTS
cd BayanSynthTTS
python -m venv .venv
.venv\Scripts\activate         # Windows
# source .venv/bin/activate   # Linux / macOS
pip install -r requirements.txt
pip install -e .               # installs bayansynthtts + bundled packages into the venv

The CosyVoice3 inference engine and Matcha-TTS decoder are bundled directly in this repo. No external private repos required.

Example voices: two reference clips (default.wav and muffled-talking.wav) live in voices/. Drop additional 5-15 s recordings there and they automatically appear in the CLI/UI dropdown.

2. Download models

python scripts/setup_models.py

This downloads everything automatically:

  • CosyVoice3 base weights (~2 GB) from Hugging Face โ†’ pretrained_models/CosyVoice3/
  • Arabic LoRA checkpoint from Hugging Face โ†’ checkpoints/llm/epoch_28_whole.pt
  • Verifies the checkpoint SHA-256

On Windows you can also double-click scripts\setup_models.bat.

3. Run

Web UI:

scripts\run_ui.bat            # Windows GUI launcher
python bayansynthtts/app.py   # Cross-platform (run from inside BayanSynthTTS/)

Files in this repo

File Description
epoch_28_whole.pt LoRA weights (LLM, 629 keys) โ€” main checkpoint
samples/*.wav Pre-generated audio demos

Swapping the LoRA Checkpoint

Via conf/models.yaml (recommended, no code changes)

llm_lora:
  enabled: true
  checkpoint: "checkpoints/llm/my_new_epoch.pt"   # โ† change this line only

Via Python constructor (for A/B testing at runtime)

tts = BayanSynthTTS(llm_checkpoint="checkpoints/llm/epoch_40.pt")

Via CLI flag

bayansynthtts "ู…ูŽุฑู’ุญูŽุจุงู‹" --llm checkpoints/llm/epoch_40.pt

Adding Your Own Voices

Drop any 5-15 second Arabic clip into voices/. Supported formats: WAV, MP3, FLAC, OGG, M4A. Non-WAV files are auto-converted at runtime.

from bayansynthtts import BayanSynthTTS
tts = BayanSynthTTS()
print(tts.list_voices())  # e.g. ['default.wav', 'muffled-talking.wav', 'my_voice.wav']
bayansynthtts "ู…ุฑุญุจุง" --voice voices/my_voice.wav

CLI Reference

bayansynthtts "ู…ูŽุฑู’ุญูŽุจุงู‹ ุจููƒูู…ู’"                       # basic synthesis โ†’ output.wav
bayansynthtts "ู…ูŽุฑู’ุญูŽุจุงู‹" -o hello.wav                  # custom output path
bayansynthtts "ู…ูŽุฑู’ุญูŽุจุงู‹" --voice voices/speaker2.wav   # use specific voice
bayansynthtts "ู…ูŽุฑู’ุญูŽุจุงู‹" --llm checkpoints/llm/new.pt  # override LLM LoRA
bayansynthtts "ู…ูŽุฑู’ุญูŽุจุงู‹" --speed 0.85                  # slower speech
bayansynthtts "ู…ูŽุฑู’ุญูŽุจุงู‹" --no-tashkeel                 # skip auto-diacritize
bayansynthtts "ู…ูŽุฑู’ุญูŽุจุงู‹" --seed 123                    # reproducible output
bayansynthtts --help

API Reference

BayanSynthTTS

Argument Type Default Description
model_dir str from YAML CosyVoice3 weights directory
llm_checkpoint str from YAML LLM LoRA .pt path
ref_audio str from YAML Default reference voice path
instruct str from YAML Instruct prompt text
config_path str conf/models.yaml Custom config file path

synthesize(text, *, ...)

Argument Type Default Description
text str required Arabic text (plain or diacritized)
ref_audio str default voice Voice clone source (any format)
instruct str from config Instruct prompt override
speed float 1.0 Speed multiplier (0.5-2.0)
stream bool False Yield chunks vs return full array
seed int None Random seed for reproducibility
auto_tashkeel bool True Auto-diacritize input text

Tashkeel utilities

from bayansynthtts import auto_diacritize, has_harakat, strip_harakat, list_available_backends

auto_diacritize("ู…ุฑุญุจุง ุจูƒู…")          # โ†’ "ู…ูŽุฑู’ุญูŽุจุงู‹ ุจููƒูู…ู’"
has_harakat("ู…ูŽุฑู’ุญูŽุจุงู‹")              # โ†’ True
strip_harakat("ู…ูŽุฑู’ุญูŽุจุงู‹")            # โ†’ "ู…ุฑุญุจุง"
list_available_backends()              # โ†’ ['mishkal']  (or ['tashkeel', 'mishkal'])

Troubleshooting

Problem Solution
No module named 'cosyvoice' Run pip install -e . from inside BayanSynthTTS/
No LLM checkpoint found Run python scripts/setup_models.py
mishkal not found pip install mishkal
No audio generated Check console for the specific mode that failed; verify voices/default.wav exists
MP3/M4A upload fails Install ffmpeg: winget install ffmpeg (Windows) or sudo apt install ffmpeg (Linux)

License

Apache 2.0.

The underlying CosyVoice3 model is subject to its own license. LoRA checkpoints trained on Common Voice Arabic data are released under CC-BY 4.0.

Downloads last month
298
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support