OVOS - Qwen3 TTS 0.6B Q8_0 (GGUF)

This model is an quantized gguf-format export of Qwen/Qwen3-TTS-12Hz-0.6B-Base for ease of use in edge devices and CPU-based inference environments. The original model is transformed into gguf with F16 tensors by the script convert_hf_to_gguf.py and then further quantized, if needed, using the tool quantize from the same repo.

Requirements

The requirements can be installed as

$ pip install git+https://github.com/femelo/py-qwen3-tts-cpp

Usage

from py_qwen3_tts_cpp.model import Qwen3TTSModel

# Initialize the model (it handles downloading from this repo)
model = Qwen3TTSModel(
    tts_model="qwen3-tts-0.6b-q8-0",
    n_threads=4
)

# Synthesize from text
result = model.synthesize("Hello, world!", language="en")
model.save_audio(result, "output.wav")

Refer to https://github.com/femelo/py-qwen3-tts-cpp for more details.

Licensing

The license is derived from the original model: Apache 2.0. For more details, please refer to Qwen/Qwen3-TTS-12Hz-0.6B-Base.

Downloads last month
246
GGUF
Model size
0.9B params
Architecture
qwen3-tts
Hardware compatibility
Log In to add your hardware

We're not able to determine the quantization variants.

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Collection including OpenVoiceOS/qwen3-tts-0.6b-q8-0