TTS - GGUF
Collection
Multilingual quantized TTS models in GGUF format • 4 items • Updated
This model is an quantized gguf-format export of Qwen/Qwen3-TTS-12Hz-0.6B-Base for ease of use in edge devices and CPU-based inference environments. The original model is transformed into gguf with F16 tensors by the script convert_hf_to_gguf.py and then further quantized, if needed, using the tool quantize from the same repo.
The requirements can be installed as
$ pip install git+https://github.com/femelo/py-qwen3-tts-cpp
from py_qwen3_tts_cpp.model import Qwen3TTSModel
# Initialize the model (it handles downloading from this repo)
model = Qwen3TTSModel(
tts_model="qwen3-tts-0.6b-q8-0",
n_threads=4
)
# Synthesize from text
result = model.synthesize("Hello, world!", language="en")
model.save_audio(result, "output.wav")
Refer to https://github.com/femelo/py-qwen3-tts-cpp for more details.
The license is derived from the original model: Apache 2.0. For more details, please refer to Qwen/Qwen3-TTS-12Hz-0.6B-Base.
We're not able to determine the quantization variants.