ATTS1HG1: High-Performance GGML Implementation of XTTS-v2

ATTS1HG1 is a high-speed, native C++ implementation of the Coqui XTTS-v2 model, utilizing the GGML tensor library. It features a custom integrated HiFiGAN vocoder optimized for Vulkan and CPU inference.

Source Code & GUI Base Model Backend
GitHub: ATTS1HG1 Coqui XTTS-v2 GGML / Vulkan

๐Ÿš€ Key Features

  • Blazing Fast: Generates audio in < 0.5s on consumer GPUs (RTX 3090) and ~1.0s on CPU.
  • Vulkan Support: Fully optimized HiFiGAN vocoder running on Vulkan (compatible with NVIDIA, AMD, Intel iGPUs).
  • Lightweight: Native C++ application, no heavy Python dependencies (PyTorch/TensorFlow not required at runtime).
  • Multi-Language: Supports 17 languages.
  • Voice : Supports 58 speaker (similar to XTTS).

๐ŸŒ Supported Languages

The model supports the following 17 languages:

Code Language Native Name
en English English
es Spanish Espaรฑol
fr French Franรงais
de German Deutsch
it Italian Italiano
pt Portuguese Portuguรชs
pl Polish Polski
tr Turkish Tรผrkรงe
ru Russian ะ ัƒััะบะธะน
nl Dutch Nederlands
cs Czech ฤŒeลกtina
ar Arabic ุงู„ุนุฑุจูŠุฉ
zh Chinese ไธญๆ–‡
ja Japanese ๆ—ฅๆœฌ่ชž
hu Hungarian Magyar
ko Korean ํ•œ๊ตญ์–ด
hi Hindi เคนเคฟเคจเฅเคฆเฅ€

โšก Performance

Benchmarks based on standard text generation ("Bonjour le monde") using the C++ client:

Device Backend Latency (Total) Note
NVIDIA RTX 3090 Vulkan ~0.47s ๐Ÿš€ Recommended
Intel iGPU Vulkan ~1.40s Good for laptops
CPU (Ryzen/Intel) CPU (AVX2) ~1.02s Solid fallback
NVIDIA RTX 3090 CUDA ~1.45s Slower on HiFiGAN due to kernel overhead

Note: The Vulkan backend is significantly faster for the HiFiGAN part of the pipeline compared to CUDA due to optimized command buffers and reduced kernel launch overhead for small convolutions.

๐Ÿ› ๏ธ Usage

This repository contains the converted .bin / .gguf weights required by the ATTS1HG1 software.

  1. Download the model files from this repository.
  2. Clone and compile the software from GitHub:
    git clone [https://github.com/abbndz/ATTS1HG1](https://github.com/abbndz/ATTS1HG1)
    
  3. Load the model in the GUI or CLI and select Vulkan for best performance.

๐Ÿ“œ License

This project uses the weights from Coqui XTTS-v2, which is licensed under the Coqui Public Model License (CPML).

  • Non-commercial use: You can use this model for personal, educational, and non-commercial projects.
  • Commercial use: Requires a license from Coqui (check their repository for details).

The C++ code (inference engine) is available under the MIT License (see GitHub).


Credits: Based on the excellent work by Coqui.ai and the GGML library by ggerganov.

Downloads last month
112
GGUF
Model size
0.4B params
Architecture
atts1hg
Hardware compatibility
Log In to add your hardware

4-bit

6-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for ABBNDZ/ATTS1HG1

Base model

coqui/XTTS-v2
Quantized
(1)
this model