sub200 / README.md
Revrse's picture
Upload 5 files
7875858 verified

A newer version of the Gradio SDK is available: 6.5.1

Upgrade
metadata
title: sub200
emoji: ๐ŸŽ™๏ธ
colorFrom: purple
colorTo: blue
sdk: gradio
sdk_version: 4.44.0
app_file: app.py
pinned: false
license: mit
hardware: zero-gpu-h200

sub200 - Ultra Low Latency TTS Hosting

sub200 allows you to host different open source TTS (Text-to-Speech) engines with ultra low latency.

Features

  • ๐Ÿš€ Ultra Low Latency - Optimized for real-time speech synthesis
  • ๐ŸŽฏ Multiple Engines - Support for Piper, Coqui TTS, Edge TTS, eSpeak, gTTS, and pyttsx3
  • ๐ŸŒ Web UI - Simple, modern Gradio interface
  • โšก Fast - Built with Gradio for high performance
  • ๐ŸŽฎ GPU Support - Optimized for GPU acceleration with Coqui TTS (H200 dynamic allocation)

Available TTS Engines

  1. Piper TTS - Ultra low latency, offline
  2. Coqui TTS - High quality neural TTS (GPU accelerated)
  3. Edge TTS - Microsoft Edge TTS (free, online)
  4. eSpeak - Fast, lightweight, offline
  5. Google TTS (gTTS) - Online, requires internet
  6. pyttsx3 - Offline, uses system voices

Usage

  1. Enter your text in the text box
  2. Select a TTS engine from the dropdown
  3. Adjust speed if needed (0.5x to 2.0x)
  4. Click "Generate Speech"
  5. Audio will auto-play when ready!

GPU Support

This Space is configured for zero GPU (H200 dynamic allocation):

  • GPU is allocated automatically when Coqui TTS is used
  • No GPU needed for other engines (Piper, Edge TTS, eSpeak, etc.)
  • Efficient resource usage with dynamic allocation

Model Files

Piper Models

  • Models are downloaded automatically at runtime if not present
  • Or you can include them in the repository (they're ~60MB each)

Coqui Models

  • Models are downloaded automatically on first use
  • They're cached in the Space's storage

Local Development

# Install dependencies
pip install -r requirements.txt

# Run server
python app.py

Then open http://localhost:7860

Performance Tips

  1. Use GPU - Coqui TTS benefits significantly from GPU acceleration
  2. Choose Right Engine:
    • Piper - Fastest, lowest latency, offline
    • Edge TTS - Best quality, requires internet
    • Coqui - High quality, GPU accelerated
    • eSpeak - Fastest, basic quality, offline

Troubleshooting

  • No audio generated: Check engine status in the accordion
  • GPU not working: Ensure zero GPU is enabled in Space settings
  • Model download fails: Check internet connection for online engines