sub200 / README.md
Revrse's picture
Upload 5 files
7875858 verified
---
title: sub200
emoji: ๐ŸŽ™๏ธ
colorFrom: purple
colorTo: blue
sdk: gradio
sdk_version: 4.44.0
app_file: app.py
pinned: false
license: mit
hardware: zero-gpu-h200
---
# sub200 - Ultra Low Latency TTS Hosting
sub200 allows you to host different open source TTS (Text-to-Speech) engines with ultra low latency.
## Features
- ๐Ÿš€ **Ultra Low Latency** - Optimized for real-time speech synthesis
- ๐ŸŽฏ **Multiple Engines** - Support for Piper, Coqui TTS, Edge TTS, eSpeak, gTTS, and pyttsx3
- ๐ŸŒ **Web UI** - Simple, modern Gradio interface
- โšก **Fast** - Built with Gradio for high performance
- ๐ŸŽฎ **GPU Support** - Optimized for GPU acceleration with Coqui TTS (H200 dynamic allocation)
## Available TTS Engines
1. **Piper TTS** - Ultra low latency, offline
2. **Coqui TTS** - High quality neural TTS (GPU accelerated)
3. **Edge TTS** - Microsoft Edge TTS (free, online)
4. **eSpeak** - Fast, lightweight, offline
5. **Google TTS (gTTS)** - Online, requires internet
6. **pyttsx3** - Offline, uses system voices
## Usage
1. Enter your text in the text box
2. Select a TTS engine from the dropdown
3. Adjust speed if needed (0.5x to 2.0x)
4. Click "Generate Speech"
5. Audio will auto-play when ready!
## GPU Support
This Space is configured for **zero GPU** (H200 dynamic allocation):
- GPU is allocated automatically when Coqui TTS is used
- No GPU needed for other engines (Piper, Edge TTS, eSpeak, etc.)
- Efficient resource usage with dynamic allocation
## Model Files
### Piper Models
- Models are downloaded automatically at runtime if not present
- Or you can include them in the repository (they're ~60MB each)
### Coqui Models
- Models are downloaded automatically on first use
- They're cached in the Space's storage
## Local Development
```bash
# Install dependencies
pip install -r requirements.txt
# Run server
python app.py
```
Then open http://localhost:7860
## Performance Tips
1. **Use GPU** - Coqui TTS benefits significantly from GPU acceleration
2. **Choose Right Engine**:
- **Piper** - Fastest, lowest latency, offline
- **Edge TTS** - Best quality, requires internet
- **Coqui** - High quality, GPU accelerated
- **eSpeak** - Fastest, basic quality, offline
## Troubleshooting
- **No audio generated**: Check engine status in the accordion
- **GPU not working**: Ensure zero GPU is enabled in Space settings
- **Model download fails**: Check internet connection for online engines