LiveTranslate / README.md
fasdfsa's picture
init
bd95217
# LiveTrans
**English** | [中文](README_zh.md)
Real-time audio translation tool for Windows. Captures system audio via WASAPI loopback, runs speech recognition (ASR), translates through LLM APIs, and displays results in a transparent overlay window.
Perfect for watching foreign-language videos, livestreams, and meetings — no player modifications needed, works with any system audio.
![Python 3.10+](https://img.shields.io/badge/Python-3.10%2B-blue)
![Windows](https://img.shields.io/badge/Platform-Windows-0078d4)
![License](https://img.shields.io/badge/License-MIT-green)
## Features
- **Real-time translation**: System audio → ASR → LLM translation → subtitle overlay, fully automatic
- **Multiple ASR engines**: faster-whisper, FunASR SenseVoice (optimized for Japanese), FunASR Nano
- **Flexible translation backend**: Compatible with any OpenAI-format API (DeepSeek, Grok, Qwen, GPT, etc.)
- **Low-latency VAD**: 32ms audio chunks + Silero VAD with adaptive silence detection
- **Transparent overlay**: Always-on-top, click-through, draggable — doesn't interfere with your workflow
- **CUDA acceleration**: GPU-accelerated ASR inference
- **Automatic model management**: First-launch setup wizard, supports ModelScope / HuggingFace dual sources
- **Translation benchmark**: Built-in benchmark tool for comparing model performance
## Screenshots
**English → Chinese** (Twitch livestream)
![English to Chinese](screenshot/en-to-cn.png)
**Japanese → Chinese** (Japanese livestream)
![Japanese to Chinese](screenshot/jp-to-cn.png)
## Requirements
- **OS**: Windows 10/11
- **Python**: 3.10+
- **GPU** (recommended): NVIDIA GPU with CUDA 12.6 (for ASR acceleration)
- **Network**: Access to a translation API (DeepSeek, OpenAI, etc.)
## Installation
### 1. Clone the repository
```bash
git clone https://github.com/TheDeathDragon/LiveTranslate.git
cd LiveTranslate
```
### 2. Create a virtual environment
```bash
python -m venv .venv
.venv\Scripts\activate
```
### 3. Install PyTorch (with CUDA)
Choose the install command based on your CUDA version. See [PyTorch official site](https://pytorch.org/get-started/locally/):
```bash
# CUDA 12.6 (recommended)
pip install torch torchaudio --index-url https://download.pytorch.org/whl/cu126
# CPU only (no NVIDIA GPU)
pip install torch torchaudio --index-url https://download.pytorch.org/whl/cpu
```
### 4. Install remaining dependencies
```bash
pip install -r requirements.txt
pip install funasr --no-deps
```
> **Note**: FunASR is installed with `--no-deps` because its dependency `editdistance` requires a C++ compiler. The pure-Python alternative `editdistance-s` is included in `requirements.txt` as a drop-in replacement.
### 5. Launch
```bash
.venv\Scripts\python.exe main.py
```
Or double-click `start.bat`.
## First Launch
1. A **setup wizard** will appear on first launch — choose your model download source (ModelScope for China, HuggingFace for international) and model cache path
2. Silero VAD and SenseVoice ASR models will be downloaded automatically (~1GB)
3. The main UI appears once downloads complete
## Configuring the Translation API
Click **Settings** on the overlay → **Translation** tab:
| Parameter | Description |
|-----------|-------------|
| API Base | API endpoint, e.g. `https://api.deepseek.com/v1` |
| API Key | Your API key |
| Model | Model name, e.g. `deepseek-chat` |
| Proxy | `none` (direct) / `system` (system proxy) / custom proxy URL |
Works with any OpenAI-compatible API, including:
- [DeepSeek](https://platform.deepseek.com/)
- [xAI Grok](https://console.x.ai/)
- [Alibaba Qwen](https://dashscope.aliyuncs.com/)
- [OpenAI GPT](https://platform.openai.com/)
- Self-hosted [Ollama](https://ollama.ai/), [vLLM](https://github.com/vllm-project/vllm), etc.
## Usage
1. Play a video or livestream with foreign-language audio
2. Launch LiveTrans — the overlay appears automatically
3. Recognized text and translations are displayed in real time
### Overlay Controls
- **Pause/Resume**: Pause or resume translation
- **Clear**: Clear current subtitles
- **Click-through**: Mouse clicks pass through the subtitle window
- **Always on top**: Keep overlay above all windows
- **Auto-scroll**: Automatically scroll to the latest subtitle
- **Model selector**: Switch between configured translation models
- **Target language**: Change the translation target language
### Settings Panel
Open via the **Settings** button on the overlay or the system tray menu:
- **VAD/ASR**: ASR engine selection, VAD mode, sensitivity parameters
- **Translation**: API configuration, system prompt, multi-model management
- **Benchmark**: Translation speed and quality benchmarks
- **Cache**: Model cache path management
## Architecture
```
Audio (WASAPI 32ms) → VAD (Silero) → ASR (Whisper/SenseVoice/Nano) → LLM Translation → Overlay
```
```
main.py Entry point & pipeline orchestration
├── audio_capture.py WASAPI loopback audio capture
├── vad_processor.py Silero VAD speech detection
├── asr_engine.py faster-whisper ASR backend
├── asr_sensevoice.py FunASR SenseVoice backend
├── asr_funasr_nano.py FunASR Nano backend
├── translator.py OpenAI-compatible translation client
├── model_manager.py Model detection, download & cache management
├── subtitle_overlay.py PyQt6 transparent overlay window
├── control_panel.py Settings panel UI
├── dialogs.py Setup wizard & model download dialogs
├── log_window.py Real-time log viewer
├── benchmark.py Translation benchmark
└── config.yaml Default configuration
```
## Known Limitations
- Windows only (depends on WASAPI loopback)
- ASR model first load takes a few seconds (GPU) to tens of seconds (CPU)
- Translation quality depends on the LLM API used
- Recognition degrades in noisy environments or with overlapping speakers
## License
[MIT License](LICENSE)