|
|
--- |
|
|
license: apache-2.0 |
|
|
tags: |
|
|
- Hibernates |
|
|
- HVC-Audio-Convert |
|
|
|
|
|
pipeline_tag: audio-to-audio |
|
|
--- |
|
|
|
|
|
# HVC-Audio-Convert Base Models |
|
|
|
|
|
## Overview |
|
|
These models serve as the foundational components for HVC-Audio-Convert (Soft-VC Voice Conversion), an advanced voice conversion framework that combines SoftVC feature extraction with the VITS (Conditional Variational Autoencoder with Adversarial Learning) architecture. |
|
|
|
|
|
## Key Features |
|
|
- High-quality voice conversion capabilities |
|
|
- Pre-trained on diverse vocal datasets |
|
|
- Supports cross-lingual voice conversion |
|
|
- Compatible with HVC-Audio-Convert v4.0 and newer |
|
|
|
|
|
## Technical Details |
|
|
- **Architecture**: Based on VITS (Conditional Variational Autoencoder) |
|
|
- **Feature Extraction**: Hibernates content encoder |
|
|
- **Training Data**: Curated multi-speaker datasets |
|
|
- **Model Format**: PyTorch checkpoints |
|
|
|
|
|
## Usage |
|
|
1. Download the desired base model |
|
|
2. Use with HVC-Audio-Convert framework |
|
|
3. Fine-tune on target voice data |
|
|
4. Perform voice conversion |
|
|
|
|
|
## Requirements |
|
|
- HVC-Audio-Convert framework |
|
|
- Python 3.8+ |
|
|
- PyTorch 1.13.0+ |
|
|
- CUDA compatible GPU (recommended) |
|
|
|
|
|
## License |
|
|
This project is licensed under the Apache License 2.0 - see the LICENSE file for details. |
|
|
|
|
|
## Citation |
|
|
If you use these models in your research, please cite: |
|
|
|