Update README.md
Browse files
README.md
CHANGED
|
@@ -1,11 +1,24 @@
|
|
| 1 |
-
---
|
| 2 |
-
license: mpl-2.0
|
| 3 |
-
|
| 4 |
-
|
| 5 |
-
|
| 6 |
-
|
| 7 |
-
|
| 8 |
-
-
|
| 9 |
-
|
| 10 |
-
|
| 11 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: mpl-2.0
|
| 3 |
+
datasets:
|
| 4 |
+
- CommonVoice
|
| 5 |
+
- LibriSpeech
|
| 6 |
+
- LibriTTS
|
| 7 |
+
- VCTK
|
| 8 |
+
- VoxCeleb
|
| 9 |
+
tags:
|
| 10 |
+
- speaker-encoding
|
| 11 |
+
- speaker-embedding
|
| 12 |
+
- speaker-verification
|
| 13 |
+
- speaker
|
| 14 |
+
- speech
|
| 15 |
+
library_name: PyTorch
|
| 16 |
+
---
|
| 17 |
+
|
| 18 |
+
This repository contains released Mozilla SpeakerEncoder models.
|
| 19 |
+
|
| 20 |
+
Due to original codebase migration and SpeakerEncoder depracation, I decided to extract inference code into my own repository. For ease of integration, I also release relevant models:
|
| 21 |
+
- `model_mozilla_128.pth.tar` - which is [Speaker-Encoder-iter25k](https://drive.google.com/open?id=1CwXxwm0qKk2lcPL_4EQDirD8XMyvHIqR) that was trained on LibriSpeech,
|
| 22 |
+
- `model_mozilla_256.pth.tar` - which is [Speaker-Encoder](https://drive.google.com/drive/folders/15oeBYf6Qn1edONkVLXe82MzdIi3O_9m3?usp=sharing) by [@mueller91](https://github.com/mueller91) that was trained on LibriTTS, VCTK, VoxCeleb and CommonVoice.
|
| 23 |
+
|
| 24 |
+
Original SpeakerEncoder implementation is available [here](https://github.com/coqui-ai/TTS/tree/f9001a4bdd8a169da8b4eefd481461f370def8f9/TTS/speaker_encoder) and originally released models are available [here](https://github.com/mozilla/TTS/wiki/Released-Models). If you are interested in the current version of Coqui.ai TTS go [here](https://github.com/coqui-ai/TTS).
|