Kanade: Compact Linguistically Rich Speech Tokens for Spoken Language Models
Kanade is a speech tokenizer that encodes speech into compact content tokens and global embeddings and decodes them back to mel spectrograms.
Updates
- 2026-01-09: Released
kanade-25hz-cleanmodel trained on LibriTTS-R with HiFT vocoder for better audio quality. LibriTTS-R is a restored version of LibriTTS removing noise, so the model trained on it can produce cleaner synthesis. Because of that, however, this version can no longer faithfully reflect the recording environment such as background noise and microphone characteristics. Also, the vocoder is changed to the HiFT model used in CosyVoice 2 for better quality. The content encoder part remains the same as the previouskanade-25hzmodel. We made tiny code change to support different vocoders during inference (specificallyload_vocoder). Please refer to the updated usage section below.
Models
| Model | Token Rate | Vocab Size | Bit Rate | Dataset | SSL Encoder | Vocoder | Parameters |
|---|---|---|---|---|---|---|---|
kanade-12.5hz |
12.5 Hz | 12800 | 171 bps | LibriTTS | WavLM-base+ | Vocos 24kHz | 120M |
kanade-25hz |
25 Hz | 12800 | 341 bps | LibriTTS | WavLM-base+ | Vocos 24kHz | 118M |
kanade-25hz-clean |
25 Hz | 12800 | 341 bps | LibriTTS-R | WavLM-base+ | HiFT 24kHz | 142M |
Installation
For simple inference, you can install Kanade tokenizer to your virtual environment:
# In your own project's virtual environment
uv add git+https://github.com/frothywater/kanade-tokenizer
# or using pip
pip install git+https://github.com/frothywater/kanade-tokenizer
We use FlashAttention for efficient local window attention in our training. We recommend installing it following the instructions in their repository to get the best performance and the closest match to our setup. The model will fall back to regular PyTorch SDPA implementation if FlashAttention is not available. In this case, we cannot guarantee the same quality as reported in the paper.
If using uv, you can install FlashAttention like:uv pip install flash-attn --no-build-isolation. (Ensureninjais installed in your system or the build will be very slow.)
Usage
Example code to load the model from HuggingFace Hub and run inference:
from kanade_tokenizer import KanadeModel, load_audio, load_vocoder, vocode
# Load Kanade model
model = KanadeModel.from_pretrained("frothywater/kanade-12.5hz")
model = model.eval().cuda()
# Load vocoder
vocoder = load_vocoder(model.config.vocoder_name).cuda()
# Load audio (samples,)
audio = load_audio("path/to/audio.wav", sample_rate=model.config.sample_rate).cuda()
# Extract features
features = model.encode(waveform)
# Synthesize audio from extracted features
mel_spectrogram = model.decode(
content_token_indices=features.content_token_indices, # (seq_len,)
global_embedding=features.global_embedding, # (dim,)
) # (n_mels, T)
# Resynthesize waveform using vocoder
resynthesized_waveform = vocode(vocoder, mel_spectrogram.unsqueeze(0)) # (1, samples)
For details about voice conversion and how to train and fine-tune the model, please refer to our github repository.
- Downloads last month
- 64