| """ |
| V6 Inference — encoder-decoder TTS with MioCodec + speaker cloning |
| =================================================================== |
| 1. Encode text with encoder (bidirectional, once) |
| 2. Autoregressively decode audio tokens with decoder + speaker embedding |
| 3. Decode tokens with MioCodec using global_embedding |
| """ |
|
|
| import torch |
| import argparse |
| import time |
| from pathlib import Path |
| from config import ( |
| AUDIO_OFFSET, NUM_AUDIO_TOKENS, END_OF_SPEECH_TOKEN_ID, |
| START_OF_SPEECH_TOKEN_ID, CODEC_SAMPLE_RATE, CODEC_FRAME_RATE, |
| ) |
| from tokenizer import TTSTokenizer |
| from codec import CodecV6 |
| from model import load_for_inference |
|
|
|
|
| @torch.no_grad() |
| def generate(model, tokenizer, text, speaker_emb, |
| max_new_tokens=512, temperature=0.7, top_k=250, |
| top_p=0.95, rep_penalty=1.1, device="cuda"): |
| """ |
| Generate audio tokens from text. |
| |
| Args: |
| model: TTSEncoderDecoder |
| tokenizer: TTSTokenizer |
| text: input text string |
| speaker_emb: [128] MioCodec global_embedding |
| max_new_tokens: max decoder steps |
| temperature: sampling temperature |
| top_k: top-k filtering |
| top_p: nucleus sampling threshold |
| rep_penalty: repetition penalty on recent tokens |
| device: cuda/cpu |
| |
| Returns: |
| torch.Tensor of MioCodec codes [num_frames], or None |
| """ |
| |
| enc_ids = tokenizer.build_encoder_input(text).unsqueeze(0).to(device) |
| enc_mask = torch.ones_like(enc_ids) |
|
|
| enc_out = model.encode(enc_ids, enc_mask) |
|
|
| |
| spk = speaker_emb.unsqueeze(0).to(device) |
|
|
| |
| dec_ids = torch.tensor([[START_OF_SPEECH_TOKEN_ID]], device=device) |
| past = None |
| generated_tokens = [] |
|
|
| for step in range(max_new_tokens): |
| inp = dec_ids[:, -1:] if past is not None else dec_ids |
|
|
| |
| |
| |
| dec_out = model.decoder( |
| input_ids=inp, |
| encoder_output=enc_out, |
| encoder_mask=enc_mask, |
| speaker_emb=spk, |
| past_key_values=past, |
| use_cache=True, |
| ) |
| past = dec_out["past_key_values"] |
| logits = dec_out["logits"][:, -1, :] |
|
|
| |
| mask = torch.full_like(logits, float("-inf")) |
| mask[:, AUDIO_OFFSET:AUDIO_OFFSET + NUM_AUDIO_TOKENS] = 0 |
| mask[:, END_OF_SPEECH_TOKEN_ID] = 0 |
| logits = logits + mask |
|
|
| |
| if rep_penalty != 1.0 and generated_tokens: |
| recent = set(generated_tokens[-100:]) |
| for tid in recent: |
| if AUDIO_OFFSET <= tid < AUDIO_OFFSET + NUM_AUDIO_TOKENS: |
| logits[:, tid] /= rep_penalty |
|
|
| logits = logits / temperature |
|
|
| |
| if top_k > 0: |
| kth = torch.topk(logits, min(top_k, logits.shape[-1])).values[:, -1:] |
| logits[logits < kth] = float("-inf") |
|
|
| |
| if top_p < 1.0: |
| sorted_l, sorted_i = torch.sort(logits, descending=True) |
| cum = torch.cumsum(torch.softmax(sorted_l, -1), -1) |
| remove = cum > top_p |
| remove[:, 1:] = remove[:, :-1].clone() |
| remove[:, 0] = False |
| logits[remove.scatter(1, sorted_i, remove)] = float("-inf") |
|
|
| next_tok = torch.multinomial(torch.softmax(logits, -1), 1) |
| tok_id = next_tok.item() |
|
|
| if tok_id == END_OF_SPEECH_TOKEN_ID: |
| break |
|
|
| generated_tokens.append(tok_id) |
| dec_ids = torch.cat([dec_ids, next_tok], dim=-1) |
|
|
| if not generated_tokens: |
| return None |
|
|
| result = torch.tensor(generated_tokens, dtype=torch.long) |
| audio_mask = (result >= AUDIO_OFFSET) & (result < AUDIO_OFFSET + NUM_AUDIO_TOKENS) |
| return result[audio_mask] - AUDIO_OFFSET |
|
|
|
|
| def synthesize(checkpoint, text, output="output.wav", |
| speaker_wav=None, speaker_emb_path=None, |
| temperature=0.7, top_k=250, top_p=0.95, |
| rep_penalty=1.1, max_tokens=512, device="cuda"): |
| """ |
| Full TTS pipeline: text → audio file. |
| |
| Speaker can be provided as: |
| 1. speaker_wav: path to reference audio (will encode with MioCodec) |
| 2. speaker_emb_path: path to saved .pt embedding |
| """ |
| print(f"'{text[:80]}' | T={temperature}") |
| model = load_for_inference(checkpoint, device=device) |
| tokenizer = TTSTokenizer() |
| codec = CodecV6(device=device) |
|
|
| |
| if speaker_emb_path: |
| import numpy as np |
| if speaker_emb_path.endswith('.npy'): |
| speaker_emb = torch.from_numpy(np.load(speaker_emb_path)).to(device) |
| else: |
| speaker_emb = torch.load(speaker_emb_path, map_location=device, weights_only=False) |
| if isinstance(speaker_emb, dict): |
| speaker_emb = speaker_emb.get("global_embedding", |
| speaker_emb.get("embedding")) |
| if speaker_emb.dim() > 1: |
| speaker_emb = speaker_emb.squeeze() |
| print(f"Speaker from preset: {speaker_emb.shape}") |
| elif speaker_wav: |
| result = codec.encode(speaker_wav) |
| speaker_emb = result['global_embedding'].to(device) |
| print(f"Speaker from wav: {speaker_wav}") |
| else: |
| raise ValueError("Provide speaker_wav or speaker_emb_path") |
|
|
| |
| t0 = time.time() |
| codes = generate(model, tokenizer, text, speaker_emb, max_tokens, |
| temperature, top_k, top_p, rep_penalty, device) |
| gen_time = time.time() - t0 |
|
|
| if codes is None or len(codes) == 0: |
| print("No audio generated!") |
| return |
|
|
| audio_dur = len(codes) / CODEC_FRAME_RATE |
| rtf = gen_time / audio_dur if audio_dur > 0 else float('inf') |
| print(f"{len(codes)} tokens ({audio_dur:.1f}s audio, {gen_time:.2f}s gen, RTF={rtf:.3f})") |
|
|
| |
| wav = codec.tokens_to_wav(codes, speaker_emb, output) |
| print(f"Saved: {output} ({len(wav)/CODEC_SAMPLE_RATE:.2f}s)") |
| return wav |
|
|
|
|
| def main(): |
| p = argparse.ArgumentParser(description="V6 TTS Inference") |
| p.add_argument("--checkpoint", required=True) |
| p.add_argument("--text", required=True) |
| p.add_argument("--output", default="output.wav") |
| p.add_argument("--speaker-wav", help="Reference audio for voice cloning") |
| p.add_argument("--speaker-emb", help="Path to saved speaker embedding .pt") |
| p.add_argument("--temperature", type=float, default=0.7) |
| p.add_argument("--top-k", type=int, default=250) |
| p.add_argument("--top-p", type=float, default=0.95) |
| p.add_argument("--rep-penalty", type=float, default=1.1) |
| p.add_argument("--max-tokens", type=int, default=512) |
| a = p.parse_args() |
| synthesize(a.checkpoint, a.text, a.output, |
| speaker_wav=a.speaker_wav, |
| speaker_emb_path=a.speaker_emb, |
| temperature=a.temperature, top_k=a.top_k, |
| top_p=a.top_p, rep_penalty=a.rep_penalty, |
| max_tokens=a.max_tokens) |
|
|
| if __name__ == "__main__": |
| main() |
|
|