| import json |
| import logging |
| import math |
| from argparse import Namespace |
| from pathlib import Path |
| from typing import Dict, List, Optional |
|
|
| import torch |
| import torch.nn as nn |
| from torch import Tensor |
| from fairseq import utils |
| from fairseq.data import Dictionary |
| from fairseq.data.audio.data_cfg import MultitaskConfig, S2SDataConfig |
| from fairseq.data.audio.speech_to_speech_dataset import SpeechToSpeechDatasetCreator |
| from fairseq.data.audio.speech_to_text_dataset import ( |
| SpeechToTextDataset, |
| TextTargetMultitaskData, |
| ) |
| from fairseq.tasks import LegacyFairseqTask, register_task |
| from fairseq.tasks.speech_to_text import DummyMultiTask |
| from fairseq.tasks.text_to_speech import batch_mel_cepstral_distortion |
|
|
| logger = logging.getLogger(__name__) |
|
|
|
|
| class CTCSequenceGenerator(nn.Module): |
| def __init__(self, tgt_dict, models, use_incremental_states=False): |
| super().__init__() |
| self.pad = tgt_dict.pad() |
| self.eos = tgt_dict.eos() |
| self.unk = tgt_dict.unk() |
| self.models = models |
| self.tgt_dict = tgt_dict |
| self.use_incremental_states = use_incremental_states |
| self.incremental_states = None |
|
|
| def reset_incremental_states(self): |
| self.incremental_states = None |
|
|
| @torch.no_grad() |
| def generate(self, encoder_out, prefix=None, aux_task_name=None, **kwargs): |
| if self.use_incremental_states: |
| if self.incremental_states is None: |
| incremental_states = torch.jit.annotate( |
| List[Dict[str, Dict[str, Optional[Tensor]]]], |
| [ |
| torch.jit.annotate(Dict[str, Dict[str, Optional[Tensor]]], {}) |
| for i in range(1) |
| ], |
| ) |
| self.incremental_states = incremental_states |
| else: |
| incremental_states = self.incremental_states |
| else: |
| incremental_states = None |
|
|
| |
| model = self.models[0] |
| model.eval() |
|
|
| max_len = model.max_decoder_positions() |
| |
|
|
| incremental_state = {} |
| pred_out, attn, scores = [], [], [] |
|
|
| prev_output_tokens = None |
| decoder_name = f"{aux_task_name}_decoder" if aux_task_name else "decoder" |
| ctc_decoder = getattr(model, decoder_name) |
| ctc_out, ctc_extra = ctc_decoder( |
| None, |
| encoder_out=encoder_out, |
| incremental_state=( |
| incremental_states[0] if self.use_incremental_states else None |
| ), |
| **kwargs, |
| ) |
| lprobs = model.get_normalized_probs([ctc_out], log_probs=True) |
|
|
| |
| lprobs[:, :, self.pad] = -math.inf |
| lprobs[:, :, self.unk] = -math.inf |
|
|
| cur_pred_lprob, cur_pred_out = torch.max(lprobs, dim=2) |
| scores = cur_pred_lprob |
| pred_out = cur_pred_out |
|
|
| attn = ctc_extra["attn"][0] |
| alignment = None |
|
|
| def _ctc_postprocess(tokens): |
| _toks = tokens.int().tolist() |
| deduplicated_toks = [ |
| v for i, v in enumerate(_toks) if i == 0 or v != _toks[i - 1] |
| ] |
| hyp = [ |
| v |
| for v in deduplicated_toks |
| if (v != self.tgt_dict.blank_index) and (v != self.tgt_dict.pad_index) |
| ] |
| return torch.tensor(hyp) |
|
|
| if prefix is not None: |
| if self.use_incremental_states: |
| pred_out = torch.cat((prefix, pred_out), dim=1) |
| else: |
| pred_out = torch.cat((prefix, pred_out[:, prefix.size(1) :]), dim=1) |
|
|
| hypos = [ |
| [ |
| { |
| "tokens": _ctc_postprocess(pred_out[b]), |
| "org_tokens": pred_out[b], |
| "attn": None, |
| "alignment": None, |
| "positional_scores": scores[b], |
| "score": utils.item(scores[b].sum().data), |
| } |
| ] |
| for b in range(pred_out.size(0)) |
| ] |
|
|
| return hypos |
|
|