| from ..filterbanks import make_enc_dec
|
| from ..masknn import DPTransformer
|
| from .base_models import BaseEncoderMaskerDecoder
|
|
|
|
|
| class DPTNet(BaseEncoderMaskerDecoder):
|
| """DPTNet separation model, as described in [1].
|
|
|
| Args:
|
| n_src (int): Number of masks to estimate.
|
| out_chan (int or None): Number of bins in the estimated masks.
|
| Defaults to `in_chan`.
|
| bn_chan (int): Number of channels after the bottleneck.
|
| Defaults to 128.
|
| hid_size (int): Number of neurons in the RNNs cell state.
|
| Defaults to 128.
|
| chunk_size (int): window size of overlap and add processing.
|
| Defaults to 100.
|
| hop_size (int or None): hop size (stride) of overlap and add processing.
|
| Default to `chunk_size // 2` (50% overlap).
|
| n_repeats (int): Number of repeats. Defaults to 6.
|
| norm_type (str, optional): Type of normalization to use. To choose from
|
|
|
| - ``'gLN'``: global Layernorm
|
| - ``'cLN'``: channelwise Layernorm
|
| mask_act (str, optional): Which non-linear function to generate mask.
|
| bidirectional (bool, optional): True for bidirectional Inter-Chunk RNN
|
| (Intra-Chunk is always bidirectional).
|
| rnn_type (str, optional): Type of RNN used. Choose between ``'RNN'``,
|
| ``'LSTM'`` and ``'GRU'``.
|
| num_layers (int, optional): Number of layers in each RNN.
|
| dropout (float, optional): Dropout ratio, must be in [0,1].
|
| in_chan (int, optional): Number of input channels, should be equal to
|
| n_filters.
|
| fb_name (str, className): Filterbank family from which to make encoder
|
| and decoder. To choose among [``'free'``, ``'analytic_free'``,
|
| ``'param_sinc'``, ``'stft'``].
|
| n_filters (int): Number of filters / Input dimension of the masker net.
|
| kernel_size (int): Length of the filters.
|
| stride (int, optional): Stride of the convolution.
|
| If None (default), set to ``kernel_size // 2``.
|
| **fb_kwargs (dict): Additional kwards to pass to the filterbank
|
| creation.
|
|
|
| References:
|
| [1]: Jingjing Chen et al. "Dual-Path Transformer Network: Direct
|
| Context-Aware Modeling for End-to-End Monaural Speech Separation"
|
| Interspeech 2020.
|
| """
|
|
|
| def __init__(
|
| self,
|
| n_src,
|
| ff_hid=256,
|
| chunk_size=100,
|
| hop_size=None,
|
| n_repeats=6,
|
| norm_type="gLN",
|
| ff_activation="relu",
|
| encoder_activation="relu",
|
| mask_act="relu",
|
| bidirectional=True,
|
| dropout=0,
|
| in_chan=None,
|
| fb_name="free",
|
| kernel_size=16,
|
| n_filters=64,
|
| stride=8,
|
| **fb_kwargs,
|
| ):
|
| encoder, decoder = make_enc_dec(
|
| fb_name, kernel_size=kernel_size, n_filters=n_filters, stride=stride, **fb_kwargs
|
| )
|
| n_feats = encoder.n_feats_out
|
| if in_chan is not None:
|
| assert in_chan == n_feats, (
|
| "Number of filterbank output channels"
|
| " and number of input channels should "
|
| "be the same. Received "
|
| f"{n_feats} and {in_chan}"
|
| )
|
|
|
| masker = DPTransformer(
|
| n_feats,
|
| n_src,
|
| ff_hid=ff_hid,
|
| ff_activation=ff_activation,
|
| chunk_size=chunk_size,
|
| hop_size=hop_size,
|
| n_repeats=n_repeats,
|
| norm_type=norm_type,
|
| mask_act=mask_act,
|
| bidirectional=bidirectional,
|
| dropout=dropout,
|
| )
|
| super().__init__(encoder, masker, decoder, encoder_activation=encoder_activation)
|
|
|