This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Uformer: A Unet based dilated complex & real dual-path conformer network for simultaneous speech enhancement and dereverberation

Abstract

Complex spectrum and magnitude are considered as two major features of speech enhancement and dereverberation. Traditional approaches always treat these two features separately, ignoring their underlying relationship. In this paper, we propose Uformer, a Unet based dilated complex & real dual-path conformer network in both complex and magnitude domain for simultaneous speech enhancement and dereverberation. We exploit time attention (TA) and dilated convolution (DC) to leverage local and global contextual information and frequency attention (FA) to model dimensional information. These three sub-modules contained in the proposed dilated complex & real dual-path conformer module effectively improve the speech enhancement and dereverberation performance. Furthermore, hybrid encoder and decoder are adopted to simultaneously model the complex spectrum and magnitude and promote the information interaction between two domains. Encoder decoder attention is also applied to enhance the interaction between encoder and decoder. Our experimental results outperform all SOTA time and complex domain models objectively and subjectively. Specifically, Uformer reaches 3.6032 DNSMOS on the blind test set of Interspeech 2021 DNS Challenge, which outperforms all top-performed models. We also carry out ablation experiments to tease apart all proposed sub-modules that are most important.

Index Terms—  speech enhancement and dereverberation, Uformer, dilated complex dual-path conformer, hybrid encoder and decoder, encoder decoder attention

1 Introduction

Owing to the success of deep learning, deep neural network (DNN) based speech enhancement and dereverberation have progressed dramatically. For a long time, DNN-based speech enhancement algorithms attempt to only enhance the noisy magnitude using ideal ratio mask (IRM) and keep the noisy phase when reconstructing speech waveform [1]. This operation can be attributed to the unclear structure of phase, which is considered difficult to estimate [2]. Then, literature has emerged that offers contradictory findings about the importance of phase from perception aspect [3]. Subsequently, complex ratio mask (CRM) was proposed by Williamson et al. [4], reconstructing speech perfectly by enhancing both real and imaginary components of the noisy speech simultaneously. Later, Tan et al. proposed a convolution recurrent network (CRN) with one encoder and two decoders for complex spectral mapping (CSM) to estimate the real and imaginary components of mixture speech simultaneously [5]. It is worth considering that CRM and CSM possess the full information (magnitude and phase) of speech signal so that they can achieve the best oracle speech enhancement performance in theory. Then, the study in [6] updated the CRN network by introducing complex convolution and LSTM, resulting in the deep complex convolution recurrent network (DCCRN), which ranked first in MOS evaluation in the Interspeech 2020 DNS challenge real-time-track. The study in [7] further optimized DCCRN by proposing DCCRN+, which enables the model with subband processing ability by learnable neural network filters. It formulated the decoder as a multi-task learning framework with an auxiliary task of a priori SNR estimation. Besides speech enhancement, dereverberation is also a challenging task since it is hard to pinpoint the direct path signal and differentiate it from its copies, especially when reverberation is strong and non-stationary noise is also present. Some previous works have already been done on simultaneous multi-channel enhancement and dereverberation [8, 9, 10]. However, it is still a relatively hard task for the single-channel scenario due to the lack of spatial information. In this scenario, the feature they can only use is single-channel magnitude or complex spectrum. Although some researches exploit magnitude and complex spectrum as two input features, no parameter interaction between them are conducted.

Updated from transformer [11], conformer [12] is an outstanding convolution-augmented model for speech recognition with strong contextual modeling ability. The original conformer model presents two important modules: self-attention module and convolution module to model the global and local information, respectively. These modules have also proven their effectiveness, especially for temporal modeling ability from speech separation task [13]. Other transformer-based models, e.g., DPT-FSNET [14] and TSTNN [15], adopted dual-path transformer, also showed stronger interpretability and temporal-dimensional modeling ability than single-path method. TeCANet [16] applied transformer to the frames within a context window thus estimating correlations between frames. To further achieve stronger contextual modeling ability, the combination of conformer and dual-path method is an instinctive idea.

Inspired by the considerations mentioned above, we propose Uformer, a Unet based dilated complex & real dual-path conformer network for simultaneous speech enhancement and dereverberation. In magnitude branch, we seek to construct the filtering system which only applies to the magnitude domain. Specifically in the magnitude branch, we seek to construct a filtering system which only applies to the magnitude domain. In this branch, most of the noise is expected to be effectively suppressed. By contrast, the complex domain branch is established as a decorating system to compensate for the possible loss of spectral details and phase mismatch. Two branches work collaboratively to facilitate the overall spectrum recovery. The overall architecture of Uformer is shown in Figure 1. Our contribution in this work is three-fold:

  • We apply dilated complex & real dual-path conformer on the bottle-neck feature between encoder and decoder. Within the proposed module, Time attention (TA) with a context window is applied to model the local time dependency while dilated convolution (DC) is used to model global time dependency. Frequency attention (FA) is used to model subband information.

  • We apply a hybrid encoder and decoder to model complex spectrum and magnitude simultaneously. The rationale is that there exists close relation between magnitude and phase in complex spectrum. Superb magnitude estimation can profit better recovery for phase and vice versa.

  • We utilize encoder decoder attention to estimate attention mask, instead of skip connection, to reveal the relevance between the corresponding hybrid encoder and decoder layers.

Our experimental results outperform all SOTA time domain and complex domain speech front-end models on both objective and subjective performance. Specifically, Uformer reaches 3.6032 DNSMOS on the blind test set of Interspeech 2021 DNS challenge, which surpass SDD-Net [17], the top-performed model in the challenge. We also carry out ablation study to tease apart all proposed sub-modules that are most important.

Refer to caption
Fig. 1: The overall architecture of Uformer.
Refer to caption
Fig. 2: The architecture of dilated complex dual-path conformer.

2 Proposed method

2.1 Problem Formulation

In the time domain, let 𝐬(t)\mathbf{s}(t), 𝐡(t)\mathbf{h}(t), 𝐧(t)\mathbf{n}(t) denote the anechoic speech, room impulsive reponse (RIR) and noise, respectively. The observed signal of microphone 𝐨(t)\mathbf{o}(t) can be denoted as:

𝐨(t)=𝐬(t)𝐡(t)+𝐧(t)=𝐬_𝐞(t)+𝐬_𝐥(t)+𝐧(t),\mathbf{o}(t)=\mathbf{s}(t)*\mathbf{h}(t)+\mathbf{n}(t)=\mathbf{s_{\_}{e}}(t)+\mathbf{s_{\_}{l}}(t)+\mathbf{n}(t),\vspace{-0.2cm} (1)

where * denotes convolution operation, 𝐬_𝐞(t)\mathbf{s_{\_}{e}}(t) refers to direct sound plus early reflections, and 𝐬_𝐥(t)\mathbf{s_{\_}{l}}(t) denotes late reverberation. The corresponding frequency domain signal model is shown as:

𝐎(t,f)=𝐒_𝐞(t,f)+𝐒_𝐥(t,f)+𝐍(t,f),\mathbf{O}(t,f)=\mathbf{S_{\_}{e}}(t,f)+\mathbf{S_{\_}{l}}(t,f)+\mathbf{N}(t,f),\vspace{-0.2cm} (2)

where 𝐎(t,f)\mathbf{O}(t,f), 𝐒_𝐞(t,f)\mathbf{S_{\_}{e}}(t,f), 𝐒_𝐥(t,f)\mathbf{S_{\_}{l}}(t,f) and 𝐍(t,f)\mathbf{N}(t,f) are the corresponding variable with frame index tt and frequency index ff in frequency domain. Our target is to estimate the 𝐬_𝐞\mathbf{s_{\_}{e}} and 𝐒_𝐞\mathbf{S_{\_}{e}} in this work.

2.2 Complex Self Attention

In the original self attention, the input 𝐗\mathbf{X} is mapped with different learnable linear transformation 𝐖\mathbf{W} to get queries (𝐐\mathbf{Q}), keys (𝐊\mathbf{K}) and values (𝐕\mathbf{V}) respectively:

𝐐=𝐗𝐖_Q,𝐊=𝐗𝐖_K,𝐕=𝐗𝐖_V.\mathbf{Q}=\mathbf{X}\mathbf{W}_{\_}{Q},\mathbf{K}=\mathbf{X}\mathbf{W}_{\_}{K},\mathbf{V}=\mathbf{X}\mathbf{W}_{\_}{V}.\vspace{-0.25cm} (3)

Then, the dot product results of queries with keys are computed, followed by division of a constant value kk, representing the projection dimension of 𝐖\mathbf{W}. After applying the softmax function to generate the weights, the weighted result is obtained by:

Attention(𝐐,𝐊,𝐕)=softmax(𝐐𝖳𝐊k)𝐕.\text{Attention}(\mathbf{Q},\mathbf{K},\mathbf{V})=\text{softmax}(\frac{\mathbf{Q}^{\mathsf{T}}\mathbf{K}}{\sqrt{k}})\mathbf{V}.\vspace{-0.25cm} (4)

Inspired by [18], given the complex input 𝐗\mathbf{X}, the complex valued 𝐐\mathbf{Q} is calculated by:

𝐐\displaystyle\mathbf{Q}^{\Re} =𝐗𝐖_Q𝐗𝐖_Q,\displaystyle=\mathbf{X}^{\Re}\mathbf{W}_{\_}{Q}^{\Re}-\mathbf{X}^{\Im}\mathbf{W}_{\_}{Q}^{\Im}, (5)
𝐐\displaystyle\mathbf{Q}^{\Im} =𝐗𝐖_Q+𝐗𝐖_Q,\displaystyle=\mathbf{X}^{\Re}\mathbf{W}_{\_}{Q}^{\Im}+\mathbf{X}^{\Im}\mathbf{W}_{\_}{Q}^{\Re},\vspace{-0.2cm}

where \Re and \Im indicate the real and imaginary parts, respectively. 𝐊\mathbf{K} and 𝐕\mathbf{V} are calculated in the same way. Thus, the complex self attention is calculated by:

ComplexAttention(𝐐,𝐊,𝐕)=\displaystyle\text{ComplexAttention}(\mathbf{Q},\mathbf{K},\mathbf{V})= (6)
(\displaystyle( Attention(𝐐,𝐊,𝐕)Attention(𝐐,𝐊,𝐕)\displaystyle\text{Attention}(\mathbf{Q}^{\Re},\mathbf{K}^{\Re},\mathbf{V}^{\Re})-\text{Attention}(\mathbf{Q}^{\Re},\mathbf{K}^{\Im},\mathbf{V}^{\Im})-
Attention(𝐐,𝐊,𝐕)Attention(𝐐,𝐊,𝐕))+\displaystyle\text{Attention}(\mathbf{Q}^{\Im},\mathbf{K}^{\Re},\mathbf{V}^{\Im})-\text{Attention}(\mathbf{Q}^{\Im},\mathbf{K}^{\Im},\mathbf{V}^{\Re}))+
i(\displaystyle i( Attention(𝐐,𝐊,𝐕)+Attention(𝐐,𝐊,𝐕)+\displaystyle\text{Attention}(\mathbf{Q}^{\Re},\mathbf{K}^{\Re},\mathbf{V}^{\Im})+\text{Attention}(\mathbf{Q}^{\Re},\mathbf{K}^{\Im},\mathbf{V}^{\Re})+
Attention(𝐐,𝐊,𝐕)Attention(𝐐,𝐊,𝐕)).\displaystyle\text{Attention}(\mathbf{Q}^{\Im},\mathbf{K}^{\Re},\mathbf{V}^{\Re})-\text{Attention}(\mathbf{Q}^{\Im},\mathbf{K}^{\Im},\mathbf{V}^{\Im})).

2.3 Dilated Complex Dual-path Conformer

The overall architecture of dilated complex dual-path conformer is shown Figure 2 (a). The module consists of time attention (TA), frequency attention (FA), dilated convolution (DC) and two feed forward (FF) layers. As for dilated real dual-path conformer, all modules have the same real-valued architectures with dilated complex dual-path conformer.

As shown in Figure 2 (b), FF consists of two linear layers with a swish activation in between. Inspired by [12], we employ half-step residual weights in our FF.

In the indoor acoustic scenario, sudden noises like clicking, coughing, etc., are common noise types without very long contextual dependency. On the other hand, the received signal is a collection of many delayed and attenuated copies of the original speech signal, resulting in strong feature correlations in a restricted range of time steps of the input sequence. To better model the local temporal relevance information mentioned above, inspired by [16], we use TA, which utilizes the frames within a context window, instead of the whole sentence, to calculate the local temporal features. For the input feature of TA: 𝐗_TA(t,f)C\mathbf{X}_{\_}{TA}(t,f)\in\mathbb{C}^{C} in Figure 2 (c), a cc-frame expansion 𝐗_TA¯(t,f)c×C\overline{\mathbf{X}_{\_}{TA}}(t,f)\in\mathbb{C}^{c\times C} is generated as contextual feature, where CC denotes channel numbers of bottle-neck feature. The 𝐐_TA(t,f)d_TA×c\mathbf{Q}_{\_}{TA}(t,f)\in\mathbb{C}^{d_{\_}{TA}\times c}, 𝐊_TA(t,f)d_TA×1\mathbf{K}_{\_}{TA}(t,f)\in\mathbb{C}^{d_{\_}{TA}\times 1} and 𝐕_TA(t,f)c×d_TA\mathbf{V}_{\_}{TA}(t,f)\in\mathbb{C}^{c\times d_{\_}{TA}} are obtained from 𝐗_TA¯\overline{\mathbf{X}_{\_}{TA}}, 𝐗_TA\mathbf{X}_{\_}{TA} and 𝐗_TA¯\overline{\mathbf{X}_{\_}{TA}} respectively by Eq. (3). The weight distribution on the context frames is computed based on the similarities between 𝐐_TA(t,f)\mathbf{Q}_{\_}{TA}(t,f) and 𝐊_TA(t,f)\mathbf{K}_{\_}{TA}(t,f) by the scaled dot product:

𝐀_TA(t,f)=softmax(𝐐_TA𝖳𝐊_TAd_TA).\mathbf{A}_{\_}{TA}(t,f)=\text{softmax}(\frac{\mathbf{Q}_{\_}{TA}^{\mathsf{T}}\mathbf{K}_{\_}{TA}}{\sqrt{d_{\_}{TA}}}).\vspace{-0.2cm} (7)

Multiplying the weight 𝐀_TA(t,f)c×1\mathbf{A}_{\_}{TA}(t,f)\in\mathbb{C}^{c\times 1} with 𝐕(i)\mathbf{V}(i), we can get the weighted feature with local temporal relevance information:

𝐘_TA(t,f)=𝐀_TA(t,f)𝐕_TA(t,f),\mathbf{Y}_{\_}{TA}(t,f)=\mathbf{A}_{\_}{TA}(t,f)\odot\mathbf{V}_{\_}{TA}(t,f),\vspace{-0.2cm} (8)

where \odot denotes the dot product. A fully connected layer is followed to project 𝐘_TA\mathbf{Y}_{\_}{TA} to the same dimension with 𝐗_TA\mathbf{X}_{\_}{TA}.

The lower frequency bands tend to contain high energies while the higher frequency bands tend to contain low energies. Therefore, we also pay different attention to different frequency subbands. In our proposed frequency attention (FA), for the input feature 𝐗_FA(t)F×C\mathbf{X}_{\_}{FA}(t)\in\mathbb{C}^{F\times C} in Figure 2 (c), projected features 𝐐_FA(t)d_FA×F\mathbf{Q}_{\_}{FA}(t)\in\mathbb{C}^{d_{\_}{FA}\times F}, 𝐊_FA(t)d_FA×F\mathbf{K}_{\_}{FA}(t)\in\mathbb{C}^{d_{\_}{FA}\times F} and 𝐕_FA(t)F×d_FA\mathbf{V}_{\_}{FA}(t)\in\mathbb{C}^{F\times d_{\_}{FA}} are obtained by Eq. (3). After generating the weight distribution on different subbands 𝐀(t)_FAF×F\mathbf{A}(t)_{\_}{FA}\in\mathbb{C}^{F\times F}, the FA result 𝐘(t)_FAF×d_FA\mathbf{Y}(t)_{\_}{FA}\in\mathbb{C}^{F\times d_{\_}{FA}} is calculated by Eq. (4). A fully connected layer is followed to project 𝐘_TA\mathbf{Y}_{\_}{TA} to the same dimension with 𝐗_FA\mathbf{X}_{\_}{FA}. The detailed architecture of TA and FA are shown in Figure 2 (c).

TasNet [19] has shown superior performance in speech separation, which uses stacked temporal convolution network (TCN) to better capture long range sequence dependencies. Original TCN first projects the input to a higher channel space with Conv1d. Then a dilated depthwise convolution (D-Conv) is applied to get a larger receptive field. The output Conv1d projects the number of channel the same as the input. Residual connection is applied to enforce the network to focus on the missing detail and mitigate gradient disappearance. Our improvements on TCN in dilated dual-path conformer are shown in Figure 2 (d). Gated D-Conv2d is applied with opposite dilation in two D-Conv2ds. The dilation of lower D-Conv2d is 1,2,,2N11,2,...,2^{N-1} while the dilation of upper gated D-Conv2d is 2N1,2N2,,12^{N-1},2^{N-2},...,1, where NN denotes the cascaded layer number of dilated dual-path conformer.

2.4 Hybrid Encoder and Decoder

In our proposed hybrid encoder and decoder, both encoder and decoder model the complex spectrum and magnitude at the same time. Let 𝐂_i\mathbf{C}_{\_}{i} and 𝐌_i\mathbf{M}_{\_}{i} denotes the complex spectrum and magnitude output of encoder/decoder layer ii respectively. To promote the information exchange between complex spectrum and magnitude, the complex-magnitude fusion results 𝐂^_i\mathbf{\hat{C}}_{\_}i and 𝐌^_i\mathbf{\hat{M}}_{\_}i are calculated by:

𝐂^_i\displaystyle\mathbf{\hat{C}}_{\_}i^{\Re} =𝐂_i+σ(𝐌_i),\displaystyle=\mathbf{C}_{\_}i^{\Re}+\sigma(\mathbf{M}_{\_}{i}), (9)
𝐂^_i\displaystyle\mathbf{\hat{C}}_{\_}i^{\Im} =𝐂_i+σ(𝐌_i),\displaystyle=\mathbf{C}_{\_}i^{\Im}+\sigma(\mathbf{M}_{\_}{i}),
𝐌^_i\displaystyle\vspace{-0.2cm}\mathbf{\hat{M}}_{\_}i =𝐌_i+σ(𝐂_i2+𝐂_i2).\displaystyle=\mathbf{M}_{\_}{i}+\sigma(\sqrt{\mathbf{C}_{\_}i^{\Re^{2}}+\mathbf{C}_{\_}i^{\Im^{2}}}).

2.5 Encoder Decoder Attention

Let 𝐄_i\mathbf{E}_{\_}{i} denotes the output of hybrid encoder layer ii while 𝐃_i\mathbf{D}_{\_}{i} denotes the output of dilated conformer layer or hybrid decoder layer ii. Two Conv2ds are first applied to 𝐄_i\mathbf{E}_{\_}{i} and 𝐃_i\mathbf{D}_{\_}{i} to map them to high dimensional features 𝐆_i\mathbf{G}_{\_}{i} by:

𝐆_i=σ(𝐖_Ei𝐄_i+𝐖_Di𝐃_i),\mathbf{G}_{\_}{i}=\sigma(\mathbf{W}^{E}_{\_}{i}\ast\mathbf{E}_{\_}{i}+\mathbf{W}^{D}_{\_}{i}\ast\mathbf{D}_{\_}{i}),\vspace{-0.2cm} (10)

where 𝐖_Ei\mathbf{W}^{E}_{\_}{i} and 𝐖_Di\mathbf{W}^{D}_{\_}{i} denote the convolution kernel first two Conv2ds of layer ii, respectively. After applying a third Conv2d to 𝐆_i\mathbf{G}_{\_}{i}, the final output is calculated by multiplying the sigmoid attention mask with 𝐃_i\mathbf{D}_{\_}{i}:

𝐃^_i=σ(𝐖_Ai𝐆_i)𝐃_i,\mathbf{\hat{D}}_{\_}{i}=\sigma(\mathbf{W}^{A}_{\_}{i}\ast\mathbf{G}_{\_}{i})\odot\mathbf{D}_{\_}{i},\vspace{-0.2cm} (11)

where 𝐖_Ai\mathbf{W}^{A}_{\_}{i} denotes the convolution kernel the third Conv2d of layer ii. Finally, we concatenate 𝐃_i\mathbf{D}_{\_}{i} and 𝐃^_i\mathbf{\hat{D}}_{\_}{i} along channel axis as the input of the next decoder layer.

2.6 Loss Function

After generating the estimated CRM 𝐇_C\mathbf{H}_{\_}{C} and IRM 𝐇_R\mathbf{H}_{\_}{R} from the last decoder layer, we multiply the noisy complex spectrum and magnitude with them respectively to get enhanced and dereverbed complex spectrum and magnitude:

𝐌_magC=tanh(𝐇_2C+𝐇_2C),\mathbf{M}^{\text{mag}}_{\_}{C}=\tanh(\sqrt{\mathbf{H}^{\Re 2}_{\_}{C}+\mathbf{H}^{\Im 2}_{\_}{C}}),\vspace{-0.5cm} (12)
𝐌_phaC=arctan2(𝐇_C,𝐇_C),\mathbf{M}^{\text{pha}}_{\_}{C}=\text{arctan2}(\mathbf{H}^{\Im}_{\_}{{C}},\mathbf{H}^{\Re}_{\_}{{C}}), (13)
𝐌_magR=σ(𝐇_R),\mathbf{M}^{\text{mag}}_{\_}{R}=\sigma(\mathbf{H}_{\_}{R}), (14)
𝐘_magC=𝐗mag𝐌_magC,\mathbf{Y}^{\text{mag}}_{\_}{C}=\mathbf{X}^{\text{mag}}\odot\mathbf{M}^{\text{mag}}_{\_}{C}, (15)
𝐘_phaC=𝐗pha+𝐌_phaC,\mathbf{Y}^{\text{pha}}_{\_}{C}=\mathbf{X}^{\text{pha}}+\mathbf{M}^{\text{pha}}_{\_}{C}, (16)
𝐘_magR=𝐗mag𝐌_magR,\mathbf{Y}^{\text{mag}}_{\_}{R}=\mathbf{X}^{\text{mag}}\odot\mathbf{M}^{\text{mag}}_{\_}{{R}}, (17)
𝐘_magC¯=(𝐘_magC+𝐘_magR)/2,\overline{\mathbf{Y}^{\text{mag}}_{\_}{C}}=(\mathbf{Y}^{\text{mag}}_{\_}{C}+\mathbf{Y}^{\text{mag}}_{\_}{R})/2, (18)
𝐘=𝐘_magC¯cos(𝐘_phaC),\mathbf{Y}^{\Re}=\overline{\mathbf{Y}^{\text{mag}}_{\_}{C}}\odot\cos(\mathbf{Y}^{\text{pha}}_{\_}{C}), (19)
𝐘=𝐘_magC¯sin(𝐘_phaC),\mathbf{Y}^{\Im}=\overline{\mathbf{Y}^{\text{mag}}_{\_}{C}}\odot\sin(\mathbf{Y}^{\text{pha}}_{\_}{C}), (20)
𝐲=iSTFT(𝐘,𝐘).\mathbf{y}=\text{iSTFT}(\mathbf{Y}^{\Re},\mathbf{Y}^{\Im}).\vspace{-0.2cm} (21)

We use hybrid time and frequency domain loss as the target:

=α_SI-SNR+β_L1T+γ_L2C+ζ_L2M,\mathcal{L}=\alpha\mathcal{L}_{\_}{\textbf{SI-SNR}}+\beta\mathcal{L}_{\_}{\textbf{L1}}^{\textbf{T}}+\gamma\mathcal{L}_{\_}{\textbf{L2}}^{\textbf{C}}+\zeta\mathcal{L}_{\_}{\textbf{L2}}^{\textbf{M}},\vspace{-0.2cm} (22)

where _SI-SNR\mathcal{L}_{\_}{\textbf{SI-SNR}}, _L1T\mathcal{L}_{\_}{\textbf{L1}}^{\textbf{T}}, _L2C\mathcal{L}_{\_}{\textbf{L2}}^{\textbf{C}} and _L2M\mathcal{L}_{\_}{\textbf{L2}}^{\textbf{M}} denote SI-SNR [20] loss in time domain, L1 loss in time domain, complex spectrum L2 loss and magnitude L2 loss, respectively. α\alpha, β\beta, γ\gamma and ζ\zeta denote the weight of four losses. Four losses are calculated by:

_SI-SNR\displaystyle\mathcal{L}_{\_}{\textbf{SI-SNR}} =20log_10ξ𝐲^𝐲ξ𝐲^,\displaystyle=-20\log_{\_}{10}\frac{\|\xi\cdot\mathbf{\hat{y}}\|}{\|\mathbf{y}-\xi\cdot\mathbf{\hat{y}}\|}, (23)
ξ\displaystyle\xi =𝐲T𝐲^/𝐲^T𝐲^,\displaystyle=\mathbf{y}^{\text{T}}\mathbf{\hat{y}}/\mathbf{\hat{y}}^{\text{T}}\mathbf{\hat{y}},
_L1T=_t|(𝐲(t)𝐲^(t))|,\mathcal{L}_{\_}{\textbf{L1}}^{\textbf{T}}=\sum_{\_}{t}\left|(\mathbf{y}(t)-\mathbf{\hat{y}}(t))\right|, (24)
_L2C=(_t,f\displaystyle\mathcal{L}_{\_}{\textbf{L2}}^{\textbf{C}}=(\sum_{\_}{t,f} |(𝐘(t,f)𝐘^(t,f))|2+\displaystyle\left|(\mathbf{Y}^{\Re}(t,f)-\mathbf{\hat{Y}}^{\Re}(t,f))\right|^{2}+ (25)
|(𝐘(t,f)𝐘^(t,f))|2)/F,\displaystyle\left|(\mathbf{Y}^{\Im}(t,f)-\mathbf{\hat{Y}}^{\Im}(t,f))\right|^{2})/F,
_L2M=(_t,f|(𝐘_magR(t,f)𝐘^_magR(t,f))|2)/F,\mathcal{L}_{\_}{\textbf{L2}}^{\textbf{M}}=(\sum_{\_}{t,f}\left|(\mathbf{Y}^{\text{mag}}_{\_}{R}(t,f)-\mathbf{\hat{Y}}^{\text{mag}}_{\_}{R}(t,f))\right|^{2})/F,\vspace{-0.1cm} (26)

where y and y^\hat{\textbf{y}} as the estimated and ground truth time domain signal, while Y and Y^\hat{\textbf{Y}} denote their corresponding frequency domain spectrum. We conduct a fusion operation in Eq. (18) between estimated complex spectrum and magnitude since we find out that calculating the complex loss only using the estimated complex output without fusion may cause serious distortion, thus leading to extremely bad objective and subjective results.

3 Experiments and Results

3.1 Datasets

In our experiments, the source speech data comes from LibriTTS [21], AISHELL-3 [22], speech data of DNS challenge [23], and the vocal part of MUSDB [24] while the source noise data comes from MUSAN [25], noise data of DNS challenge, the music part of MUSDB, MS-SNSD [26] and collected pure music data including classical and pop music. The training set contains 982 h source speech data and 230 h source noise data, while the development set contains 73 h source speech data and source 30 h noise data, respectively. Image method [27] is used to simulate RIR with RT60 ranges from 0.2 s to 1.2 s. Early reflection within 50 ms is used as the dereverberation training target. The training data are generated on-the-fly with 16k Hz sampling rate and segmented into 4 s chunks in one batch with SNR ranges from -5 to 15 dB. For model evaluation, three SNR ranges are selected to generate simulated dataset, namely [-5, 0], [0, 5] and [5, 10] dB, respectively. We simulate 500 noisy and reverberate pieces in each set. The source data among training, development and simulated evaluation set have no overlap. The blind test set of Interspeech 2021 DNS Challenge , which contains 600 noisy and reverberate pieces, is also selected as another evaluation dataset.

Table 1: Results on different models in terms of PESQ, eSTOI, DNSMOS and MOS, where PESQ and eSTOI are calculated on simulated test set while DNSMOS and MOS are calculated on Interspeech2021 DNS challenge blind test set. Note that SDD-Net and DCCRN+ here are the results submitted to the challenge.
Model Cau. #Param. (M) PESQ eSTOI DNSMOS MOS
SNR (dB) - - [-5,0] [0,5] [5,10] [-5,0] [0,5] [5,10] - -
Noisy - - 1.4710 1.7616 1.9904 43.50 53.54 60.96 2.4139 1.8545
UFormer ×\times 9.46 2.4501 2.7472 2.9511 64.63 74.33 79.62 3.6032 3.3545
UFormer \checkmark 9.46 2.4023 2.7265 2.9250 64.22 74.29 79.46 3.5890 3.3523
- FA ×\times 9.02 2.4207 2.7273 2.9306 64.19 74.11 79.37 3.5801 -
- DC ×\times 9.31 2.3374 2.6689 2.8883 62.86 73.03 78.52 3.5654 -
- encoder-decoder
attention
×\times 5.33 2.4218 2.7217 2.9177 64.27 74.10 79.37 3.5381 -
dilated conformer
\rightarrow LSTM
×\times 9.47 2.4106 2.7243 2.9258 64.11 73.90 79.31 3.5839 -
- real-valued
sub-modules
×\times 7.26 2.4266 2.7352 2.9402 64.28 74.26 79.58 3.5751 -
- complex-valued
sub-modules
×\times 3.85 2.4039 2.7025 2.9095 63.55 73.37 78.78 3.5265 -
DCCRN [6] \checkmark 8.99 2.3652 2.6674 2.8676 62.25 72.55 77.97 3.4915 3.2773
GCRN [5] ×\times 30.83 2.2672 2.5768 2.7883 61.43 71.87 77.70 3.3452 -
PHASEN [28] ×\times 8.41 2.3203 2.6170 2.8072 62.76 72.73 78.12 3.4518 -
SDD-Net [17] \checkmark 6.38 - - - - - -
3.36/3.47/
3.56/3.60
3.3432
DCCRN+ [7] \checkmark 4.71 - - - - - - 3.4260 3.0682
TasNet [19] ×\times 8.69 2.2671 2.5649 2.7808 61.11 71.30 77.50 3.3832 -
DPRNN [29] ×\times 2.60 2.2758 2.5723 2.7752 61.30 71.61 77.10 3.2524 -

3.2 Training Setups

For the input time domain signal, a 512-point short time Fourier transform (STFT) with a window size of 25 ms and a window shift of 10 ms is applied to each frame, resulting in 257 frequency bins. The channel numbers of encoder layers are [8, 16, 32, 64, 128, 128] and are inversed for decoder layers. The kernel size and stride for time and frequency axis are (2, 5) and (1, 2). The frame expansion in TA is 9 frames. For non-causal models, we combine the current frame with 4 historical and 4 future frames together as frame expansion. For causal models, we combine the current frame with 8 historical frames as frame expansion. For dilated conformer, the hidden dimension of two linear layers in FF are 64 and 128, respectively. The projection dimension and head number of both TA and FA are 16 and 1, respectively. For DC module, inspired by [30], the output channel of the first Conv2d, D-Conv2d and the second Conv2d are 32, 32 and 128, respectively, to compress the feature dimension. The kernel size for time and frequency axis is (2, 1). The dilated conformer module is cascaded for 8 layers and the dilation expansion rate is 2. For causal models, dilated conformer only uses historical frames. For encoder-decoder attention, the kernel size of three Conv2ds is (2, 3). The dropout rate for all dropout layers in Uformer is 0.1. Both real and complex modules have the same configuration. Most recently, the effectiveness of power compressed spectrum has been demonstrated in the dereverberation task. We conduct the compression to magnitude before sending it into the network, and the compression variable is 0.5, which is the reported optimal parameter in  [31]. All models are trained with Adam optimizer for 15 epochs, thus all models are trained with 14730 h different noisy and reverberate speech data totally. The initial learning rate is 0.001 and will get halved if there is no loss decrease on the development set. Gradient clipping is applied with a maximum l2 norm of 5. α\alpha, β\beta, γ\gamma, ζ\zeta in hybrid loss are 5, 1/30, 1 and 1, respectively to make four losses are on the relatively same order of magnitude.

3.3 Experiments

We conduct ablation experiments to prove the effectiveness of each proposed sub-modules, including a) Uformer without FA, b) Uformer without DC, c) Uformer without encoder decoder attention, d) substitute dilated complex & real dual-path conformer with complex & real LSTM, e) Uformer without all real-valued sub-modules (means we only model complex spectrum) and f) Uformer without all complex-valued sub-modules (means we only model magnitude). We also compare Uformer with DCCRN, PHASEN [28], GCRN, SDD-Net [17], TasNet and DPRNN [29]. DCCRN, PHASEN and GCRN are in complex domain, which aim to model magnitude and phase simultaneously. TasNet and DPRNN are advanced time domain speech front-end models. SDD-Net and DCCRN+ won the first and third prize of Interspeech 2021 DNS Challenge. The results of SDD-Net and DCCRN+ are the final results submitted to the challenge. All complex domain models are optimized with hybrid time and frequency domain loss while time domain models are optimized with SI-SNR loss. All the models are implemented with the best configurations mentioned in the corresponding literature.

3.4 Results and Analysis

Four evaluation metrics are used, namely perceptual evaluation of speech quality (PESQ), extended short-time objective intelligibility (eSTOI), DNSMOS [32] and MOS. PESQ, and eSTOI are used to evaluate the objective performance of speech quality and intelligibility for simulated test set. DNSMOS is a non-intrusive perceptual objective metric, which is used to simulates the human subjective evaluation on DNS blind test set. We also organize 12 listeners to conduct a MOS test on 20 randomly selected clips from DNS blind test set to assess UFormer, DCCRN, DCCRN+ and SDD-Net.

The experimental results are shown in Table 1. The four DNSMOS results of SDD-Net indicate the results of four training stages [17], namely denoising, dereverberation, spectral refinement and post-processing, respectively. Uformer gives the best performance in both objective and subjective evaluation, while the result of the causal version doesn’t degrade generally. Compared with the time domain models, better performance is achieved for complex domain approaches, which indicates that it is more suitable to perform simultaneous enhancement and dereverberation in the T-F domain than the waveform level. Uformer reaches 3.6032 on DNSMOS, which is superior to all complex domain neural network based models and has relatively equal ability with the SDD-Net with post processing. However, compared with Uformer conducted in the form of end-to-end speech enhancement and dereverberation, the training procedure of SDD-Net is divided into four separate stages mentioned above, resulting in a relatively complicated training process. The idea of using dilated convolution layer shows great ability of modeling long range sequence dependencies, which lead to 0.08 and 0.04 average improvements on PESQ and DNSMOS, respectively. In addition, FA, encoder-decoder attention also contribute to a great extend. Finally, the PESQ and DNSMOS degrade 0.02 and 0.04 when only modeling complex spectrum and degrade 0.03 and 0.08 when only modeling magnitude. These two results confirm the importance of the information interaction of complex spectrum and magnitude. Uformer model also achieves the best MOS results (3.3545) among all SOTA complex domain models

4 Conclusion

In this paper, we propose Uformer 111Opensource code is available at https://github.com/felixfuyihui/Uformer for simultaneous speech enhancement and dereverberation in both magnitude and complex domains. We leverage local and global contextual information as well as frequency information to improve the speech enhancement and dereverberation ability by dilated complex & real dual-path conformer module. Hybrid encoder and decoder is adopted to model the complex spectrum and magnitude simultaneously. Encoder decoder attention is applied to better model the information interaction between encoder and decoder. Our experimental results indicate that the proposed UFormer reaches 3.6032 DNSMOS on the blind test set of Interspeech 2021 DNS Challenge, which surpasses the challenge best model – SDD-Net.

References

  • [1] Yuxuan Wang, Arun Narayanan, and DeLiang Wang, “On training targets for supervised speech separation,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 22, no. 12, pp. 1849–1858, 2014.
  • [2] D Wang and J. S Lim, “The unimportance of phase in speech enhancement,” Acoustics Speech Signal Processing IEEE Transactions on, vol. 30, no. 4, pp. 679–681, 1982.
  • [3] Kuldip Paliwal, Kamil Wójcicki, and Benjamin Shannon, “The importance of phase in speech enhancement,” Speech Communication, vol. 53, no. 4, pp. 465–494, 2011.
  • [4] Donald S Williamson and DeLiang Wang, “Time-frequency masking in the complex domain for speech dereverberation and denoising,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 25, no. 7, pp. 1492–1501, 2017.
  • [5] Ke Tan and DeLiang Wang, “Learning complex spectral mapping with gated convolutional recurrent networks for monaural speech enhancement,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 28, pp. 380–390, 2019.
  • [6] Yanxin Hu, Yun Liu, Shubo Lv, Mengtao Xing, Shimin Zhang, Yihui Fu, Jian Wu, Bihong Zhang, and Lei Xie, “DCCRN: Deep complex convolution recurrent network for phase-aware speech enhancement,” Interspeech, pp. 2472–2476, 2020.
  • [7] Shubo Lv, Yanxin Hu, Shimin Zhang, and Lei Xie, “DCCRN+: Channel-wise subband dccrn with snr estimation for speech enhancement,” arXiv preprint arXiv:2106.08672, 2021.
  • [8] Hideaki Kagami, Hirokazu Kameoka, and Masahiro Yukawa, “Joint separation and dereverberation of reverberant mixtures with determined multichannel non-negative matrix factorization,” in IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2018, pp. 31–35.
  • [9] Henri Gode, Marvin Tammen, and Simon Doclo, “Joint multi-channel dereverberation and noise reduction using a unified convolutional beamformer with sparse priors,” arXiv preprint arXiv:2106.01902, 2021.
  • [10] Lukas Pfeifenberger and Franz Pernkopf, “Blind speech separation and dereverberation using neural beamforming,” arXiv preprint arXiv:2103.13443, 2021.
  • [11] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Lukasz Kaiser, and Illia Polosukhin, “Attention is all you need,” Advances in neural information processing systems, vol. 30, 2017.
  • [12] Anmol Gulati, James Qin, Chung-Cheng Chiu, Niki Parmar, Yu Zhang, Jiahui Yu, Wei Han, Shibo Wang, Zhengdong Zhang, Yonghui Wu, and Ruoming Pang, “Conformer: Convolution-augmented transformer for speech recognition,” arXiv preprint arXiv:2005.08100, 2020.
  • [13] Sanyuan Chen, Yu Wu, Zhuo Chen, Jian Wu, Jinyu Li, Takuya Yoshioka, Chengyi Wang, Shujie Liu, and Ming Zhou, “Continuous speech separation with conformer,” in IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2021, pp. 5749–5753.
  • [14] Feng Dang, Hangting Chen, and Pengyuan Zhang, “Dpt-fsnet: Dual-path transformer based full-band and sub-band fusion network for speech enhancement,” arXiv preprint arXiv:2104.13002, 2021.
  • [15] Kai Wang, Bengbeng He, and Wei-Ping Zhu, “Tstnn: Two-stage transformer based neural network for speech enhancement in the time domain,” in IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2021, pp. 7098–7102.
  • [16] Helin Wang, Bo Wu, Lianwu Chen, Meng Yu, Jianwei Yu, Yong Xu, Shi-Xiong Zhang, Chao Weng, Dan Su, and Dong Yu, “TeCANet: Temporal-contextual attention network for environment-aware speech dereverberation,” arXiv preprint arXiv:2103.16849, 2021.
  • [17] Andong Li, Wenzhe Liu, Xiaoxue Luo, Guochen Yu, Chengshi Zheng, and Xiaodong Li, “A simultaneous denoising and dereverberation framework with target decoupling,” arXiv preprint arXiv:2106.12743, 2021.
  • [18] Muqiao Yang, Martin Q Ma, Dongyu Li, Yao-Hung Hubert Tsai, and Ruslan Salakhutdinov, “Complex transformer: A framework for modeling complex-valued sequence,” in IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2020, pp. 4232–4236.
  • [19] Yi Luo and Nima Mesgarani, “Conv-tasnet: Surpassing ideal time–frequency magnitude masking for speech separation,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 27, no. 8, pp. 1256–1266, 2019.
  • [20] Jonathan Le Roux, Scott Wisdom, Hakan Erdogan, and John R Hershey, “SDR–half-baked or well done?,” in IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2019, pp. 626–630.
  • [21] Heiga Zen, Viet Dang, Rob Clark, Yu Zhang, Ron J Weiss, Ye Jia, Zhifeng Chen, and Yonghui Wu, “LibriTTS: A corpus derived from LibriSpeech for text-to-speech,” arXiv preprint arXiv:1904.02882, 2019.
  • [22] Yao Shi, Hui Bu, Xin Xu, Shaoji Zhang, and Ming Li, “Aishell-3: A multi-speaker mandarin tts corpus and the baselines,” arXiv preprint arXiv:2010.11567, 2020.
  • [23] Chandan KA Reddy, Harishchandra Dubey, Kazuhito Koishida, Arun Nair, Vishak Gopal, Ross Cutler, Sebastian Braun, Hannes Gamper, Robert Aichner, and Sriram Srinivasan, “Interspeech 2021 deep noise suppression challenge,” arXiv preprint arXiv:2101.01902, 2021.
  • [24] Fabian-Robert Stöter, Antoine Liutkus, and Nobutaka Ito, “The 2018 signal separation evaluation campaign,” in International Conference on Latent Variable Analysis and Signal Separation. Springer, 2018, pp. 293–305.
  • [25] David Snyder, Guoguo Chen, and Daniel Povey, “MUSAN: A music, speech, and noise corpus,” arXiv preprint arXiv:1510.08484, 2015.
  • [26] Chandan KA Reddy, Ebrahim Beyrami, Jamie Pool, Ross Cutler, Sriram Srinivasan, and Johannes Gehrke, “A scalable noisy speech dataset and online subjective test framework,” arXiv preprint arXiv:1909.08050, 2019.
  • [27] Jont B Allen and David A Berkley, “Image method for efficiently simulating small-room acoustics,” The Journal of the Acoustical Society of America, vol. 65, no. 4, pp. 943–950, 1979.
  • [28] Dacheng Yin, Chong Luo, Zhiwei Xiong, and Wenjun Zeng, “PHASEN: A phase-and-harmonics-aware speech enhancement network,” in Conference on Artificial Intelligence. AAAI, 2020, vol. 34, pp. 9458–9465.
  • [29] Yi Luo, Zhuo Chen, and Takuya Yoshioka, “Dual-Path RNN: efficient long sequence modeling for time-domain single-channel speech separation,” in IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2020, pp. 46–50.
  • [30] Andong Li, Chengshi Zheng, Lu Zhang, and Xiaodong Li, “Glance and gaze: A collaborative learning framework for single-channel speech enhancement,” arXiv preprint arXiv:2106.11789, 2021.
  • [31] Andong Li, Chengshi Zheng, Renhua Peng, and Xiaodong Li, “On the importance of power compression and phase estimation in monaural speech dereverberation,” JASA Express Letters, vol. 1, no. 1, pp. 014802, 2021.
  • [32] Chandan KA Reddy, Vishak Gopal, and Ross Cutler, “DNSMOS: A non-intrusive perceptual objective speech quality metric to evaluate noise suppressors,” in IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2021, pp. 6493–6497.