This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

MLNET: An Adaptive Multiple Receptive-field Attention Neural Network for Voice Activity Detection

Abstract

Voice activity detection (VAD) makes a distinction between speech and non-speech and its performance is of crucial importance for speech based services. Recently, deep neural network (DNN)-based VADs have achieved better performance than conventional signal processing methods. The existed DNN-based models always handcrafted a fixed window to make use of the contextual speech information to improve the performance of VAD. However, the fixed window of contextual speech information can’t handle various unpredictable noise environments and highlight the critical speech information to VAD task. In order to solve this problem, this paper proposed an adaptive multiple receptive-field attention neural network, called MLNET, to finish VAD task. The MLNET leveraged multi-branches to extract multiple contextual speech information and investigated an effective attention block to weight the most crucial parts of the context for final classification. Experiments in real-world scenarios demonstrated that the proposed MLNET-based model outperformed other baselines.

Index Terms: Voice Activity Detection, Adaptive Mutiple Receptive-field Attention

1 Introduction

Voice Activity Detection (VAD), which aims at removing noise or silence from the original speech signal and obtaining valid speech signal, is an essential part of speech recognition or other speech-based applications [1, 2, 3]. Unfortunately, in the real environment, the speech signals may contain numerous background noises and have a low signal-to-noise ratio (SNR), which brings great challenges to an accurate VAD system.

With the continuous development of speech technologies, the research on VAD has become a continuing hot spot [4, 5, 6]. Early research focused on parametric methods: energy function, zero-crossing rate, statistical signal analysis or other acoustic features [7, 8, 9, 10]. Later, various machine learning based methods were established to VAD systems: Gaussian Mixture Models (GMM) [11], Hidden Markov Model(HMM) [12] or Support Vector Machines(SVM) [13]. More specifically, deep learning models were also established: deep neural network(DNN)[14, 15, 16, 17, 18], deep belief network(DBN) [19], convolutional neural network (CNN) [20, 21], recurrent neural network (RNN) [22, 23]. Moreover, speech enhancement (SE) based methods were also introduced to reply the low SNR noisy environments [24, 25].

The above machine learning based models have achieved great progress in VAD task. These models had a common characteristics: in order to get a robust VAD system, speech including contextual information, just as other speech-based systems, would be forwarded into a neural network. However, when training or testing, it was hard to select the optimal hyper-parameter to determine the amount of contextual information. Specially, when selecting more information, much noisy information may be included in short speech segments and cause false positive detections. When less contextual information was selected, the VAD system may not make use of effective information to make the right classification. Based on the above analysis, when performing VAD tasks, the VAD systems should be capable of selecting the most appropriate contextual information according to the characteristics of current speech and focusing on the most important speech segments to obtain the optimal detection results. Motivated by the successful application of attention mechanism in image and natural language understanding tasks, this paper proposed an attention-based model of MLNET to choose the appropriate speech segment for the classifier. The MLNET took advantage of different gated units to extract different contextual speech information and leveraged attention mechanism of channel selection to choose the most appropriate contextual information to adapt different noisy environments. In particular, the first attention model for VAD task was the ACAM model and it focused on the effect of a certain frame in a fixed window, which is different from our window size attention [23]. From our point of view, the most important frame for the detection result is the current frame and the surrounding frames are also important for final result but they are the auxiliary information.

With respect to the state of the art, the main contributions of this paper were as following:

  • A architecture of MLNET, which could adapt different segments and select the optimal contextual information for the final classifier, was proposed to VAD.

  • A useful mechanism of gated units was leveraged to extract different contextual speech information.

  • An attention strategy for effective and appropriate contextual information selection was investigated.

  • The proposed method was benchmarked against several state-of-art methods and the functions of each part in MLNET were also compared.

2 Methodology

2.1 Model Structure

The MLNET-based VAD achieved a frame-based speech or non-speech classifier. Suppose the training corpus can be marked as χ={(xt,yt)}t=1T\chi={\{(x_{t},y_{t})\}}_{t=1}^{T} , where xtx_{t} is an acoustic feature vector of tt-th frame audio signal and yty_{t} denotes the label of xtx_{t}. If xtx_{t} is a speech frame, then yt=1y_{t}=1; otherwise, yt=0y_{t}=0. Because of contextual information’s importance for speech applications, xtx_{t} is usually expanded to It=[xtr,,xt1,xt,xt+1,,xt+r]I_{t}=[x_{t-r},...,x_{t-1},x_{t},x_{t+1},...,x_{t+r}] when training or testing. Here, the objective of VAD is to learn a function ff as (1).

yt^=f([xtr,,xt1,xt,xt+1,,xt+r])\hat{y_{t}}=f([x_{t-r},...,x_{t-1},x_{t},x_{t+1},...,x_{t+r}]) (1)

where yt^\hat{y_{t}} denotes the predicting result of xtx_{t} and rr denoted the window size of contextal speech information.

In experiment, the value rr was a hyper-parameter and the fixed value of rr was hard to adapt various speech environments. For example, when xtx_{t} wa a speech frame and the speech segment duration around xtx_{t} was short, the large rr would contain much non-speech information and cause false positive results. In turn, when rr was short, the contextual information may not be fully utilized, which can result in false negative or false positive results. To address the above problems, this paper proposed the MLNET model, which leveraged the multiple gated affined units and attention mechanism to choose the optimal receptive-field speech adaptively to make the classification. The MLNET’s architecture is shown in Figure 1.

Refer to caption

Figure 1: The MLNET’s Architecture. The xti(i[r,+r])x_{t-i}(i\in[-r,+r]) represented 40-log Mel Features of a frame and (2r+1)(2r+1) frames’ features would be inputted into the MLNET.

2.2 Adaptive Multiple Receptive-field Attention Block

The detail architecture of multiple receptive-field attention block was showed in Figure 2. In order to realize adaptive feature selection, multiple branches were leveraged to extract different receptive-filed speech features and each branch represents a specific receptive field or a specific contextual information. Specifically, for the branch of rir_{i}, the contextual speech information can be denoted as It=[xtri,,xt1,xt,xt+1,,xt+ri]I_{t}^{{}^{\prime}}=[x_{t-r_{i}},...,x_{t-1},x_{t},x_{t+1},...,x_{t+r_{i}}] and (2×ri+1)(2\times r_{i}+1) frame features were included. For the calculation of subsequent attention modules, the feature matrix of ItI_{t}^{{}^{\prime}} with different size need to be converted feature matrix with the same size. This paper made use of the gated affined unit to finish this task. The gated affined unit as this non-linearity has proved to work better for modeling audio than the other affine function or Relu functions [26]. The gated affined unit’s definition was defined as (3):

qi=tanh(Wf,riIi+bf,ri)σ(Wg,riIi+bg,ri)q_{i}=tanh(W_{f,r_{i}}*I_{i}+b_{f,r_{i}})\odot\sigma(W_{g,r_{i}}*I_{i}+b_{g,r_{i}}) (2)

where Wf,riW_{f,r_{i}}, bf,rib_{f,r_{i}} denoted affine parameters of tanh and Wg,riW_{g,r_{i}}, bg,rib_{g,r_{i}} denoted affine parameters of sigmoid. It was obvious that the size of Wf,riW_{f,r_{i}} and Wg,riW_{g,r_{i}} would change with rir_{i} while the size of qiq_{i} was constant regardless of rir_{i}. The * denoted a convolution operator while the \odot denoted an element-wise multiplication operator.

Refer to caption


Figure 2: Adaptive Multiple Receptive-field Attention Block. As can be seen, the inputted feature with different receptive-field rir_{i} would be calculated at different branches and the attentional feature vector of time tt would be calculated based on each branch.

After gated affined unit operation, different two-dimensional extracted matrixes, which were denoted by qt,1,,qt,i,,qt,rq_{t,1},...,q_{t,i},...,q_{t,r}, were obtained and each feature matrix of qt,iq_{t,i} had the same size. Analogy with image, each feature matrix of qt,iq_{t,i} can be regarded as a channel feature map and we produced a channel attention map to decide which feature matrix should be focused on. Based on previous channel attention work in images [27, 28], the attention module’s structure was showed in Figure 3. Firstly, both average-pooling and max-pooling operations were used to aggregate information of a feature matrix and two feature descriptors were generated to represent different feature matrixes. Then, both descriptors were forward to a shared 2-layer fully connected DNN to produce two different attention ratio vector pt,maxp_{t,max} and pt,avgp_{t,avg}. Finally, by adding an norm operation, pt,maxp_{t,max} and pt,avgp_{t,avg} were used to generate final attention ratio vector ata_{t} of (4).

at\displaystyle a_{t} =σ(FC(avgpool(qt))+FC(maxpool(qt)))\displaystyle=\sigma{(FC(avgpool(q_{t}))+FC(maxpool(q_{t})))} (3)
=σ(W1(W0(avgpool(qt)))+W1(W0(maxpool(qt))))\displaystyle=\sigma{(W_{1}(W_{0}(avgpool(q_{t})))+W_{1}(W_{0}(maxpool(q_{t}))))}

where FCFC denoted the shared 2-layer fully connected DNN and W1W_{1} and W0W_{0} were the weights of DNN. Note that the weight W0W_{0} of first layer followed a leaky_reluleaky\_relu activation function. After obtaining ata_{t}, the attentioned or scaled feature matrix would be calculated through (4)-(7).

at\displaystyle a_{t} =[at,1,,at,i,,at,r]\displaystyle=[a_{t,1},...,a_{t,i},...,a_{t,r}] (4)
pt,i=σ(at,i)/(i=1rσ(at,i))p_{t,i}=\sigma{(a_{t,i})}/(\sum_{i=1}^{r}{\sigma{(a_{t,i})}}) (5)
pt=[pt,1,pt,2,,pt,r]p_{t}=[p_{t,1},p_{t,2},...,p_{t,r}] (6)
Qt=i=1rpt,iqt,iQ_{t}=\sum_{i=1}^{r}{p_{t,i}*q_{t,i}} (7)

Where ata_{t} denoted the attention vector and it can be calucated by the attention module. σ\sigma was the sigmoid function and the ptp_{t} represented the normalized value of ata_{t}. Finally, the QtQ_{t} was calculated by summing original qtq_{t} with attentional coefficient.

Refer to caption

Figure 3: Attention Module Structure. The gated affined unit qiq_{i} of receptive-field rir_{i} would produce two feature vector through maxpooling and averagepooling and the two vectors were inputted 2-layer FC to produce the attentional weights. The final attentional feature matrix can be obtained through (5)-(8)

2.3 Augmented Classifier

Through the adaptive multiple receptive-field attention block, the scaled feature matrix QtQ_{t} was obtained and QtQ_{t} would be reshaped into a feature vector mtm_{t}. The Bi-LSTM is skilled in learning contextual speech information and the DNN is an excellent classifier. Nextly, the feature sequence Mt=[m1,m2,,mT]M_{t}=[m_{1},m_{2},...,m_{T}] would be fed into a two-layer Bi-LSTM and 1-layer Fully Connected DNN to make the final classification.

The training of MLNET-based VAD could be regarded as a common supervised optimized problems with traditional cross-entropy loss function.

crossentropy=t=0t=T{ytlog(y^t)+(1yt)log(1y^t)}\mathcal{L}_{crossentropy}=\sum_{t=0}^{t=T}{\{y_{t}log(\hat{y}_{t})+(1-y_{t})log(1-\hat{y}_{t})\}} (8)

Because of the characteristic of the proposed model, this paper further investigated an additional attention loss function to adapt the attention mechanism in the training phase. The attention loss function was designed to emphasize the most important receptive field of rkr_{k}, and the definitions were showed in (9)-(11).

k=argmax𝑖(pt,i)k=\underset{i}{argmax}(p_{t,i})\\ (9)
attention=t=0t=Ti=0i=r(yt,ilog(pt,i))\mathcal{L}_{attention}=\sum_{t=0}^{t=T}\sum_{i=0}^{i=r}(y_{t,i}log(p_{t,i})) (10)
=crossentropy+attention\mathcal{L}=\mathcal{L}_{crossentropy}+\mathcal{L}_{attention} (11)

Where attention\mathcal{L}_{attention} could be denoted as a softmax problem and the most appropriate receptive-field pt,kp_{t,k} was the target and the corresponding yt,ky_{t,k} is 11 and other yt,i=0(ik)y_{t,i}=0({i\not=k}).

3 Experiments

3.1 Datasets and Evaluations

In this paper, the English corpus of Aurora4 [29] and the Chinese corpuses of Thchs30 [30] were applied to train and test the proposed model. In our experiments, firstly, because of imbalance of speech and non-speech, 2-second-long silence segments were added into forward and backward of each utterance. In training, the clean speech corpus were corrupted by public 100 noise types of HuNonspeech111http://web.cse.ohio-state.edu/pnl/corpus/HuNonspeech/. Each utterance was randomly corrupted at a level of -5dB-20dB SNR and all have an average 7.5dB SNR. The NOISEX-92222http://spib.rice.edu/spib/select_noise.html noise dataset was used to construct testing dataset and 4 types unseen noises of babble, factory, destory-engine were selected to corrupt the clean speech. Also, the SNR was setted between -5dB and 20dB and the average is 7.5dB. For Aurora4, 95% of training data were used for training and 5% were used as dev data. The testing corpus of Aurora4 were corrupted by the NOISEX-92 and leveraged as the testing data. For thchs30 data, the dataset was constructed with the same process, but the different was that dev dataset leveraged the original corresponding data.

For comparison, the metrics of F1-score and DCF were selected as a performance measurement. F1-score took into account both accuracy and recall metrics, which was a common evaluation index of binary classification problems. DCF333http://fearlesssteps.exploreapollo.org/ reflected the wrong performance of the model and DCF was defined as followed:

F1score(θ)=2TP2TP+FP+FNF1-score(\theta)=\frac{2TP}{2TP+FP+FN} (12)
DCF(θ)=0.75×PFN(θ)+0.25×PFP(θ)DCF(\theta)=0.75\times P_{FN}(\theta)+0.25\times P_{FP}(\theta) (13)

where θ\theta denoted a given system decision-threshold setting. TP represented true positive examples’ num while FP and FN represented the num of false positive and false negative examples. PFNP_{FN} was the probability of FP while PFNP_{FN} was the probability of FN. It should be noted that the larger the F1-score was, the better performance while the smaller the DCF was, the better performance. In testing, we calculated the two metrics of each recording respectively and averaged the metrics of all recordings as the final score.

Table 1: Model Configuration
Name Unit
#Attention module
shared 2-layer FC 64×\times64
#Main network
Affine Matrix 40×(2[1,3,5,7,9]+1)\times(2*[1,3,5,7,9]+1)
40×(2[1,3,5,7,9]+1)\times(2*[1,3,5,7,9]+1)
Attention module 1
2-layer Bi-LSTM (64 + 64) ×\times (64 + 64)
1-layer FC 64

3.2 MLNET Setup

The acoustic feature extracted for MLNET was 40-dimensional log-mel filterbank while the frame size was 25 ms with a shift of 10 ms. The window of attention block were setted as 19 frames, which corresponded to 190ms contextual information, while the gated affined unit’s receptive-field were setted 1, 3, 5, 7 ,9. Other parameters of MLNET were shown in Table1. For training our proposed model, the matrix weight parameters of MLNET were all initialized with random uniform initialization while the bias parameters were initialized with a constant of 0.1. In our experiments, we trained the network for 150 epoches with the Adam algorithm when the loss function got little change. The batchsize was 32 and the learning rate was set to 0.001. When training, the gradient cropping strategy was also applied and the gradient of each parameter at each iteration was limited between -1 and 1.

To demonstrate the effectiveness of our proposed model, three VAD approaches were used for comparison. The first was google’ WebRTC VAD systems [31]. Additionally, Vafeiadis proposed 2-D CRNN model and has made great success in recent speech activity detection [32]. Nextly, ACAM approach, which was the state of art of attention-based methods, was also included [23]. In our experiments, we made use of the same parameters, but all trick strategies, such as batch normalization and regularization, were not leveraged. To alleviate the effect of the input features, it should be noted that 40-log mel acoustic features were leveraged to establish the CRNN and ACAM baseline models, which were different from the original approaches.

Table 2: Result Comparision of Aurora4
Name Dev Eval
#F1-score
Google VAD (mode 0) 72.33 76.32
CRNN 89.14 87.23
ACAM 90.56 89.03
MLNET 91.38 89.27
#DCF
Google VAD (mode 0) 22.06 18.34
CRNN 9.23 10.34
ACAM 8.95 9.01
MLNET 8.77 9.23
Table 3: Result Comparision of Thchs30
Name Dev Eval
#F1-score
Google VAD (mode 0) 74.71 74.60
CRNN 91.35 89.90
ACAM 92.53 91.27
MLNET 93.25 92.58
#DCF
Google VAD (mode 0) 17.88 18.18
CRNN 8.21 9.72
ACAM 7.67 8.51
MLNET 6.89 8.12

3.3 Results and Discussions

The results were summarized in Table 2 and Table 3. We can observed that MLNET has the best performance and outformed other three baselines models. Especially, three-training-based methods achieved 10% higher than google’s VAD in F1-score and 8% lower in DCF. The ACAM and MLNET outformed the traditional deep learning methods of CRNN, which proved attention-based models were helpful for improving the accuracy of detection. Comparing with ACAM, MLNET selected the windows’ attentions instead of ACAM’s frames-based attention and experiments showed that the window based attention models achieved higher performance than the frames-based models, which also conformed to our prior knowledge that current frame information was the most important and the joined frames were auxiliary information when predicting the current frame.

In order to illustrate each part’s functionality of our proposed model, the comparison of each modules were further investigated. The base was the Bi-LSTM, which just leveraged the contextual speech information and the feature vectors of contextual speech were aggregated to a longer vector before feeding into the network. The second was the gated unit model that the gated unit operation replaced the aggregated mechanism. The third and the fourth were leveraged to certificate the multiple window’s functions while the attention’s function was also compared. The Aurora4 dataset were leveraged to make this evaluation and the results were shown in Table 4. As noted in this table, the gated affined unit based models have better performance than direct aggregation of Bi-LSTM and achieved about 2% increase in F1-score and 1.5% decrease in DCF. Adding the multiple receptive-field of non-attention, the VAD’s performance was also improved and achieved about 2.5% increase in F1-score and 2.13 % decrease in DCF. Lastly, the adaptive attention mechanism was also helpful for MLNET’s performance. In contrast, the adaptive attention only achieved a small accuracy improvement than non-attention module and the reason may be the mechanism of receptive-field selection has existed in multiple receptive-field attention block. In particular, we observed that the adaptive multiple receptive-field attention block was also helpful for speeding up models’ convergence in our experiments. To sum up, the proposed method can better deal with the VAD problems.

Table 4: Module Comparision
Model Dev Eval
#F1-score
Bi-LSTM 85.89 84.17
+Gated Unit 87.63 86.24
+Non-Attention 90.85 88.73
+Attention 91.38 89.27
#DCF
Bi-LSTM 12.24 13.54
+Gated Unit 11.16 12.01
+Non-Attention 9.11 9.88
+Attention 8.77 9.23

4 Conclusion

The existed DNN-based VAD models only leveraged fixed receptive-field contextual speech information and were unable to handle with speech segments of different lengths adaptively. To overcome this defect, this paper proposed an architecture of MLNET for VAD task. MLNET made use of different gated affined unit to extract different contextual speech information and leveraged the adaptive attention block to select the most focused speech segments. Comparing with the existed models, the experiments demonstrated that MLNET outformed other baseline models and proved that the proposed architecture was helpful to deal with VAD problems.

5 ACKNOWLEDGEMENTS

This paper is supported by National Key Research and Development Program of China under grant No. 2018YFB1003500, No. 2018YFB0204400 and No. 2017YFB1401202. Corresponding author is Jianzong Wang from Ping An Technology (Shenzhen) Co., Ltd.

References

  • [1] M. W. Mak and H. B. Yu, “A study of voice activity detection techniques for nist speaker recognition evaluations,” in Computer Speech & Language, 2014, p. 295–313.
  • [2] J. Ramirez, J. M. Gorriz, and J. C. Segura, Voice Activity Detection. Fundamentals and Speech Recognition System Robustness.   InTech, 2007.
  • [3] T. Yoshimura, T. Hayashi, K. Takeda, and S. Watanabe, “End-to-end automatic speech recognition integrated with ctc-based voice activity detection,” in ICASSP, 2020.
  • [4] B. Sharma, R. K. Das, and H. Li, “Multi-Level Adaptive Speech Activity Detector for Speech in Naturalistic Environments,” in Interspeech, 2019.
  • [5] F. M. G. D. P. Mainar and M. Cernak, “Spiking Neural Networks Trained With Backpropagation for Low Power Neuromorphic Implementation of Voice Activity Detection,” in ICASSP, 2020.
  • [6] G. D. F. Martinelli and M. Cernak, “A Bin Encoding Training of a Spiking Neural Network Based Voice Activity Detection,” in ICASSP, 2020.
  • [7] X. Yang, B. Tan, J. Ding, J. Zhang, and J. Gong, “Comparative study on voice activity detection algorithm,” in International Conference on Electrical and Control Engineering, 2010.
  • [8] K. W. Ho, Y. T. Young, P. K. Jung, and L. Chungyong, “Robust voice activity detection algorithm for estimating noise spectrum,” Electronics Letters, vol. 36, no. 2, pp. 180–181, 2002.
  • [9] S. Jongseo, S. N. Kim, and W. Sung, “A statistical model based voice activity detector,” IEEE Signal Processing Letters, vol. 6, no. 1, pp. 1 – 3, 1999.
  • [10] D. Thomas, S. Yannis, K. Yusuke, and A. Masami, “Voice activity detection: Merging source and filter-based information,” in IEEE Signal Processing Letters.   IEEE, 2016, pp. 252–256.
  • [11] T. Ng, B. Zhang, and N. L, “Developing a speech activity detection system for the darpa rats program,” in Interspeech, 2012.
  • [12] V. H and S. H, “Hidden-markov-model-based voice activity detector with high speech detection rate for speech enhancement,” Iet Signal Processing, vol. 6, no. 1, pp. 54–63, 2012.
  • [13] E. Dong, G. Liu, Y. Zhou, and X. Zhang, “Applying support vector machines to voice activity detection,” in International Conference on Signal Processing Proceedings.   IEEE, 2003.
  • [14] S. Tong, H. Gu, and K. Yu, “A comparative study of robustness of deep learning approaches for vad,” in ICASSP, 2016.
  • [15] X. Zhang and D. Wang, “Boosting contextual information for deep neural network based voice activity detection,” IEEE/ACM Transactions on Audio Speech and Language Processing, vol. 24, no. 2, pp. 252–264, 2016.
  • [16] Z. C. Fan, Z. Bai, X.-L. Zhang, S. Rahardja, and J. Chen, “Auc optimization for deep learning based voice activity detection,” in ICASSP, 2019.
  • [17] A. Ivry, I. Cohen, and B. Berdugo, “Evaluation of deep-learning-based voice activity detectors and room impulse response models in reverberant environments,” in ICASSP, 2020.
  • [18] A. Ivry, B. Berdugo, and I. Cohen, “Voice activity detection for transient noisy environment based on diffusion nets,” IEEE Journal of Selected Topics in Signal Processing, vol. 13, no. 2, pp. 254–264, 2019.
  • [19] X. Zhang and J. Wu, “Deep belief networks based voice activity detection,” IEEE Transactions on Audio Speech and Language Processing, vol. 21, no. 4, pp. 697–710, 2013.
  • [20] I. Ariav and I. Cohen, “An end-to-end multimodal voice activity detection using wavenet encoder and residual networks,” IEEE Journal of Selected Topics in Signal Processing, pp. 1–1, 2019.
  • [21] S. Y. Chang, B. Li, G. Simko, T. N. Sainath, A. Tripathi, A. van den Oord, and O. Vinyals, “Temporal modeling using dilated convolution and gating for voice-activity-detection,” in ICASSP, 2018.
  • [22] H. Thad, Mierle, and Keir, “Recurrent neural networks for voice activity detection,” in ICASSP, 2013, pp. 7378–7382.
  • [23] K. Juntae and H. Minsoo, “Voice activity detection using an adaptive context attention model,” IEEE Signal Processing Letters, vol. PP, no. 99, pp. 1–1, 2018.
  • [24] R. Lin, C. Costello, C. Jankowski, and V. Mruthyunjaya, “Optimizing Voice Activity Detection for Noisy Conditions,” in Interspeech 2019, pp. 2030–2034.
  • [25] Y. Jung, Y. Kim, Y. Choi, and H. Kim, “Joint learning using denoising variational autoencoders for voice activity detection.” in Interspeech, 2018, pp. 1210–1214.
  • [26] A. van den Oord, N. Kalchbrenner, O. Vinyals, L. Espeholt, A. Graves, and K. Kavukcuoglu, “Conditional image generation with pixelcnn decoders,” CoRR, 2016. [Online]. Available: http://arxiv.org/abs/1606.05328
  • [27] J. Hu, L. Shen, and G. Sun, “Squeeze-and-excitation networks,” in CVPR, 2018, pp. 7132–7141.
  • [28] S. Woo, J. Park, J. Y. Lee, and I. So Kweon, “Cbam: Convolutional block attention module,” in ECCV, 2018, pp. 3–19.
  • [29] N. Parihar and J. Picone, Aurora working group: DSR front end LVCSR evaluation AU/384/02.   Inst. for Signal and Information Process, Mississippi State University, 2002.
  • [30] Z. Z. Dong Wang, Xuewei Zhang, “Thchs-30 : A free chinese speech corpus,” 2015. [Online]. Available: http://arxiv.org/abs/1512.01882
  • [31] GoogleWebRTC, “https://webrtc.org/,” 2016.
  • [32] A. Vafeiadis, E. Fanioudakis, I. Potamitis, K. Votis, D. Giakoumis, D. Tzovaras, L. Chen, and R. Hamzaoui, “Two-Dimensional Convolutional Recurrent Neural Networks for Speech Activity Detection,” in Interspeech, 2019.