This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Singing Voice Synthesis Based on a Musical Note Position-aware
Attention Mechanism

Abstract

This paper proposes a novel sequence-to-sequence (seq2seq) model with a musical note position-aware attention mechanism for singing voice synthesis (SVS). A seq2seq modeling approach that can simultaneously perform acoustic and temporal modeling is attractive. However, due to the difficulty of the temporal modeling of singing voices, many recent SVS systems with an encoder-decoder-based model still rely on explicitly on duration information generated by additional modules. Although some studies perform simultaneous modeling using seq2seq models with an attention mechanism, they have insufficient robustness against temporal modeling. The proposed attention mechanism is designed to estimate the attention weights by considering the rhythm given by the musical score. Furthermore, several techniques are also introduced to improve the modeling performance of the singing voice. Experimental results indicated that the proposed model is effective in terms of both naturalness and robustness of timing.

Index Terms—  Singing voice synthesis, sequence-to-sequence model, attention mechanism, temporal modeling

1 Introduction

Statistical parametric singing voice synthesis (SVS) have been evolving with the spread of machine learning techniques [1, 2, 3, 4, 5, 6]. The essence of SVS is a sequence transform to generate an acoustic feature sequence from a score feature sequence obtained from musical scores. Since the singing voice needs to be strictly synchronized with the given musical score, how to model the temporal structure of the singing voice is one of the important issues in the SVS.

Typical deep neural network (DNN)-based SVS systems [2, 3, 4, 5] model the acoustic feature and its temporal structure (specifically as phoneme duration and vocal timing deviation) independently using acoustic, duration, and time-lag models. The acoustic model works as a mapping function that generates the acoustic feature sequence from the time-aligned score feature sequence. Although these pipeline systems are stable when running in both training and synthesis, they suffer from alignment-related issues: 1) the modeling performance of acoustic features is affected by alignment errors, and 2) they do not have the sufficient ability to adequately model the correlation between the acoustic feature and its temporal structure. These errors may cause a lack of naturalness and expressiveness of synthesized singing voices.

Inspired by the success of autoregressive (AR) and non-AR generation models for modern text-to-speech (TTS) systems [7, 8, 9, 10, 11, 12, 13, 14], several neural SVS systems have been developed [15, 16, 17, 18, 19, 20, 21, 22, 23]. Unlike TTS, it is not easy to model the temporal structure of a singing voice because the phoneme duration of singing voices varies greatly depending on the corresponding note duration, even for the same phoneme. Thus, these kinds of SVS systems [15, 16, 17, 18, 19, 20, 21] mainly adopt encoder-decoder models with an explicit length regulator. This approach has the advantage of robustness in terms of alignment while it makes the model sensitive to the performance of external duration information, similar to conventional pipeline systems. A number of studies [22, 23] use an AR sequence-to-sequence (seq2seq) model with an attention mechanism to model the acoustic feature and its temporal structure simultaneously; however, they suffer from timing mismatches between the target musical score and synthesized singing voice. This is a critical issue for SVS applications because manually collecting the timing in these attention-based systems is both difficult and impractical

In this paper, we propose a novel seq2seq model for SVS with a musical note position-aware attention mechanism. The proposed attention mechanism calculates the attention weights based on the note position informed by the musical score. Moreover, we also introduce additional techniques to help obtain robust alignment and improve naturalness: auxiliary note feature embedding, guided attention loss with a penalty matrix designed for singing voice alignments, and pitch normalization for the seq2seq model. The proposed seq2seq model can synthesize singing voices with proper vocal timing without extra supplementary temporal modeling.

2 Related work

Seq2seq modeling using an attention mechanism [24] is a key technique for modeling the correspondence between sequences with different lengths. In the TTS scenario, since the monotonicity and locality properties of TTS alignment can be exploited, hybrid location-sensitive attention combining content-based and location-based attention mechanisms [9], a forward attention mechanism computed recursively using a recursively forward algorithm [10], and monotonic attention mechanisms [25, 26] have been developed.

One of the noticeable differences between SVS and other typical seq2seq tasks is that the alignment path is strongly related to the note timing. The authors [23] have been proposed a global duration control attention, introducing a global transition token into a forward attention mechanism. Although the tempo of singing voices can be controlled through its token, a time-invariant token has the insufficient ability for duration control and cannot prevent the mismatch between the timing of synthesized singing voices and musical scores. The proposed musical note position-aware attention mechanism is based on the generalized form of forward attention with phoneme-dependent and time-variant transition probabilities, and the attention probability is calculated taking into account note position information via note position embeddings. This achieves appropriate modeling of the temporal structure of the singing voice by the attention mechanism.

Refer to caption
Fig. 1: Overview of the proposed model.

3 Proposed model

The proposed model is based on an encoder-decoder model with an attention mechanism and can generate frame-level acoustic feature sequences directly from phoneme-level score feature sequences. The score feature consists of not only the phone, note pitch, and note length but also other rich musical contexts such as beat, key, tempo, dynamics, and staccato [1]. An overview of the proposed model is shown in Fig. 1. We describe four techniques in this section for seq2seq SVS systems to satisfy the singing-specific requirements, such as high controllability and robustness against tempo and pitch.

3.1 Musical note position-aware attention mechanism

The attention mechanism [24] calculates an attention weight at each decoder time-step to perform a soft-selection of encoder hidden state 𝒙=[𝒙1,𝒙2,𝒙N]\bm{x}=[\bm{x}_{1},\bm{x}_{2},\ldots\bm{x}_{N}], which is obtained by processing the input score features by the encoder. The context vector 𝒄t\bm{c}_{t} can be obtained by 𝒄t=n=1Nαt(n)𝒙n\bm{c}_{t}=\sum_{n=1}^{N}\alpha_{t}(n)\bm{x}_{n}, where αt(n)\alpha_{t}(n) is the attention weight representing the degree of attention on the nn-th hidden state at the tt-th decoder time-step. Finally, the output vector 𝒐t\bm{o}_{t} can be the predicted conditioning of the context vector 𝒄t\bm{c}_{t} by the decoder.

The alignment obtained by the attention mechanism must be monotonic and continuous without skipping any encoder states to synthesize a singing voice following a given musical score. To satisfy this assumption, a current attention weight αt(n)\alpha_{t}(n) is calculated recursively using the previous alignment as follows:

αt(n)\displaystyle\alpha^{\prime}_{t}(n) =((1ut1(n))αt1(n)\displaystyle=\big{(}(1-u_{t-1}(n))\alpha_{t-1}(n)
+ut1(n1)αt1(n1))yt(n),\displaystyle\qquad\quad+u_{t-1}(n-1)\alpha_{t-1}(n-1)\big{)}\cdot y_{t}(n), (1)
αt(n)\displaystyle\alpha_{t}(n) =αt(n)/m=1Nαt(m),\displaystyle=\alpha^{\prime}_{t}(n)\biggl{/}\,\sum_{m=1}^{N}\alpha^{\prime}_{t}(m), (2)

where yt(n)y_{t}(n) is the output probability, and ut(n)u_{t}(n) is the transition probability where the attention mechanism notices the nn-th phoneme at the tt-th decoder step and moves to the n+1n+1-th phoneme at the t+1t+1-th decoder steps. Note that ut(n)u_{t}(n) in Eq. (1) is the phoneme-dependent time-variant transition probability; thereby this formula can be regarded as a generalized form of forward attention with a transition agent [10].

The alignment of SVS should be obtained by considering the temporal structure of musical notes, specifically note duration and tempo. Therefore, we introduce musical note positional features to compute yt(n)y_{t}(n) and ut(n)u_{t}(n).

In the proposed method, the output probability yt(n)y_{t}(n) is calculated by extended content-based attention with a musical note position-aware additional term 𝑼()𝒑t,n\bm{U}^{(\cdot)}\bm{p}_{t,n} as:

et(n)\displaystyle e_{t}(n) =𝒗(e)tanh(𝑾(e)𝒒t+𝑽(e)𝒙n+𝑼(e)𝒑t,n+𝒃(e)),\displaystyle=\bm{v}^{(e)}\!{}^{\top}\!\tanh(\bm{W}^{(e)}\bm{q}_{t}+\bm{V}^{(e)}\bm{x}_{n}+\bm{U}^{(e)}\bm{p}_{t,n}+\bm{b}^{(e)}), (3)
yt(n)\displaystyle y_{t}(n) =exp(et(n))/m=1Nexp(et(m)),\displaystyle=\exp\left(e_{t}(n)\right)\biggl{/}\,\sum_{m=1}^{N}\exp\left(e_{t}(m)\right), (4)

where 𝒒t\bm{q}_{t} denotes the tt-th time-step decoder hidden state, and 𝑾()𝒒t\bm{W}^{(\cdot)}\bm{q}_{t} and 𝑽()𝒙n\bm{V}^{(\cdot)}\bm{x}_{n} are the content-based terms that represent query/key comparisons in the attention mechanism, and 𝒃()\bm{b}^{(\cdot)} is the bias term. The note position embedded feature 𝒑t,n\bm{p}_{t,n} is obtained by feeding the note position representation [pt,n1,pt,n2,pt,n3][p^{1}_{t,n},p^{2}_{t,n},p^{3}_{t,n}] with a single tanh hidden layer. Each note position representation is computed from the note lengths of the given musical score as follows:

pt,n1\displaystyle p^{1}_{t,n} =tsn,\displaystyle=t-s_{n}, (5)
pt,n2\displaystyle p^{2}_{t,n} =ent,\displaystyle=e_{n}-t, (6)
pt,n3\displaystyle p^{3}_{t,n} ={snt,(t<sn)0,(snten)ten,(en<t)\displaystyle=\left\{\begin{aligned} &s_{n}-t,&\quad(t<s_{n})\\ &0,&\quad(s_{n}\leq t\leq e_{n})\\ &t-e_{n},&\quad(e_{n}<t)\\ \end{aligned}\right. (7)

where sns_{n} and ene_{n} denote the start and end positions of the nn-th musical note, respectively.

Since transition probabilities should be explicitly derived from past alignments, we adopt a location-sensitive attention [9]-like formula to calculate the transition probability as follows:

ut(n)=σ(𝒗(u)tanh(𝑾(u)𝒒t\displaystyle u_{t}(n)=\sigma\bigl{(}\bm{v}^{(u)}\!{}^{\top}\!\tanh(\bm{W}^{(u)}\bm{q}_{t} +𝑽(u)𝒙n+𝑼(u)𝒑t,n\displaystyle+\bm{V}^{(u)}\bm{x}_{n}+\bm{U}^{(u)}\bm{p}_{t,n}
+𝑻(u)𝒇t,n+𝒃(u))),\displaystyle+\bm{T}^{(u)}\bm{f}_{t,n}+\bm{b}^{(u)})\bigr{)}, (8)

where σ()\sigma(\cdot) is a logistic sigmoid function, and 𝑻(u)𝒇t,n\bm{T}^{(u)}\bm{f}_{t,n} denotes the location-sensitive term that uses convolutional features computed from the previous cumulative alignments [27].

3.2 Auxiliary note feature embedding

As the temporal structure of the singing voices depends on the context of the notes in the input score, the alignment of singing voices should also be predicted on the basis of it. To encourage this, we embed musical note context associated with the current note as an auxiliary note feature into the attention query.

The auxiliary note features contained musical note-related contexts, which were obtained by removing phoneme and mora-related contexts from the score features, and were expanded to the frame-level sequence using note length. This upsampled feature is then fed to a single dense layer, which is concatenated with the output of Pre-Net to form the decoder input. Since the auxiliary note feature delivers the attention mechanism to the current frame position and note context in the corresponding musical note via the attention query, it is expected to enable the attention mechanism to adjust the alignment to fit the rhythm provided by the musical score.

Refer to caption
Fig. 2: Penalty matrix examples. This matrix is generated on the basis of pseudo-determined mora boundaries from note durations described in the musical score.

3.3 Guided attention loss for SVS

Since singing voices are generally sung to follow the rhythm of a musical score, it is natural to assume that the alignment of the singing voice should be close to the path determined by the note timing in the score. On the basis of this idea, we customize the guided attention loss [28] for SVS systems.

The difference from the original guided attention loss for TTS is how to construct the penalty matrix. We generate a penalty matrix based on the note boundaries, unlike the one for the TTS whose diagonal elements are zero. Specifically, we design a penalty matrix 𝑮N×T\bm{G}\in\mathbb{R}^{N\times T} based on the pseudo-determined mora boundary so that alignment estimation is robust even when multiple morae are included in the same note, as shown in Fig. 2. These pseudo-boundaries are obtained by equally dividing note duration in accordance with the number of morae in each note. The soft matrix is generated by linearly decaying over a width of 60 frames. Note that the start positions of boundaries for constructing the penalty matrix are shifted 15 frames earlier to consider vocal timing deviation.

Let the alignment matrix 𝑨N×T\bm{A}\in\mathbb{R}^{N\times T} be a matrix, where the element at (i,j)(i,j) corresponds to αj(i)\alpha_{j}(i) and TT is the total number of frames. The guided attention loss is defined as follows:

att(𝑮,𝑨)\displaystyle\mathcal{L}_{\mathrm{att}}(\bm{G},\bm{A}) =1NT𝑮𝑨1,\displaystyle=\frac{1}{NT}\left\lVert\bm{G}\odot\bm{A}\right\rVert_{1}, (9)

where \odot denotes an element-wise product. The final loss function \mathcal{L} of the proposed model is given by

=feat(𝒐,𝒐^)+feat(𝒐,𝒐^)+λatt(𝑮,𝑨),\displaystyle\mathcal{L}=\mathcal{L}_{\mathrm{feat}}(\bm{o},\hat{\bm{o}})+\mathcal{L}_{\mathrm{feat}}(\bm{o},\hat{\bm{o}}^{\prime})+\lambda\mathcal{L}_{\mathrm{att}}(\bm{G},\bm{A}), (10)
feat(𝒐,𝒐^)=1TDt=1T𝒐𝒐^22,\displaystyle\mathcal{L}_{\mathrm{feat}}(\bm{o},\hat{\bm{o}})=\frac{1}{TD}\sum_{t=1}^{T}\left\lVert\bm{o}-\hat{\bm{o}}\right\rVert^{2}_{2}, (11)

where 𝒐^\hat{\bm{o}} and 𝒐^\hat{\bm{o}}^{\prime} represent the acoustic features predicted by the decoder and Post-Net, respectively, and DD is the number of dimensions in the acoustic features. In Eq. (10), λ\lambda represents an adjustment parameter for guided attention loss.

3.4 Pitch normalization

The pitch of the synthesized singing voice must accurately follow the note pitch of the musical score. Thus, following our previous work [2], we integrate pitch normalization, where the log fundamental frequency (F0F_{0}) is modeled as the difference from the log F0F_{0} determined by the musical score (note pitch), into the proposed model.

Pitch normalization requires a time-aligned frame-level note pitch sequence to process the generated F0F_{0} sequence frame-by-frame. In the proposed system, the frame-level note pitch sequence can be obtained by weighting the phone-level input note pitch sequence using the attention weight at each decoder time-step.

4 Experiments

4.1 Experimental conditions

To evaluate the proposed models, we conducted experiments using 70 Japanese children’s songs (total: 70 min) performed by one female singer. Sixty songs were used for training, and the rest were used for testing. Singing voice signals were sampled at 48 kHz, and each sample was quantized by 16 bits. The acoustic feature consisted of 0-th through 49-th mel-cepstral coefficients, a continuous log F0F_{0} value, 25-dimensional analysis aperiodicity measures, 1-dimensional vibrato component, and a voiced/unvoiced binary flag. Mel-cepstral coefficients were extracted by WORLD [29]. The difference between the original log F0F_{0} and the median-smoothed log F0F_{0} were used as the vibrato component.

The model architectures of encoder, decoder, Pre-Net, and Post-Net are the same as those of [9]. A linear projection layer was used instead of the embedding layer. We added an extra linear projection layer to process the frame-level auxiliary note feature sequence. The score feature was a 267-dimensional feature vector. The auxiliary note feature was an 87-dimensional feature vector. The frame position in the current note and the note duration were concatenated with an expanded frame-level auxiliary note feature. For all systems, the reduction factor was set to 33, and pitch normalization was applied for F0F_{0} modeling. The hyperparameter λ\lambda in Eq. (10) was set to 10.010.0. All systems were combined with the same pre-trained PeriodNet [30], a non-AR neural vocoder with a parallel structure, to reconstruct waveforms from predicted acoustic features.

4.2 Subjective evaluation

Table 1: Results of experiment 1 with 95% confidence intervals.
Systems 𝒑t,n\bm{p}_{t,n} Aux.​ note feat. att\mathcal{L}_{\mathrm{att}} MOS
(Sec.​ 3.1) (Sec.​ 3.2) (Sec.​ 3.3)
Base failed
NF \checkmark failed
NP \checkmark 3.12±0.133.12\pm 0.13
NP+NF \checkmark \checkmark 3.81±0.123.81\pm 0.12
Prop \checkmark \checkmark \checkmark 3.95 ±\pm 0.12

We performed 5-scale mean opinion score (MOS) tests111Audio samples are available at the following URL: https://www.sp.nitech.ac.jp/~hono/demos/icassp2023/ to evaluate the naturalness of the synthesized singing voices. Each of the 15 native Japanese-speaking participants evaluated ten phrases randomly selected from the test songs. A click sound generated based on the basis of the tempo of the score was superimposed on the synthesized singing voice to evaluate the overall naturalness considering the vocal timing.

4.2.1 Experiment 1

Refer to caption
((a)) Base
Refer to caption
((b)) NF
Refer to caption
((c)) NP
Refer to caption
((d)) NP+NF
Refer to caption
((e)) NoTrans
Refer to caption
((f)) P-Trans
Refer to caption
((g)) T-Trans
Refer to caption
((h)) Prop
Fig. 3: Examples of alignments obtained by each system. White dashed lines represent note boundaries.
Refer to caption
((a)) P-Trans
Refer to caption
((b)) T-Trans
Refer to caption
((c)) Prop
Fig. 4: Examples of output probabilities yt(n)y_{t}(n) (left column) and transition probabilities ut(n)u_{t}(n) (right column) in P-Trans, T-Trans, and Prop. White dashed lines represent note boundaries.

We first evaluated the effectiveness of the proposed techniques for modeling the temporal structure of the singing voice described in Section 3. We used five systems as shown in Table 1. In this experiment, the attention weights were calculated following Eqs. (1) and (2). In Base and NF, we calculated the output and transition probabilities without using 𝑼()𝒑t,n\bm{U}^{(\cdot)}\bm{p}_{t,n} in Eqs. (4) and (8).

Table 1 shows the subjective evaluation results, and Fig. 3 shows a number of examples of the alignments predicted by each system. From Fig. 3, Base and NF cannot obtain an appropriate alignment of the singing voice, which indicates the proposed musical note position-aware attention mechanism is effective. Since synthesized samples by Base and NF contained many timing and lyrics errors, they were excluded from the MOS test. NP+NF achieved much higher MOS scores than NP. This shows that the use of auxiliary note features is effective. The figure also shows that Prop can obtain a more monotonic alignment and better MOS scores than NP+NF and thus demonstrating the effectiveness of the proposed guided attention loss described in Section 3.3. Note that a previous study [22] reported that a simple attention model similar to Base could synthesize singing voices; however, this required 70 hours of singing voice data with data augmentation. In contrast, our proposed system does not require large training data or data augmentation approaches.

4.2.2 Experiment 2

Table 2: Results of experiment 2 with 95% confidence intervals.
Attention mechanism Transition probability
Systems Phoneme- dependent Time- variant MOS
NoAtt 3.61±0.133.61\pm 0.13
NoTrans \checkmark 3.78±0.123.78\pm 0.12
P-Trans \checkmark \checkmark 3.83±0.133.83\pm 0.13
T-Trans \checkmark \checkmark 2.16±0.122.16\pm 0.12
Prop \checkmark \checkmark \checkmark 3.87 ±\pm 0.12

We performed another experiment focusing on the attention mechanisms using the five systems listed in Table 2. Note that Prop in Tables 1 and 2 denote the same systems. In NoAtt, instead of an attention mechanism, the encoder hidden state is time-aligned by using forced alignment at the training stage and estimating phoneme boundaries using DNN duration and time-lag models at the synthesis stage. Five-state, left-to-right, no-skip hidden semi-Markov models (HSMMs) were used to obtain forced alignment [31]. NoTrans is the system where the transition probability is fixed at ut(n)=0.5u_{t}(n)=0.5, P-Trans is the system where the transition probability is only phoneme-dependent, and T-Trans is the system where transition probability is time-variant regardless of the phonemes. The other conditions of NoTrans, P-Trans and T-Trans are the same as those in Prop. The attention mechanism of P-Trans is similar to [23] in exploiting the phoneme-dependent transition probability.

Table 2 and Fig. 3 show the subjective evaluation results and examples of the alignments predicted by each system, respectively. P-Trans had a slightly better score than NoTrans. As shown in Fig. 3(f), the alignment obtained by P-Trans increased monotonically; however, it can be seen from Fig. 4(a) that estimating the alignments is highly dependent on the output probabilities. On the other hand, as shown in Figs. 3(g) and 4(b), the alignment of T-Trans readily undergoes transitions, which causes consonant skipping. As the attention mechanism of T-Trans cannot model temporal structures appropriately, T-Trans had the lowest MOS scores in the subjective evaluation. In contrast, Prop achieved the best MOS score. Figures 3(h) and 4 show that the transition probability of Prop works more adequately than others, which leads to obtaining the appropriate alignment. Furthermore, Prop achieved a better MOS score than NoAtt. This indicates the effectiveness of the proposed method whereby acoustic features and temporal structures of the singing voice are modeled simultaneously using the seq2seq model with the proposed attention mechanism.

5 Conclusion

We proposed singing voice synthesis based on the seq2seq model with a musical note position-aware attention mechanism. Our proposed model can properly estimate the alignment of a singing voice while taking note positions into account and can generate acoustic feature sequence without additional temporal modeling. Experimental results show that the proposed model is effective for simultaneously modeling the acoustic feature sequence and its temporal structure. Future work includes introducing an HSMM-based structured attention mechanism [32] to improve the duration controllability and more robust alignment estimation.

6 Acknowledgments

The authors would like to thank Mr. Shumma Murata for joining the discussion and the experiment.

References

  • [1] K. Oura, A. Mase, T. Yamada, S. Muto, Y. Nankaku, and K. Tokuda, “Recent development of the HMM-based singing voice synthesis system–Sinsy,” in Proc. ISCA SSW7, 2010, pp. 211–216.
  • [2] Y. Hono, K. Hashimoto, K. Oura, Y. Nankaku, and K. Tokuda, “Sinsy: A deep neural network-based singing voice synthesis system,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 29, pp. 2803–2815, 2021.
  • [3] M. Blaauw and J. Bonada, “A neural parametric singing synthesizer,” in Proc. Interspeech, 2017, pp. 4001–4005.
  • [4] Y. Hono, K. Hashimoto, K. Oura, Y. Nankaku, and K. Tokuda, “Singing voice synthesis based on generative adversarial networks,” in Proc. ICASSP, 2019, pp. 6955–6959.
  • [5] K. Nakamura, S. Takaki, K. Hashimoto, K. Oura, Y. Nankaku, and K. Tokuda, “Fast and high-quality singing voice synthesis system based on convolutional neural networks,” in Proc. ICASSP, 2020, pp. 7239–7243.
  • [6] Y.-H. Yi, Y. Ai, Z.-H. Ling, and L.-R. Dai, “Singing voice synthesis using deep autoregressive neural networks for acoustic modeling,” in Proc. Interspeech, 2019, pp. 2593–2597.
  • [7] J. Sotelo, S. Mehri, K. Kumar, J. F. Santos, K. Kastner, A. Courville, and Y. Bengio, “Char2wav: End-to-end speech synthesis,” in Proc. ICLR Workshop Track, 2017.
  • [8] Y. Wang, R. Skerry-Ryan, D. Stanton, Y. Wu, R. J. Weiss, N. Jaitly, Z. Yang, Y. Xiao, Z. Chen, S. Bengio et al., “Tacotron: Towards end-to-end speech synthesis,” in Proc. Interspeech, 2017, pp. 4004–4010.
  • [9] J. Shen, R. Pang, R. J. Weiss, M. Schuster, N. Jaitly, Z. Yang, Z. Chen, Y. Zhang, Y. Wang, R. Skerrv-Ryan et al., “Natural TTS synthesis by conditioning WaveNet on mel spectrogram predictions,” in Proc. ICASSP, 2018, pp. 4779–4783.
  • [10] J.-X. Zhang, Z.-H. Ling, and L.-R. Dai, “Forward attention in sequence-to-sequence acoustic modeling for speech synthesis,” in Proc. ICASSP, 2018, pp. 4789–4793.
  • [11] N. Li, S. Liu, Y. Liu, S. Zhao, and M. Liu, “Neural speech synthesis with transformer network,” in Proc. AAAI Conference on Artificial Intelligence, vol. 33, 2019, pp. 6706–6713.
  • [12] Y. Ren, C. Hu, X. Tan, T. Qin, S. Zhao, Z. Zhao, and T.-Y. Liu, “FastSpeech 2: Fast and high-quality end-to-end text to speech,” in Proc. ICLR, 2021.
  • [13] J. Shen, Y. Jia, M. Chrzanowski, Y. Zhang, I. Elias, H. Zen, and Y. Wu, “Non-attentive Tacotron: Robust and controllable neural tts synthesis including unsupervised duration modeling,” arXiv preprint arXiv:2010.04301, 2020.
  • [14] I. Elias, H. Zen, J. Shen, Y. Zhang, Y. Jia, R. J. Weiss, and Y. Wu, “Parallel Tacotron: Non-autoregressive and controllable tts,” in Proc. ICASSP, 2021, pp. 5709–5713.
  • [15] J. Lee, H.-S. Choi, C.-B. Jeon, J. Koo, and K. Lee, “Adversarially trained end-to-end korean singing voice synthesis system,” in Proc. Interspeech, 2019, pp. 2588–2592.
  • [16] M. Blaauw and J. Bonada, “Sequence-to-sequence singing synthesis using the feed-forward transformer,” in Proc. ICASSP, 2020, pp. 7229–7233.
  • [17] P. Lu, J. Wu, J. Luan, X. Tan, and L. Zhou, “XiaoiceSing: A high-quality and integrated singing voice synthesis system,” in Proc. Interspeech, 2020, pp. 1306–1310.
  • [18] J. Chen, X. Tan, J. Luan, T. Qin, and T.-Y. Liu, “HiFiSinger: Towards high-fidelity neural singing voice synthesis,” arXiv preprint arXiv:2009.01776, 2020.
  • [19] J. Wu and J. Luan, “Adversarially trained multi-singer sequence-to-sequence singing synthesizer,” in Proc. Interspeech, 2020, pp. 1296–1300.
  • [20] Y. Gu, X. Yin, Y. Rao, Y. Wan, B. Tang, Y. Zhang, J. Chen, Y. Wang, and Z. Ma, “ByteSing: A chinese singing voice synthesis system using duration allocated encoder-decoder acoustic models and wavernn vocoders,” in Proc. ISCSLP, 2021, pp. 1–5.
  • [21] J. Shi, S. Guo, N. Huo, Y. Zhang, and Q. Jin, “Sequence-to-sequence singing voice synthesis with perceptual entropy loss,” in Proc. ICASSP, 2021, pp. 76–80.
  • [22] O. Angelini, A. Moinet, K. Yanagisawa, and T. Drugman, “Singing synthesis: with a little help from my attention,” in Proc. Interspeech, 2020, pp. 1221–1225.
  • [23] T. Wang, R. Fu, J. Yi, J. Tao, and Z. Wen, “Singing-Tacotron: Global duration control attention and dynamic filter for end-to-end singing voice synthesis,” arXiv preprint arXiv:2202.07907, 2022.
  • [24] D. Bahdanau, K. Cho, and Y. Bengio, “Neural machine translation by jointly learning to align and translate,” in Proc. ICLR, 2015.
  • [25] M. He, Y. Deng, and L. He, “Robust sequence-to-sequence acoustic modeling with stepwise monotonic attention for neural TTS,” in Proc. Interspeech, 2019, pp. 1293–1297.
  • [26] Y. Yasuda, X. Wang, and J. Yamagishi, “Initial investigation of encoder-decoder end-to-end tts using marginalization of monotonic hard alignments,” in Proc. ISCA SSW10, 2019, pp. 211–216.
  • [27] J. Chorowski, D. Bahdanau, D. Serdyuk, K. Cho, and Y. Bengio, “Attention-based models for speech recognition,” in Advances in Neural Information Processing Systems, 2015, pp. 577–585.
  • [28] H. Tachibana, K. Uenoyama, and S. Aihara, “Efficiently trainable text-to-speech system based on deep convolutional networks with guided attention,” in Proc. ICASSP, 2018, pp. 4784–4788.
  • [29] M. Morise, F. Yokomori, and K. Ozawa, “WORLD: A vocoder-based high-quality speech synthesis system for real-time applications,” IEICE Transactions on Information and Systems, vol. 99, no. 7, pp. 1877–1884, 2016.
  • [30] Y. Hono, S. Takaki, K. Hashimoto, K. Oura, Y. Nankaku, and K. Tokuda, “PeriodNet: A non-autoregressive raw waveform generative model with a structure separating periodic and aperiodic components,” IEEE Access, vol. 9, pp. 137 599–137 612, 2021.
  • [31] H. Zen, K. Tokuda, T. Masuko, T. Kobayasih, and T. Kitamura, “A hidden semi-Markov model-based speech synthesis system,” IEICE Transactions on Information and Systems, vol. E90-D, no. 5, pp. 825–834, 2007.
  • [32] Y. Nankaku, K. Sumiya, T. Yoshimura, S. Takaki, K. Hashimoto, K. Oura, and K. Tokuda, “Neural sequence-to-sequence speech synthesis using a hidden semi-markov model based structured attention mechanism,” arXiv preprint arXiv:2108.13985, 2021.