Explicit Intensity Control for Accented Text-to-Speech


Abstract
Accented text-to-speech (TTS) synthesis seeks to generate speech with an accent (L2) as a variant of the standard version (L1). How to control the intensity of accent in the process of TTS is a very interesting research direction, and has attracted more and more attention. Recent work design a speaker-adversarial loss to disentangle the speaker and accent information, and then adjust the loss weight to control the accent intensity. However, such a control method lacks interpretability, and there is no direct correlation between the controlling factor and natural accent intensity. To this end, this paper propose a new intuitive and explicit accent intensity control scheme for accented TTS. Specifically, we first extract the posterior probability, called as "Goodness of Pronunciation (GoP)" from the L1 speech recognition model to quantify the phoneme accent intensity for accented speech, then design a FastSpeech2 based TTS model, named Ai-TTS, to take the accent intensity expression into account during speech generation. Experiments show that the our method outperforms the baseline model in terms of accent rendering and intensity control.



Main Results
Controllability Evaluation on Utterance-level

Unconsciously, our yells and exclamations yielded to this rhythm.
(Speaker: TXHC; Accent: Mandarin)

Ground Truth DAW (Intensity = "strong") Ai-TTS (Intensity = "strong")

He made no reply as he waited for Whittemore to continue.
(Speaker: NCC; Accent: Mandarin)

Ground Truth DAW (Intensity = "strong") Ai-TTS (Intensity = "strong")
Controllability Evaluation on Phoneme-level

Unconsciously, our yells and exclamations yielded to this rhythm.
(Speaker: TXHC; Accent: Mandarin)

Phoneme Sequence: AH2 N K AA1 N SH AH0 S L IY0 sp AW1 ER0 Y EH1 L Z AE1 N D sp EH2 K S K L AH0 M EY1 SH AH0 N Z sp Y IY1 L D IH0 D T UW1 DH IH1 S R IH1 DH AH0 M.
Sample (1):
AH2 (0.9) N (0.9) K (0.9) AA1 (0.9) N (0.9) SH (0.9) AH0 (0.9) S (0.9) L (0.9) IY0 (0.9) sp AW1 (0.1) ER0 (0.1) Y (0.9) EH1 (0.9) L (0.9) Z (0.9) AE1 (0.1) N (0.1) (0.1) D (0.1) sp EH2 (0.9) K (0.9) S (0.9) K (0.9) L (0.9) AH0 (0.9) M (0.9) EY1 (0.9) SH (0.9) AH0 (0.9) N (0.9) Z (0.9) sp Y (0.1) IY1 (0.1) L (0.1) D (0.1) IH0 (0.1) D (0.1) T (0.1) UW1 (0.1) DH (0.1) IH1 (0.1) S (0.1) R (0.1) IH1 (0.1) DH (0.1) AH0 (0.1) M (0.1).
Sample (2):
AH2 (0.9) N (0.9) K (0.9) AA1 (0.9) N (0.9) SH (0.9) AH0 (0.9) S (0.9) L (0.9) IY0 (0.9) sp AW1 (0.1) ER0 (0.1) Y (0.1) EH1 (0.1) L (0.1) Z (0.1) AE1 (0.1) N (0.1) (0.1) D (0.1) sp EH2 (0.1) K (0.1) S (0.1) K (0.1) L (0.1) AH0 (0.1) M (0.1) EY1 (0.1) SH (0.1) AH0 (0.1) N (0.1) Z (0.1) sp Y (0.1) IY1 (0.1) L (0.1) D (0.1) IH0 (0.1) D (0.1) T (0.1) UW1 (0.1) DH (0.1) IH1 (0.1) S (0.1) R (0.1) IH1 (0.1) DH (0.1) AH0 (0.1) M (0.1).
Sample (3):
AH2 (0.1) N (0.1) K (0.1) AA1 (0.1) N (0.1)SH (0.1) AH0 (0.1) S (0.1) L (0.1) IY0 (0.1) sp AW1 (0.1) ER0 (0.1) Y (0.1) EH1 (0.1) L (0.1) Z (0.1) AE1 (0.1) N (0.1) (0.1) D (0.1) sp EH2 (0.1) K (0.1) S (0.1) K (0.1) L (0.1) AH0 (0.1) M (0.1) EY1 (0.1) SH (0.1) AH0 (0.1) N (0.1) Z (0.1) sp Y (0.1) IY1 (0.1) L (0.1) D (0.1) IH0 (0.1) D (0.1) T (0.1) UW1 (0.1) DH (0.1) IH1 (0.1) S (0.1) R (0.1) IH1 (0.1) DH (0.1) AH0 (0.1) M (0.1).
References:
[1] Yi Ren, Chenxu Hu, Xu Tan, Tao Qin, Sheng Zhao, Zhou Zhao, and Tie-Yan Liu, “Fastspeech 2: Fast and high-quality end-to-end text to speech,” in International Conference on Learning Representations, 2020.

[2] Meinard M ̈uller, “Dynamic time warping,” Information retrieval for music and motion, pp. 69–84, 2007.