BERT, can HE predict contrastive focus? Predicting and controlling prominence in neural TTS using a language model - Archive ouverte HAL Accéder directement au contenu
Communication Dans Un Congrès Année : 2022

BERT, can HE predict contrastive focus? Predicting and controlling prominence in neural TTS using a language model

Résumé

Several recent studies have tested the use of transformer language model representations to infer prosodic features for textto-speech synthesis (TTS). While these studies have explored prosody in general, in this work, we look specifically at the prediction of contrastive focus on personal pronouns. This is a particularly challenging task as it often requires semantic, discursive and/or pragmatic knowledge to predict correctly. We collect a corpus of utterances containing contrastive focus and we evaluate the accuracy of a BERT model, finetuned to predict quantized acoustic prominence features, on these samples. We also investigate how past utterances can provide relevant information for this prediction. Furthermore, we evaluate the controllability of pronoun prominence in a TTS model conditioned on acoustic prominence features.
Fichier principal
Vignette du fichier
BERT__can_HE_predict_contrastive_focus_.pdf (1.48 Mo) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)

Dates et versions

hal-03791472 , version 1 (29-09-2022)

Identifiants

Citer

Brooke Stephenson, Laurent Besacier, Laurent Girin, Thomas Hueber. BERT, can HE predict contrastive focus? Predicting and controlling prominence in neural TTS using a language model. Interspeech 2022 - 23rd Annual Conference of the International Speech Communication Association, Sep 2022, Incheon, South Korea. pp.3383-3387, ⟨10.21437/Interspeech.2022-10116⟩. ⟨hal-03791472⟩
109 Consultations
35 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More