Predicting unseen articulations from multi-speaker articulatory models - Archive ouverte HAL Accéder directement au contenu
Communication Dans Un Congrès Année : 2010

Predicting unseen articulations from multi-speaker articulatory models

Résumé

In order to study inter-speaker variability, this work aims to assess the generalization capabilities of data-based multi-speaker articulatory models. We use various three-mode factor analysis techniques to model the variations of midsagittal vocal tract contours obtained from MRI images for three French speakers articulating 73 vowels and consonants. Articulations of a given speaker for phonemes not present in the training set are then predicted by inversion of the models from measurements of these phonemes articulated by the other subjects. On the average, the prediction RMSE was 5.25 mm for tongue contours, and 3.3 mm for 2D midsagittal vocal tract distances. Besides, this study has established a methodology to determine the optimal number of factors for such models.
Fichier non déposé

Dates et versions

hal-00508267 , version 1 (02-08-2010)

Identifiants

  • HAL Id : hal-00508267 , version 1

Citer

Gopal Ananthakrishnan, Pierre Badin, Julián Andrés Valdés Vargas, Olov Engwall. Predicting unseen articulations from multi-speaker articulatory models. Interspeech 2010 - 11th Annual Conference of the International Speech Communication Association, Sep 2010, Makuhari, Japan. pp.n.c. ⟨hal-00508267⟩
161 Consultations
0 Téléchargements

Partager

Gmail Facebook X LinkedIn More