An audiovisual talking head for augmented speech generation: models and animations based on a real speaker's articulatory data - Archive ouverte HAL Accéder directement au contenu
Chapitre D'ouvrage Année : 2008

An audiovisual talking head for augmented speech generation: models and animations based on a real speaker's articulatory data

Résumé

We present a methodology developed to derive three-dimensional models of speech articulators from volume MRI and multiple view video images acquired on one speaker. Linear component analysis is used to model these highly deformable articulators as the weighted sum of a small number of basic shapes corresponding to the articulators' degrees of freedom for speech. These models are assembled into an audiovisual talking head that can produce augmented audiovisual speech, i.e. can display usually non visible articulators such as tongue or velum. The talking head is then animated by recovering its control parameters by inversion from the coordinates of a small number of points of the articulators of the same speaker tracked by Electro-Magnetic Articulography. The augmented speech produced points the way to promising applications in the domain of speech therapy for speech retarded children, perception and production rehabilitation of hearing impaired children, and pronunciation training for second language learners.

Dates et versions

hal-00296599 , version 1 (13-07-2008)

Identifiants

Citer

Pierre Badin, Frédéric Elisei, Gérard Bailly, Yuliya Tarabalka. An audiovisual talking head for augmented speech generation: models and animations based on a real speaker's articulatory data. F.J. Perales & R.B. Fisher. Proceedings of the Vth Conference on Articulated Motion and Deformable Objects (AMDO 2008), 5098, Springer Verlag: Berlin, Heidelberg, Germany, pp.132-143, 2008, Lecture Notes in Computer Science, 5098, ⟨10.1007/978-3-540-70517-8_14⟩. ⟨hal-00296599⟩
229 Consultations
0 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More