High-level geometry-based features of video modality for emotion prediction. - Archive ouverte HAL Accéder directement au contenu
Communication Dans Un Congrès Année : 2016

High-level geometry-based features of video modality for emotion prediction.

Résumé

The automatic analysis of emotion remains a challenging task in unconstrained experimental conditions. In this paper, we present our contribution to the 6th Audio/Visual Emotion Challenge (AVEC 2016), which aims at predicting the continuous emotional dimensions of arousal and valence. First, we propose to improve the performance of the multimodal prediction with low-level features by adding high-level geometry-based features, namely head pose and expression signature. The head pose is estimated by fitting a reference 3D mesh to the 2D facial landmarks. The expression signature is the projection of the facial landmarks in an unsupervised person-specific model. Second, we propose to fuse the unimodal predictions trained on each training subject before performing the multimodal fusion. The results show that our high-level features improve the performance of the multimodal prediction of arousal and that the subjects fusion works well in unimodal prediction but generalizes poorly in multimodal prediction, particularly on valence.
Fichier non déposé

Dates et versions

hal-01831390 , version 1 (05-07-2018)

Identifiants

Citer

Raphael Weber, Vincent Barrielle, Catherine Soladie, Renaud Seguier. High-level geometry-based features of video modality for emotion prediction.. 6th International Workshop on Audio/Visual Emotion Challenge (AVEC'16), Oct 2016, Amsterdam, Netherlands. p.51-58, ⟨10.1145/2988257.2988262⟩. ⟨hal-01831390⟩
222 Consultations
0 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More