Temporal Multimodal Fusion for Video Emotion Classification in the Wild - Archive ouverte HAL Accéder directement au contenu
Communication Dans Un Congrès Année : 2017

Temporal Multimodal Fusion for Video Emotion Classification in the Wild

Résumé

This paper addresses the question of emotion classification. The task consists in predicting emotion labels (taken among a set of possible labels) best describing the emotions contained in short video clips. Building on a standard framework – lying in describing videos by audio and visual features used by a supervised classifier to infer the labels – this paper investigates several novel directions. First of all, improved face descriptors based on 2D and 3D Convo-lutional Neural Networks are proposed. Second, the paper explores several fusion methods, temporal and multimodal, including a novel hierarchical method combining features and scores. In addition, we carefully reviewed the different stages of the pipeline and designed a CNN architecture adapted to the task; this is important as the size of the training set is small compared to the difficulty of the problem, making generalization difficult. The so-obtained model ranked 4th at the 2017 Emotion in the Wild challenge with the accuracy of 58.8 %.
Fichier principal
Vignette du fichier
main.pdf (1.76 Mo) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)

Dates et versions

hal-01590608 , version 1 (20-09-2017)

Identifiants

Citer

Valentin Vielzeuf, Stéphane Pateux, Frédéric Jurie. Temporal Multimodal Fusion for Video Emotion Classification in the Wild. ACM - ICMI 2017, Nov 2017, Glasgow, United Kingdom. ⟨hal-01590608⟩
247 Consultations
493 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More