MULTIMODAL INFORMATION FUSION AND TEMPORAL INTEGRATION FOR VIOLENCE DETECTION IN MOVIES - Archive ouverte HAL Accéder directement au contenu
Communication Dans Un Congrès Année : 2012

MULTIMODAL INFORMATION FUSION AND TEMPORAL INTEGRATION FOR VIOLENCE DETECTION IN MOVIES

Résumé

This paper presents a violent shots detection system that studies several methods for introducing temporal and multimodal information in the framework. It also investigates different kinds of Bayesian network structure learning algorithms for modelling these problems. The system is trained and tested using the MediaEval 2011 Affect Task corpus, which comprises of 15 Hollywood movies. It is experimentally shown that both multimodality and temporality add interesting information into the system. Moreover, the analysis of the links between the variables of the resulting graphs yields important observations about the quality of the structure learning algorithms. Overall, our best system achieved 50% false alarms and 3% missed detection, which is among the best submissions in the MediaEval campaign.
Fichier principal
Vignette du fichier
CedricPenet_MultimodalInformationFusionAndTemporalIntegrationForViolenceDetectionInMoviesVFINAL.pdf (280.01 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)
Loading...

Dates et versions

hal-00671016 , version 1 (16-02-2012)

Identifiants

  • HAL Id : hal-00671016 , version 1

Citer

Cédric Penet, Claire-Hélène Demarty, Guillaume Gravier, Patrick Gros. MULTIMODAL INFORMATION FUSION AND TEMPORAL INTEGRATION FOR VIOLENCE DETECTION IN MOVIES. ICASSP - 37th International Conference on Acoustics, Speech, and Signal Processing (2012), Mar 2012, Kyoto, Japan. ⟨hal-00671016⟩
260 Consultations
506 Téléchargements

Partager

Gmail Facebook X LinkedIn More