Statistical Audio-Visual Data Fusion for Video Scene Segmentation

Abstract : Automatic video segmentation into semantic units is important to organize an effective content based access to long video. In this work we focus on the problem of video segmentation into narrative units called scenes - aggregates of shots unified by a common dramatic event or locale. In this work we derive a statistical video scene segmentation approach which detects scenes boundaries in one pass fusing multimodal audio-visual features in a symmetrical and scalable manner. The approach deals properly with the variability of real-valued features and models their conditional dependence on the context. It also integrates prior information concerning the duration of scenes. Two kinds of features extracted in visual and audio domain are proposed. The results of experimental evaluations carried out on ground truth video are reported. They show that our approach effectively fuse multiple modalities with higher performance as compared with an alternative rule-based fusion technique.
Document type :
Book sections
Complete list of metadatas

https://hal.archives-ouvertes.fr/hal-01589551
Contributor : Équipe Gestionnaire Des Publications Si Liris <>
Submitted on : Monday, September 18, 2017 - 4:37:48 PM
Last modification on : Thursday, November 21, 2019 - 2:20:01 AM

Identifiers

Citation

Vyacheslav Parshin, Liming Chen. Statistical Audio-Visual Data Fusion for Video Scene Segmentation. Pr. Yujin Zhang. Semantic-Based Visual Information Retrieval, Idea Group Inc., pp.68-89, 2007, ⟨10.4018/978-1-59904-370-8.ch004⟩. ⟨hal-01589551⟩

Share

Metrics

Record views

386