Visual attention modeling for 3D video using neural networks - Archive ouverte HAL Accéder directement au contenu
Communication Dans Un Congrès Année : 2014

Visual attention modeling for 3D video using neural networks

Iana Iatsun
  • Fonction : Auteur
SIC
Mohamed-Chaker Larabi
SIC
Christine Fernandez-Maloigne
SIC

Résumé

Visual attention is one of the most important mechanisms in the human visual perception. Recently, its modeling becomes a principal requirement for the optimization of the image processing systems. Numerous algorithms have already been designed for 2D saliency prediction. However, only few works can be found for 3D content. In this study, we propose a saliency model for stereoscopic 3D video. This algorithm extracts information from three dimensions of content, i.e. spatial, temporal and depth. This model benefits from the properties of interest points to be close to human fixations in order to build spatial salient features. Besides, as the perception of depth relies strongly on monocular cues, our model extracts the depth salient features using the pictorial depth sources. Since weights for fusion strategy are often selected in ad-hoc manner, in this work, we suggest to use a machine learning approach. The used artificial Neural Network allows to define adaptive weights based on the eye-tracking data. The results of the proposed algorithm are tested versus ground-truth information using the state-of-the-art techniques.
Fichier non déposé

Dates et versions

hal-01159126 , version 1 (02-06-2015)

Identifiants

Citer

Iana Iatsun, Mohamed-Chaker Larabi, Christine Fernandez-Maloigne. Visual attention modeling for 3D video using neural networks. International Conference on 3D Imaging (IC3D), Dec 2014, Liège, Belgium. pp.1 - 8, ⟨10.1109/IC3D.2014.7032602⟩. ⟨hal-01159126⟩
57 Consultations
0 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More