Skip to Main content Skip to Navigation
Conference papers

Visual attention modeling for stereoscopic video

Abstract : In this paper, we propose a computational model of visual attention for stereoscopic video. Low-level visual features including color, luminance, texture and depth are used to calculate feature contrast for spatial saliency of stereoscopic video frames. Besides, the proposed model adopts motion features to compute the temporal saliency. Here, we extract the relative planar and depth motion for temporal saliency calculation. The final saliency map is computed by fusing the spatial and temporal saliency together. Experimental results show the promising performance of the proposed method in saliency prediction for stereoscopic video.
Document type :
Conference papers
Complete list of metadatas

Cited literature [25 references]  Display  Hide  Download
Contributor : Matthieu Perreira da Silva <>
Submitted on : Tuesday, January 17, 2017 - 3:53:28 PM
Last modification on : Tuesday, January 14, 2020 - 3:59:53 PM
Long-term archiving on: : Tuesday, April 18, 2017 - 3:00:55 PM


Files produced by the author(s)




Yuming Fang, Chi Zhang, Jing Li, Matthieu Perreira da Silva, Patrick Le Callet. Visual attention modeling for stereoscopic video. 2016 IEEE International Conference on Multimedia & Expo Workshops (ICMEW), Jul 2016, Seattle, United States. pp.1 - 6, ⟨10.1109/ICMEW.2016.7574768⟩. ⟨hal-01438315⟩



Record views


Files downloads