Blind Prediction of Natural Video Quality - Archive ouverte HAL Accéder directement au contenu
Article Dans Une Revue IEEE Transactions on Image Processing Année : 2014

Blind Prediction of Natural Video Quality

Michele Saad
  • Fonction : Auteur
Alan C. Bovik
  • Fonction : Auteur

Résumé

We propose a blind (no reference or NR) video quality evaluation model that is nondistortion specific. The approach relies on a spatio-temporal model of video scenes in the discrete cosine transform domain, and on a model that characterizes the type of motion occurring in the scenes, to predict video quality. We use the models to define video statistics and perceptual features that are the basis of a video quality assessment (VQA) algorithm that does not require the presence of a pristine video to compare against in order to predict a perceptual quality score. The contributions of this paper are threefold. 1) We propose a spatio-temporal natural scene statistics (NSS) model for videos. 2) We propose a motion model that quantifies motion coherency in video scenes. 3) We show that the proposed NSS and motion coherency models are appropriate for quality assessment of videos, and we utilize them to design a blind VQA algorithm that correlates highly with human judgments of quality. The proposed algorithm, called video BLIINDS, is tested on the LIVE VQA database and on the EPFL-PoliMi video database and shown to perform close to the level of top performing reduced and full reference VQA algorithms.

Dates et versions

hal-01211821 , version 1 (05-10-2015)

Identifiants

Citer

Michele Saad, Alan C. Bovik, Christophe Charrier. Blind Prediction of Natural Video Quality. IEEE Transactions on Image Processing, 2014, 23 (3), pp.1352 - 1365. ⟨10.1109/TIP.2014.2299154⟩. ⟨hal-01211821⟩
86 Consultations
0 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More