4.6 Article

No-Reference Video Quality Assessment Using Multi-Pooled, Saliency Weighted Deep Features and Decision Fusion

Journal

SENSORS
Volume 22, Issue 6, Pages -

Publisher

MDPI
DOI: 10.3390/s22062209

Keywords

no-reference video quality assessment; convolutional neural network; decision fusion

Ask authors/readers for more resources

This paper presents an innovative deep learning-based approach for no-reference video quality assessment (NR-VQA) that utilizes parallel pre-trained convolutional neural networks to extract deep features and map them to perceptual quality scores. The quality scores from different regressors are then fused to obtain the overall perceptual quality of a video sequence. Experimental results show that the proposed method achieves state-of-the-art performance on benchmark databases with authentic distortions, and the fusion of multiple deep architectures significantly improves NR-VQA.
With the constantly growing popularity of video-based services and applications, no-reference video quality assessment (NR-VQA) has become a very hot research topic. Over the years, many different approaches have been introduced in the literature to evaluate the perceptual quality of digital videos. Due to the advent of large benchmark video quality assessment databases, deep learning has attracted a significant amount of attention in this field in recent years. This paper presents a novel, innovative deep learning-based approach for NR-VQA that relies on a set of in parallel pre-trained convolutional neural networks (CNN) to characterize versatitely the potential image and video distortions. Specifically, temporally pooled and saliency weighted video-level deep features are extracted with the help of a set of pre-trained CNNs and mapped onto perceptual quality scores independently from each other. Finally, the quality scores coming from the different regressors are fused together to obtain the perceptual quality of a given video sequence. Extensive experiments demonstrate that the proposed method sets a new state-of-the-art on two large benchmark video quality assessment databases with authentic distortions. Moreover, the presented results underline that the decision fusion of multiple deep architectures can significantly benefit NR-VQA.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available