scispace - formally typeset
Open AccessJournal ArticleDOI

UGC-VQA: Benchmarking Blind Video Quality Assessment for User Generated Content

Reads0
Chats0
TLDR
In this article, the VIDeo quality EVALuator (VIDEVAL) is proposed to improve the performance of VQA models for UGC/consumer videos.
Abstract
Recent years have witnessed an explosion of user-generated content (UGC) videos shared and streamed over the Internet, thanks to the evolution of affordable and reliable consumer capture devices, and the tremendous popularity of social media platforms. Accordingly, there is a great need for accurate video quality assessment (VQA) models for UGC/consumer videos to monitor, control, and optimize this vast content. Blind quality prediction of in-the-wild videos is quite challenging, since the quality degradations of UGC videos are unpredictable, complicated, and often commingled. Here we contribute to advancing the UGC-VQA problem by conducting a comprehensive evaluation of leading no-reference/blind VQA (BVQA) features and models on a fixed evaluation architecture, yielding new empirical insights on both subjective video quality studies and objective VQA model design. By employing a feature selection strategy on top of efficient BVQA models, we are able to extract 60 out of 763 statistical features used in existing methods to create a new fusion-based model, which we dub the VIDeo quality EVALuator (VIDEVAL), that effectively balances the trade-off between VQA performance and efficiency. Our experimental results show that VIDEVAL achieves state-of-the-art performance at considerably lower computational cost than other leading models. Our study protocol also defines a reliable benchmark for the UGC-VQA problem, which we believe will facilitate further research on deep learning-based VQA modeling, as well as perceptually-optimized efficient UGC video processing, transcoding, and streaming. To promote reproducible research and public evaluation, an implementation of VIDEVAL has been made available online: https://github.com/vztu/VIDEVAL .

read more

Citations
More filters
Journal ArticleDOI

RAPIQUE: Rapid and Accurate Video Quality Prediction of User Generated Content

TL;DR: In this paper, the Rapid and Accurate Video Quality Evaluator (RAPIQUE) model is proposed for video quality prediction, which combines and leverages the advantages of both quality-aware scene statistics features and semantics-aware deep convolutional features.
Proceedings ArticleDOI

Exploring the Effectiveness of Video Perceptual Representation in Blind Video Quality Assessment

TL;DR: Wang et al. as mentioned in this paper proposed a temporal perceptual quality index (TPQI) to measure the temporal distortion by describing the graphic morphology of the representation, which can be applied to any dataset without parameter tuning.
Journal ArticleDOI

Spatiotemporal Representation Learning for Blind Video Quality Assessment

TL;DR: Sissuire et al. as discussed by the authors proposed HEKE, a feature encoder specific to blind video quality assessment (BVQA), to extract spatio-temporal representation from videos.
Journal ArticleDOI

Contrastive Self-Supervised Pre-Training for Video Quality Assessment

TL;DR: In this article , a self-supervised pre-training for video quality assessment (VQA) task is proposed, which exploits the plentiful unlabeled video data to learn feature representation in a simple-yet-effective way.
Journal ArticleDOI

A Subjective and Objective Study of Space-Time Subsampled Video Quality

TL;DR: The ETRI-LIVE Space-Time Sub-sampled Video Quality (ETRI-Live STSVQ) dataset as mentioned in this paper contains 437 videos generated by applying various levels of combined space-time subsampling and video compression on 15 diverse video contents.
References
More filters
Proceedings ArticleDOI

Deep Residual Learning for Image Recognition

TL;DR: In this article, the authors proposed a residual learning framework to ease the training of networks that are substantially deeper than those used previously, which won the 1st place on the ILSVRC 2015 classification task.
Proceedings Article

Very Deep Convolutional Networks for Large-Scale Image Recognition

TL;DR: This work investigates the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting using an architecture with very small convolution filters, which shows that a significant improvement on the prior-art configurations can be achieved by pushing the depth to 16-19 weight layers.
Proceedings ArticleDOI

ImageNet: A large-scale hierarchical image database

TL;DR: A new database called “ImageNet” is introduced, a large-scale ontology of images built upon the backbone of the WordNet structure, much larger in scale and diversity and much more accurate than the current image datasets.
Journal ArticleDOI

LIBSVM: A library for support vector machines

TL;DR: Issues such as solving SVM optimization problems theoretical convergence multiclass classification probability estimates and parameter selection are discussed in detail.
Journal ArticleDOI

Image quality assessment: from error visibility to structural similarity

TL;DR: In this article, a structural similarity index is proposed for image quality assessment based on the degradation of structural information, which can be applied to both subjective ratings and objective methods on a database of images compressed with JPEG and JPEG2000.
Related Papers (5)