Large-Scale Study of Perceptual Video Quality

The great variations of videographic skills in videography, camera designs, compression and processing protocols, communication and bandwidth environments, and displays leads to an enormous variety of video impairments. Current no-reference (NR) video quality models are unable to handle this diversity of distortions. This is true in part because available video quality assessment databases contain very limited content, fixed resolutions, were captured using a small number of camera devices by a few videographers and have been subjected to a modest number of distortions. As such, these databases fail to adequately represent real world videos, which contain very different kinds of content obtained under highly diverse imaging conditions and are subject to authentic, complex, and often commingled distortions that are difficult or impossible to simulate. As a result, NR video quality predictors tested on real-world video data often perform poorly. Toward advancing NR video quality prediction, we have constructed a large-scale video quality assessment database containing 585 videos of unique content, captured by a large number of users, with wide ranges of levels of complex, authentic distortions. We collected a large number of subjective video quality scores via crowdsourcing. A total of 4776 unique participants took part in the study, yielding over 205 000 opinion scores, resulting in an average of 240 recorded human opinions per video. We demonstrate the value of the new resource, which we call the live video quality challenge database (LIVE-VQC), by conducting a comparison with leading NR video quality predictors on it. This paper is the largest video quality assessment study ever conducted along several key dimensions: number of unique contents, capture devices, distortion types and combinations of distortions, study participants, and recorded subjective scores. The database is available for download on this link: http://live.ece.utexas.edu/research/LIVEVQC/index.html.

[1]  Hari Kalva,et al.  Assessing internet video quality using crowdsourcing , 2013, CrowdMM '13.

[2]  Deva Ramanan,et al.  Efficiently Scaling up Crowdsourced Video Annotation , 2012, International Journal of Computer Vision.

[3]  Christian Timmerer,et al.  Quality of Experience of Web-based Adaptive HTTP Streaming Clients in Real-World Environments using Crowdsourcing , 2014, VideoNext '14.

[4]  Vijayan K. Asari,et al.  No-Reference Video Quality Assessment Based on Artifact Measurement and Statistical Analysis , 2015, IEEE Transactions on Circuits and Systems for Video Technology.

[5]  Stefano Tubaro,et al.  A H.264/AVC video database for the evaluation of quality metrics , 2010, 2010 IEEE International Conference on Acoustics, Speech and Signal Processing.

[6]  Yanjiao Chen,et al.  From QoS to QoE: A Tutorial on Video Quality Assessment , 2015, IEEE Communications Surveys & Tutorials.

[7]  Dietmar Saupe,et al.  The Konstanz natural video database (KoNViD-1k) , 2017, 2017 Ninth International Conference on Quality of Multimedia Experience (QoMEX).

[8]  Alan C. Bovik,et al.  Massive Online Crowdsourced Study of Subjective and Objective Picture Quality , 2015, IEEE Transactions on Image Processing.

[9]  Michael S. Bernstein,et al.  ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.

[10]  C.-C. Jay Kuo,et al.  MCL-V: A streaming video quality assessment database , 2015, J. Vis. Commun. Image Represent..

[11]  Pavel Korshunov,et al.  Crowdsourcing-based multimedia subjective evaluations: a case study on image recognizability and aesthetic appeal , 2013, CrowdMM '13.

[12]  M. Stone Cross‐Validatory Choice and Assessment of Statistical Predictions , 1976 .

[13]  Christian Keimel,et al.  QualityCrowd — A framework for crowd-based quality evaluation , 2012, 2012 Picture Coding Symposium.

[14]  Mikko Nuutinen,et al.  CID2013: A Database for Evaluating No-Reference Image Quality Assessment Algorithms , 2015, IEEE Transactions on Image Processing.

[15]  Dietmar Saupe,et al.  Empirical evaluation of no-reference VQA methods on a natural video quality database , 2017, 2017 Ninth International Conference on Quality of Multimedia Experience (QoMEX).

[16]  Michael Seufert,et al.  One Shot Crowdtesting: Approaching the Extremes of Crowdsourced Subjective Quality Testing , 2016 .

[17]  Christian Keimel,et al.  The TUM high definition video datasets , 2012, 2012 Fourth International Workshop on Quality of Multimedia Experience.

[18]  Cisco Visual Networking Index: Forecast and Methodology 2016-2021.(2017) http://www.cisco.com/c/en/us/solutions/collateral/service-provider/visual- networking-index-vni/complete-white-paper-c11-481360.html. High Efficiency Video Coding (HEVC) Algorithms and Architectures https://jvet.hhi.fraunhofer. , 2017 .

[19]  Chin-Laung Lei,et al.  Quadrant of euphoria: a crowdsourcing platform for QoE assessment , 2010, IEEE Network.

[20]  Alan C. Bovik,et al.  Making a “Completely Blind” Image Quality Analyzer , 2013, IEEE Signal Processing Letters.

[21]  Alan C. Bovik,et al.  In-Capture Mobile Video Distortions: A Study of Subjective Behavior and Objective Algorithms , 2018, IEEE Transactions on Circuits and Systems for Video Technology.

[22]  Qingming Huang,et al.  CNN-MR for No Reference Video Quality Assessment , 2017, 2017 4th International Conference on Information Science and Control Engineering (ICISCE).

[23]  Jacob Søgaard,et al.  Crowdsourcing based subjective quality assessment of adaptive video streaming , 2014, 2014 Sixth International Workshop on Quality of Multimedia Experience (QoMEX).

[24]  Sugato Chakravarty,et al.  Methodology for the subjective assessment of the quality of television pictures , 1995 .

[25]  Chih-Jen Lin,et al.  LIBSVM: A library for support vector machines , 2011, TIST.

[26]  Christophe Charrier,et al.  Blind Prediction of Natural Video Quality , 2014, IEEE Transactions on Image Processing.

[27]  Phuoc Tran-Gia,et al.  Best Practices for QoE Crowdtesting: QoE Assessment With Crowdsourcing , 2014, IEEE Transactions on Multimedia.

[28]  Bernhard Schölkopf,et al.  New Support Vector Algorithms , 2000, Neural Computation.

[29]  Israel Abramov,et al.  Sex and vision II: color appearance of monochromatic lights , 2012, Biology of Sex Differences.

[30]  Alan C. Bovik,et al.  No-Reference Image Quality Assessment in the Spatial Domain , 2012, IEEE Transactions on Image Processing.

[31]  Christian Keimel,et al.  Visual quality of current coding technologies at high definition IPTV bitrates , 2010, 2010 IEEE International Workshop on Multimedia Signal Processing.

[32]  Rajiv Soundararajan,et al.  Study of Subjective and Objective Quality Assessment of Video , 2010, IEEE Transactions on Image Processing.

[33]  Gustavo de Veciana,et al.  Video Quality Assessment on Mobile Devices: Subjective, Behavioral and Objective Studies , 2012, IEEE Journal of Selected Topics in Signal Processing.

[34]  Mikko Nuutinen,et al.  CVD2014—A Database for Evaluating No-Reference Video Quality Assessment Algorithms , 2016, IEEE Transactions on Image Processing.

[35]  Xin Jin,et al.  VideoSet: A large-scale compressed video quality dataset based on JND measurement , 2017, J. Vis. Commun. Image Represent..

[36]  Xuelong Li,et al.  Spatiotemporal Statistics for Video Quality Assessment , 2016, IEEE Transactions on Image Processing.

[37]  Gustavo de Veciana,et al.  Modeling the Time—Varying Subjective Quality of HTTP Video Streams With Rate Adaptations , 2013, IEEE Transactions on Image Processing.

[38]  B. Rovner,et al.  Effect of depression on vision function in age-related macular degeneration. , 2002, Archives of ophthalmology.

[39]  Alan C. Bovik,et al.  A Completely Blind Video Integrity Oracle , 2016, IEEE Transactions on Image Processing.

[40]  Alan C. Bovik,et al.  Large Scale Subjective Video Quality Study , 2018, 2018 25th IEEE International Conference on Image Processing (ICIP).