Massive Online Crowdsourced Study of Subjective and Objective Picture Quality

Most publicly available image quality databases have been created under highly controlled conditions by introducing graded simulated distortions onto high-quality photographs. However, images captured using typical real-world mobile camera devices are usually afflicted by complex mixtures of multiple distortions, which are not necessarily well-modeled by the synthetic distortions found in existing databases. The originators of existing legacy databases usually conducted human psychometric studies to obtain statistically meaningful sets of human opinion scores on images in a stringently controlled visual environment, resulting in small data collections relative to other kinds of image analysis databases. Toward overcoming these limitations, we designed and created a new database that we call the LIVE In the Wild Image Quality Challenge Database, which contains widely diverse authentic image distortions on a large number of images captured using a representative variety of modern mobile devices. We also designed and implemented a new online crowdsourcing system, which we have used to conduct a very large-scale, multi-month image quality assessment (IQA) subjective study. Our database consists of over 350 000 opinion scores on 1162 images evaluated by over 8100 unique human observers. Despite the lack of control over the experimental environments of the numerous study participants, we demonstrate excellent internal consistency of the subjective data set. We also evaluate several top-performing blind IQA algorithms on it and present insights on how the mixtures of distortions challenge both end users as well as automatic perceptual quality prediction models. The new database is available for public use at <;uri xlink:href="http://live.ece.utexas.edu/research/ChallengeDB/index.html" xlink:type="simple">http://live.ece.utexas.edu/research/ChallengeDB/index.html<;/uri>.

[1]  Christophe Charrier,et al.  Blind Image Quality Assessment: A Natural Scene Statistics Approach in the DCT Domain , 2012, IEEE Transactions on Image Processing.

[2]  Chih-Jen Lin,et al.  Ranking individuals by group comparisons , 2006, ICML.

[3]  Alan C. Bovik,et al.  Blind image quality assessment on real distorted images using deep belief nets , 2014, 2014 IEEE Global Conference on Signal and Information Processing (GlobalSIP).

[4]  Christian Timmerer,et al.  A web based subjective evaluation platform , 2013, 2013 Fifth International Workshop on Quality of Multimedia Experience (QoMEX).

[5]  Alan C. Bovik,et al.  Automatic Prediction of Perceptual Image and Video Quality , 2013, Proceedings of the IEEE.

[6]  D. Heeger Normalization of cell responses in cat striate cortex , 1992, Visual Neuroscience.

[7]  Alan C. Bovik,et al.  A Statistical Evaluation of Recent Full Reference Image Quality Assessment Algorithms , 2006, IEEE Transactions on Image Processing.

[8]  Christian Keimel,et al.  Crowdsourcing in QoE Evaluation , 2014, Quality of Experience.

[9]  Qingming Huang,et al.  Online crowdsourcing subjective image quality assessment , 2012, ACM Multimedia.

[10]  Yee Whye Teh,et al.  A Fast Learning Algorithm for Deep Belief Nets , 2006, Neural Computation.

[11]  D. Rene Rasmussen The mobile image quality survey game , 2011, Electronic Imaging.

[12]  Antonio Torralba,et al.  LabelMe: A Database and Web-Based Tool for Image Annotation , 2008, International Journal of Computer Vision.

[13]  Eero P. Simoncelli,et al.  Image quality assessment: from error visibility to structural similarity , 2004, IEEE Transactions on Image Processing.

[14]  Alan C. Bovik,et al.  Wireless Video Quality Assessment: A Study of Subjective Scores and Objective Algorithms , 2010, IEEE Transactions on Circuits and Systems for Video Technology.

[15]  Michael Seufert,et al.  Crowdsourcing 2.0: Enhancing execution speed and reliability of web-based QoE testing , 2014, 2014 IEEE International Conference on Communications (ICC).

[16]  Alan C. Bovik,et al.  Feature maps driven no-reference image quality prediction of authentically distorted images , 2015, Electronic Imaging.

[17]  Zhou Wang,et al.  Multiscale structural similarity for image quality assessment , 2003, The Thrity-Seventh Asilomar Conference on Signals, Systems & Computers, 2003.

[18]  Ryszard Stasinski,et al.  Internet based subjective assessment of image quality experiment , 2011, 2011 18th International Conference on Systems, Signals and Image Processing.

[19]  Guoping Qiu,et al.  Social image quality , 2011, Electronic Imaging.

[20]  Bülent Sankur,et al.  Statistical evaluation of image quality measures , 2002, J. Electronic Imaging.

[21]  Laura A. Dabbish,et al.  Labeling images with a computer game , 2004, AAAI Spring Symposium: Knowledge Collection from Volunteer Contributors.

[22]  Qingming Huang,et al.  Random partial paired comparison for subjective video quality assessment via hodgerank , 2011, ACM Multimedia.

[23]  Phuoc Tran-Gia,et al.  Best Practices for QoE Crowdtesting: QoE Assessment With Crowdsourcing , 2014, IEEE Transactions on Multimedia.

[24]  Damon M. Chandler,et al.  ${\bf S}_{3}$: A Spectral and Spatial Measure of Local Perceived Sharpness in Natural Images , 2012, IEEE Transactions on Image Processing.

[25]  Alan C. Bovik,et al.  Making a “Completely Blind” Image Quality Analyzer , 2013, IEEE Signal Processing Letters.

[26]  Martin Schader,et al.  Exploring task properties in crowdsourcing - an empirical study on mechanical turk , 2011, ECIS.

[27]  Ashish Kapoor,et al.  Learning a blind measure of perceptual image quality , 2011, CVPR 2011.

[28]  Alan C. Bovik,et al.  Crowdsourced study of subjective image quality , 2014, 2014 48th Asilomar Conference on Signals, Systems and Computers.

[29]  Alan C. Bovik,et al.  No-Reference Image Quality Assessment in the Spatial Domain , 2012, IEEE Transactions on Image Processing.

[30]  Alan C. Bovik,et al.  RRED Indices: Reduced Reference Entropic Differencing for Image Quality Assessment , 2012, IEEE Transactions on Image Processing.

[31]  Geoffrey M. Boynton,et al.  New model of human luminance pattern vision mechanisms: analysis of the effects of pattern orientation, spatial phase, and temporal frequency , 1994, Other Conferences.

[32]  David S. Doermann,et al.  Active Sampling for Subjective Image Quality Assessment , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[33]  Hocine Cherifi,et al.  A comparison of image quality models and metrics based on human visual sensitivity , 1998, Proceedings 1998 International Conference on Image Processing. ICIP98 (Cat. No.98CB36269).

[34]  David S. Doermann,et al.  No-Reference Image Quality Assessment Using Visual Codebooks , 2012, IEEE Transactions on Image Processing.

[35]  O. Dykstra Rank Analysis of Incomplete Block Designs: A Method of Paired Comparisons Employing Unequal Repetitions on Pairs , 1960 .

[36]  Yuan Yao,et al.  Statistical ranking and combinatorial Hodge theory , 2008, Math. Program..

[37]  Alan C. Bovik,et al.  Objective quality assessment of multiply distorted images , 2012, 2012 Conference Record of the Forty Sixth Asilomar Conference on Signals, Systems and Computers (ASILOMAR).

[38]  Chin-Laung Lei,et al.  A crowdsourceable QoE evaluation framework for multimedia content , 2009, ACM Multimedia.

[39]  Graham D. Finlayson,et al.  Web-based Image Preference , 2013 .

[40]  Christian Keimel,et al.  QualityCrowd — A framework for crowd-based quality evaluation , 2012, 2012 Picture Coding Symposium.

[41]  David S. Doermann,et al.  No-reference image quality assessment based on visual codebook , 2011, 2011 18th IEEE International Conference on Image Processing.

[42]  Nikolay N. Ponomarenko,et al.  TID2008 – A database for evaluation of full-reference visual quality assessment metrics , 2004 .

[43]  Javed A. Aslam,et al.  Condorcet fusion for improved retrieval , 2002, CIKM '02.

[44]  Paul S. Fisher,et al.  Image quality measures and their performance , 1995, IEEE Trans. Commun..

[45]  Nikolay N. Ponomarenko,et al.  Color image database TID2013: Peculiarities and preliminary results , 2013, European Workshop on Visual Information Processing (EUVIP).

[46]  Margaret H. Pinson,et al.  A new standardized method for objectively measuring video quality , 2004, IEEE Transactions on Broadcasting.

[47]  Rajiv Soundararajan,et al.  Study of Subjective and Objective Quality Assessment of Video , 2010, IEEE Transactions on Image Processing.

[48]  Zhou Wang,et al.  Image Quality Assessment: From Error Measurement to Structural Similarity , 2004 .

[49]  Pavel Korshunov,et al.  Crowdsourcing-based multimedia subjective evaluations: a case study on image recognizability and aesthetic appeal , 2013, CrowdMM '13.

[50]  Dinei A. F. Florêncio,et al.  Crowdsourcing subjective image quality evaluation , 2011, 2011 18th IEEE International Conference on Image Processing.

[51]  Christophe Charrier,et al.  Blind Prediction of Natural Video Quality , 2014, IEEE Transactions on Image Processing.

[52]  Alan C. Bovik,et al.  Automatic quality prediction of authentically distorted pictures , 2015 .

[53]  Alan C. Bovik,et al.  Blind Image Quality Assessment: From Natural Scene Statistics to Perceptual Quality , 2011, IEEE Transactions on Image Processing.

[54]  Alan C. Bovik,et al.  C-DIIVINE: No-reference image quality assessment based on local magnitude and phase statistics of natural scenes , 2014, Signal Process. Image Commun..

[55]  Ashish Kapoor,et al.  Blind Image Quality Assessment Using Semi-supervised Rectifier Networks , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[56]  Alan C. Bovik,et al.  The Essential Guide to Video Processing , 2009, J. Electronic Imaging.