Using Crowdsourcing for Multi-label Biomedical Compound Figure Annotation

Information analysis or retrieval for images in the biomedical literature needs to deal with a large amount of compound figures (figures containing several subfigures), as they constitute probably more than half of all images in repositories such as PubMed Central, which was the data set used for the task. The ImageCLEFmed benchmark proposed among other tasks in 2015 and 2016 a multi-label classification task, which aims at evaluating the automatic classification of figures into 30 image types. This task was based on compound figures and thus the figures were distributed to participants as compound figures but also in a separated form. Therefore, the generation of a gold standard was required, so that algorithms of participants can be evaluated and compared. This work presents the process carried out to generate the multi-labels of \(\sim \,2650\) compound figures using a crowdsourcing approach. Automatic algorithms to separate compound figures into subfigures were used and the results were then validated or corrected via crowdsourcing. The image types (MR, CT, X–ray, ...) were also annotated by crowdsourcing including detailed quality control. Quality control is necessary to insure quality of the annotated data as much as possible. \(\sim \,625\) h were invested with a cost of \(\sim \,870\$\).

[1]  Tat-Seng Chua,et al.  NUS-WIDE: a real-world web image database from National University of Singapore , 2009, CIVR '09.

[2]  Henning Müller,et al.  Overview of the ImageCLEF 2015 Medical Classification Task , 2015, CLEF.

[3]  Henning Müller,et al.  The medGIFT Group in ImageCLEFmed 2013 , 2013, CLEF.

[4]  Stefanie Nowak,et al.  How reliable are annotations via crowdsourcing: a study about inter-annotator agreement for multi-label image annotation , 2010, MIR '10.

[5]  Henning Müller,et al.  Overview of the ImageCLEF 2013 Medical Tasks , 2013, CLEF.

[6]  Henning Müller,et al.  Crowdsourcing for Medical Image Classification , 2014 .

[7]  Henning Müller,et al.  Ground truth generation in medical imaging: a crowdsourcing-based iterative approach , 2012, CrowdMM '12.

[8]  Henning Müller,et al.  Semi-supervised Learning for Image Modality Classification , 2015, MRDM@ECIR.

[9]  Henning Müller,et al.  Overview of the ImageCLEF 2016 Medical Task , 2016, CLEF.

[10]  William R. Hersh,et al.  Automatic Image Modality Based Classification and Annotation to Improve Medical Image Retrieval , 2007, MedInfo.

[11]  Henning Müller,et al.  Separating compound figures in journal articles to allow for subfigure classification , 2013, Medical Imaging.

[12]  T. Peto,et al.  Crowdsourcing as a Novel Technique for Retinal Fundus Photography Classification: Analysis of Images in the EPIC Norfolk Cohort on Behalf of the UKBiobank Eye and Vision Consortium , 2013, PloS one.

[13]  Elisa Bertino,et al.  Quality Control in Crowdsourcing Systems: Issues and Directions , 2013, IEEE Internet Computing.

[14]  Nassir Navab,et al.  AggNet: Deep Learning From Crowds for Mitosis Detection in Breast Cancer Histology Images , 2016, IEEE Trans. Medical Imaging.

[15]  Matthew Lease,et al.  On Quality Control and Machine Learning in Crowdsourcing , 2011, Human Computation.

[16]  Lena Maier-Hein,et al.  Crowdsourcing for Reference Correspondence Generation in Endoscopic Images , 2014, MICCAI.

[17]  Pierre Tirilly,et al.  On modality classification and its use in text-based image retrieval in medical databases , 2011, 2011 9th International Workshop on Content-Based Multimedia Indexing (CBMI).

[18]  Lei Zhang,et al.  Multi-label sparse coding for automatic image annotation , 2009, CVPR.