Combining Heterogeneously Labeled Datasets For Training Segmentation Networks

Accurate segmentation of medical images is an important step towards analyzing and tracking disease related morphological alterations in the anatomy. Convolutional neural networks (CNNs) have recently emerged as a powerful tool for many segmentation tasks in medical imaging. The performance of CNNs strongly depends on the size of the training data and combining data from different sources is an effective strategy for obtaining larger training datasets. However, this is often challenged by heterogeneous labeling of the datasets. For instance, one of the dataset may be missing labels or a number of labels may have been combined into a super label. In this work we propose a cost function which allows integration of multiple datasets with heterogeneous label subsets into a joint training. We evaluated the performance of this strategy on thigh MR and a cardiac MR datasets in which we artificially merged labels for half of the data. We found the proposed cost function substantially outperforms a naive masking approach, obtaining results very close to using the full annotations.

[1]  Christoph Meinel,et al.  Deep Learning for Medical Image Analysis , 2018, Journal of Pathology Informatics.

[2]  Erika Schneider,et al.  The osteoarthritis initiative: report on the design rationale for the magnetic resonance imaging protocol for the knee. , 2008, Osteoarthritis and cartilage.

[3]  Thomas Brox,et al.  U-Net: Convolutional Networks for Biomedical Image Segmentation , 2015, MICCAI.

[4]  Syed Muhammad Anwar,et al.  Deep Learning in Medical Image Analysis , 2017 .

[5]  Heung-Il Suk,et al.  Deep Learning in Medical Image Analysis. , 2017, Annual review of biomedical engineering.

[6]  Thomas Brox,et al.  3D U-Net: Learning Dense Volumetric Segmentation from Sparse Annotation , 2016, MICCAI.

[7]  Jeffrey W. Prescott,et al.  Quantitative Imaging Biomarkers: The Application of Advanced Image Processing and Analysis to Clinical and Preclinical Decision Making , 2013, Journal of Digital Imaging.

[8]  Xin Yang,et al.  Deep Learning Techniques for Automatic MRI Cardiac Multi-Structures Segmentation and Diagnosis: Is the Problem Solved? , 2018, IEEE Transactions on Medical Imaging.

[9]  Mert R. Sabuncu,et al.  An algorithm for optimal fusion of atlases with different labeling protocols , 2015, NeuroImage.

[10]  W. Wirth,et al.  The role of thigh muscle and adipose tissue in knee osteoarthritis progression in women: data from the Osteoarthritis Initiative. , 2018, Osteoarthritis and cartilage.

[11]  Marc Pollefeys,et al.  An Exploration of 2D and 3D Deep Learning Techniques for Cardiac MR Image Segmentation , 2017, STACOM@MICCAI.

[12]  F. Eckstein,et al.  Thigh Muscle Cross‐Sectional Areas and Strength in Advanced Versus Early Painful Osteoarthritis: An Exploratory Between‐Knee, Within‐Person Comparison in Osteoarthritis Initiative Participants , 2013, Arthritis care & research.

[13]  F. Eckstein,et al.  Thigh muscle and adipose tissue changes during symptomatic and radiographic knee osteoarthritis progression – data from the osteoarthritis initiative , 2018 .

[14]  Bram van Ginneken,et al.  A survey on deep learning in medical image analysis , 2017, Medical Image Anal..

[15]  F Eckstein,et al.  Longitudinal (4 year) change of thigh muscle and adipose tissue distribution in chronically painful vs painless knees--data from the Osteoarthritis Initiative. , 2015, Osteoarthritis and cartilage.