Robust Multimodal Brain Tumor Segmentation via Feature Disentanglement and Gated Fusion

Accurate medical image segmentation commonly requires effective learning of the complementary information from multimodal data. However, in clinical practice, we often encounter the problem of missing imaging modalities. We tackle this challenge and propose a novel multimodal segmentation framework which is robust to the absence of imaging modalities. Our network uses feature disentanglement to decompose the input modalities into the modality-specific appearance code, which uniquely sticks to each modality, and the modality-invariant content code, which absorbs multimodal information for the segmentation task. With enhanced modality-invariance, the disentangled content code from each modality is fused into a shared representation which gains robustness to missing data. The fusion is achieved via a learning-based strategy to gate the contribution of different modalities at different locations. We validate our method on the important yet challenging multimodal brain tumor segmentation task with the BRATS challenge dataset. With competitive performance to the state-of-the-art approaches for full modality, our method achieves outstanding robustness under various missing modality(ies) situations, significantly exceeding the state-of-the-art method by over \(16\%\) in average for Dice on whole tumor segmentation.

[1]  Konstantinos Kamnitsas,et al.  Efficient multi‐scale 3D CNN with fully connected CRF for accurate brain lesion segmentation , 2016, Medical Image Anal..

[2]  Kuan-Lun Tseng,et al.  Joint Sequence Learning and Cross-Modality Convolution for 3D Biomedical Segmentation , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[3]  Dacheng Tao,et al.  One-Pass Multi-task Convolutional Neural Networks for Efficient Brain Tumor Segmentation , 2018, MICCAI.

[4]  Maneesh Kumar Singh,et al.  DRIT++: Diverse Image-to-Image Translation via Disentangled Representations , 2019, International Journal of Computer Vision.

[5]  Ruslan Salakhutdinov,et al.  Learning Factorized Multimodal Representations , 2018, ICLR.

[6]  Brian B. Avants,et al.  The Multimodal Brain Tumor Image Segmentation Benchmark (BRATS) , 2015, IEEE Transactions on Medical Imaging.

[7]  Marleen de Bruijne,et al.  Learning Cross-Modality Representations From Multi-Modal Images , 2019, IEEE Trans. Medical Imaging.

[8]  Hayit Greenspan,et al.  Improving CNN Training using Disentanglement for Liver Lesion Classification in CT , 2018, 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC).

[9]  Yan Shen,et al.  Brain Tumor Segmentation on MRI with Missing Modalities , 2019, IPMI.

[10]  Jan Kautz,et al.  Multimodal Unsupervised Image-to-Image Translation , 2018, ECCV.

[11]  Daniel Rueckert,et al.  Unsupervised Deformable Registration for Multi-Modal Images via Disentangled Representations , 2019, IPMI.

[12]  Mohammad Havaei,et al.  HeMIS: Hetero-Modal Image Segmentation , 2016, MICCAI.

[13]  Marleen de Bruijne,et al.  Why Does Synthesized Data Improve Multi-sequence Classification? , 2015, MICCAI.

[14]  Sotirios A. Tsaftaris,et al.  Multimodal MR Synthesis via Modality-Invariant Latent Representation , 2018, IEEE Transactions on Medical Imaging.

[15]  Thomas Brox,et al.  3D U-Net: Learning Dense Volumetric Segmentation from Sparse Annotation , 2016, MICCAI.

[16]  Gijs van Tulder,et al.  Learning Cross-Modality Representations From Multi-Modal Images , 2019, IEEE Transactions on Medical Imaging.

[17]  Sébastien Ourselin,et al.  Scalable multimodal convolutional networks for brain tumour segmentation , 2017, MICCAI.

[18]  Yong Fan,et al.  A deep learning model integrating FCNNs and CRFs for brain tumor segmentation , 2017, Medical Image Anal..

[19]  Sotirios A. Tsaftaris,et al.  Factorised spatial representation learning: application in semi-supervised myocardial segmentation , 2018, MICCAI.