A Novel Loss Calibration Strategy for Object Detection Networks Training on Sparsely Annotated Pathological Datasets

Recently, object detection frameworks based on Convolutional Neural Networks (CNNs) have become powerful methods for various tasks of medical image analysis; however, they often struggle with most pathological datasets, which are impossible to annotate all the cells. Obviously, sparse annotations may lead to a seriously miscalculated loss in training, which limits the performance of networks. To address this limitation, we investigate the internal training process of object detection networks. Our core observation is that there is a significant density difference between the regression boxes of the positive instances and negative instances. Our novel Boxes Density Energy (BDE) focuses on utilizing the densities of regression boxes to conduct loss-calibration, which is dedicated to reducing the miscalculated loss, meanwhile to penalizing mispredictions with a relatively more significant loss. Thus BDE can guide networks to be trained along the right direction. Extensive experiments have demonstrated that, BDE on the sparsely annotated pathological dataset can significantly boost the performance of networks, and even with 1.0–1.5% higher recall than networks trained on the fully annotated dataset.

[1]  Yi Yang,et al.  Self-produced Guidance for Weakly-supervised Object Localization , 2018, ECCV.

[2]  Jin Xiong,et al.  HR-NET: A Highly Reliable Message-Passing Mechanism for Cluster File System , 2011, 2011 IEEE Sixth International Conference on Networking, Architecture, and Storage.

[3]  Quoc V. Le,et al.  NAS-FPN: Learning Scalable Feature Pyramid Architecture for Object Detection , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[4]  Takuya Akiba,et al.  Sampling Techniques for Large-Scale Object Detection From Sparsely Annotated Objects , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[5]  Bernard Ghanem,et al.  Missing Labels in Object Detection , 2019, CVPR Workshops.

[6]  Hao Chen,et al.  IRNet: Instance Relation Network for Overlapping Cervical Cell Segmentation , 2019, MICCAI.

[7]  Kaiming He,et al.  Feature Pyramid Networks for Object Detection , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[8]  Yuxing Peng,et al.  ThunderNet: Towards Real-Time Generic Object Detection on Mobile Devices , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[9]  Kiyoharu Aizawa,et al.  Cross-Domain Weakly-Supervised Object Detection Through Progressive Domain Adaptation , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[10]  Fuchun Sun,et al.  HyperNet: Towards Accurate Region Proposal Generation and Joint Object Detection , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[11]  Lianwen Jin,et al.  DeepText: A new approach for text proposal generation and text detection in natural images , 2017, 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[12]  Eugene W. Myers,et al.  Cell Detection with Star-convex Polygons , 2018, MICCAI.

[13]  Yi Yang,et al.  Adversarial Complementary Learning for Weakly Supervised Object Localization , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[14]  Kaiming He,et al.  Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[15]  Shiming Xiang,et al.  AugFPN: Improving Multi-Scale Feature Learning for Object Detection , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[16]  Yu Liu,et al.  Gradient Harmonized Single-stage Detector , 2018, AAAI.

[17]  Nassir Navab,et al.  Multiclass Deep Active Learning for Detecting Red Blood Cell Subtypes in Brightfield Microscopy , 2019, MICCAI.

[18]  Kaiming He,et al.  Focal Loss for Dense Object Detection , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).