Identifying Label Errors in Object Detection Datasets by Loss Inspection

Labeling datasets for supervised object detection is a dull and time-consuming task. Errors can be easily introduced during annotation and overlooked during review, yielding inaccurate benchmarks and performance degradation of deep neural networks trained on noisy labels. In this work, we for the first time introduce a benchmark for label error detection methods on object detection datasets as well as a label error detection method and a number of baselines. We simulate four different types of randomly introduced label errors on train and test sets of well-labeled object detection datasets. For our label error detection method we assume a two-stage object detector to be given and consider the sum of both stages' classification and regression losses. The losses are computed with respect to the predictions and the noisy labels including simulated label errors, aiming at detecting the latter. We compare our method to three baselines: a naive one without deep learning, the object detector's score and the entropy of the classification softmax distribution. We outperform all baselines and demonstrate that among the considered methods, ours is the only one that detects label errors of all four types efficiently. Furthermore, we detect real label errors a) on commonly used test datasets in object detection and b) on a proprietary dataset. In both cases we achieve low false positives rates, i.e., when considering 200 proposals from our method, we detect label errors with a precision for a) of up to 71.5% and for b) with 97%.

[1]  M. Rottmann,et al.  Automated Detection of Label Errors in Semantic Segmentation Datasets via Deep Learning and Uncertainty Quantification , 2022, 2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV).

[2]  M. Rottmann,et al.  Towards Rapid Prototyping and Comparability in Active Learning for Deep Object Detection , 2022, ArXiv.

[3]  Curtis G. Northcutt,et al.  Identifying Incorrect Annotations in Multi-Label Classification Data , 2022, ArXiv.

[4]  Nikos Arechiga,et al.  Finding Label and Model Errors in Perception Data With Learned Observation Assertions , 2022, SIGMOD Conference.

[5]  Ismail Elezi,et al.  Not All Labels Are Equal: Rationalizing The Labeling Costs for Training Object Detection , 2021, 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[6]  Masayoshi Tomizuka,et al.  Labels are Not Perfect: Inferring Spatial Uncertainty in Object Detection , 2020, IEEE Transactions on Intelligent Transportation Systems.

[7]  Nirvair Neeru,et al.  A Survey on Deep Learning Approaches to Medical Images and a Systematic Look up into Real-Time Object Detection , 2021, Archives of Computational Methods in Engineering.

[8]  Jonas Mueller,et al.  Pervasive Label Errors in Test Sets Destabilize Machine Learning Benchmarks , 2021, NeurIPS Datasets and Benchmarks.

[9]  Richard Bowden,et al.  A Survey of Deep Learning Applications to Autonomous Vehicle Control , 2019, IEEE Transactions on Intelligent Transportation Systems.

[10]  Isaac L. Chuang,et al.  Confident Learning: Estimating Uncertainty in Dataset Labels , 2019, J. Artif. Intell. Res..

[11]  Klaus C. J. Dietmayer,et al.  Deep Multi-Modal Object Detection and Semantic Segmentation for Autonomous Driving: Datasets, Methods, and Challenges , 2019, IEEE Transactions on Intelligent Transportation Systems.

[12]  R. Socher,et al.  Learning From Noisy Anchors for One-Stage Object Detection , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[13]  Trevor Darrell,et al.  BDD100K: A Diverse Driving Dataset for Heterogeneous Multitask Learning , 2018, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[14]  James Bailey,et al.  Symmetric Cross Entropy for Robust Learning With Noisy Labels , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[15]  Wei Guo,et al.  An Adaptive Supervision Framework for Active Learning in Object Detection , 2019, BMVC.

[16]  Haichao Zhang,et al.  Towards Adversarially Robust Object Detection , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[17]  Kai Chen,et al.  MMDetection: Open MMLab Detection Toolbox and Benchmark , 2019, ArXiv.

[18]  Pengfei Chen,et al.  Understanding and Utilizing Deep Neural Networks Trained with Noisy Labels , 2019, ICML.

[19]  Klaus H. Maier-Hein,et al.  Retina U-Net: Embarrassingly Simple Exploitation of Segmentation Supervision for Medical Object Detection , 2018, ML4H@NeurIPS.

[20]  Joachim Denzler,et al.  Active Learning for Deep Object Detection , 2018, VISIGRAPP.

[21]  Larry S. Davis,et al.  Soft Sampling for Robust Object Detection , 2018, BMVC.

[22]  Yizhou Wang,et al.  L_DMI: A Novel Information-theoretic Loss Function for Training Deep Nets Robust to Label Noise , 2019, NeurIPS.

[23]  Bernard Ghanem,et al.  Missing Labels in Object Detection , 2019, CVPR Workshops.

[24]  Sherali Zeadally,et al.  Autonomous Cars: Research Results, Issues, and Future Challenges , 2019, IEEE Communications Surveys & Tutorials.

[25]  Xingrui Yu,et al.  Co-teaching: Robust training of deep neural networks with extremely noisy labels , 2018, NeurIPS.

[26]  Kevin Gimpel,et al.  Using Trusted Data to Train Deep Networks on Labels Corrupted by Severe Noise , 2018, NeurIPS.

[27]  Li Fei-Fei,et al.  MentorNet: Learning Data-Driven Curriculum for Very Deep Neural Networks on Corrupted Labels , 2017, ICML.

[28]  Hongyi Zhang,et al.  mixup: Beyond Empirical Risk Minimization , 2017, ICLR.

[29]  Niko Sünderhauf,et al.  Dropout Sampling for Robust Object Detection in Open-Set Conditions , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[30]  Vinay P. Namboodiri,et al.  Deep active learning for object detection , 2018, BMVC.

[31]  Chen Sun,et al.  Revisiting Unreasonable Effectiveness of Data in Deep Learning Era , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[32]  Gregory Cohen,et al.  EMNIST: Extending MNIST to handwritten letters , 2017, 2017 International Joint Conference on Neural Networks (IJCNN).

[33]  Jacob Goldberger,et al.  Training deep neural-networks using a noise adaptation layer , 2016, ICLR.

[34]  Dumitru Erhan,et al.  Training Deep Neural Networks on Noisy Labels with Bootstrapping , 2014, ICLR.

[35]  Pietro Perona,et al.  Microsoft COCO: Common Objects in Context , 2014, ECCV.

[36]  Andreas Geiger,et al.  Are we ready for autonomous driving? The KITTI vision benchmark suite , 2012, 2012 IEEE Conference on Computer Vision and Pattern Recognition.

[37]  Luc Van Gool,et al.  The Pascal Visual Object Classes (VOC) Challenge , 2010, International Journal of Computer Vision.

[38]  Burr Settles,et al.  Active Learning Literature Survey , 2009 .

[39]  Mark Goadrich,et al.  The relationship between Precision-Recall and ROC curves , 2006, ICML.

[40]  Walt Detmar Meurers,et al.  Detecting Errors in Part-of-Speech Annotation , 2003, EACL.

[41]  L. R. Dice Measures of the Amount of Ecologic Association Between Species , 1945 .