BundleNet: Learning with Noisy Label via Sample Correlations

Sequential patterns are important, because they can be exploited to improve the prediction accuracy of our classifiers. Sequential data, such as time series/video frames, and event data are becoming more and more ubiquitous in a wide spectrum of application scenarios especially in the background of large data and deep learning. However, large data sets used in training modern machine-learning models, such as deep neural networks, are often affected by label noise. Existing noisy learning approaches mainly focus on building an additional network to clean the noise or find a robust loss function. Few works tackle this problem by exploiting sample correlations. In this paper, we propose BundleNet, a framework of sequential structure (named bundle-module, see Fig. 1) for deep neural networks to handle the label noise. The bundle module naturally takes into account sample correlations by constructing bundles of samples class-by-class, and treats them as independent inputs. Moreover, we prove that the bundle-module performs a form of regularization, which is similar to dropout as regularization during training. The regularization effect endows the BundleNet with strong robustness to the label noise. Extensive experiments on public data sets prove that the proposed approach is effective and promising.

[1]  Aritra Ghosh,et al.  Robust Loss Functions under Label Noise for Deep Neural Networks , 2017, AAAI.

[2]  Qi Tian,et al.  Boosted random contextual semantic space based representation for visual recognition , 2016, Inf. Sci..

[3]  Ali Farhadi,et al.  Learning Everything about Anything: Webly-Supervised Visual Concept Learning , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[4]  Junchi Yan,et al.  Sales pipeline win propensity prediction: A regression approach , 2015, 2015 IFIP/IEEE International Symposium on Integrated Network Management (IM).

[5]  Nagarajan Natarajan,et al.  Learning with Noisy Labels , 2013, NIPS.

[6]  Sven Behnke,et al.  Object class segmentation of RGB-D video using recurrent convolutional neural networks , 2017, Neural Networks.

[7]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[8]  Qi Tian,et al.  Contextual Exemplar Classifier-Based Image Representation for Classification , 2017, IEEE Transactions on Circuits and Systems for Video Technology.

[9]  Fei-Fei Li,et al.  What's the Point: Semantic Segmentation with Point Supervision , 2015, ECCV.

[10]  Zenglin Xu,et al.  Multi-View Learning with Limited and Noisy Tagging , 2016, IJCAI.

[11]  William H. Richardson,et al.  Bayesian-Based Iterative Method of Image Restoration , 1972 .

[12]  Rob Fergus,et al.  Learning from Noisy Labels with Deep Neural Networks , 2014, ICLR.

[13]  Junchi Yan,et al.  Adaptive Discrete Hypergraph Matching , 2018, IEEE Transactions on Cybernetics.

[14]  Pietro Perona,et al.  Learning Object Categories From Internet Image Searches , 2010, Proceedings of the IEEE.

[15]  Sida I. Wang,et al.  Dropout Training as Adaptive Regularization , 2013, NIPS.

[16]  Kilian Q. Weinberger,et al.  Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[17]  Albert Fornells,et al.  A study of the effect of different types of noise on the precision of supervised learning techniques , 2010, Artificial Intelligence Review.

[18]  Qi Tian,et al.  Structured Weak Semantic Space Construction for Visual Categorization , 2018, IEEE Transactions on Neural Networks and Learning Systems.

[19]  Jonathan Krause,et al.  The Unreasonable Effectiveness of Noisy Data for Fine-Grained Recognition , 2015, ECCV.

[20]  Li Fei-Fei,et al.  ImageNet: Constructing a large-scale image database , 2010 .

[21]  Hongyuan Zha,et al.  A constrained clustering based approach for matching a collection of feature sets , 2016, 2016 23rd International Conference on Pattern Recognition (ICPR).

[22]  Antonio Criminisi,et al.  Harvesting Image Databases from the Web , 2007, 2007 IEEE 11th International Conference on Computer Vision.

[23]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[24]  Changsheng Li,et al.  Active Sample Learning and Feature Selection: A Unified Approach , 2015, ArXiv.

[25]  Richard Nock,et al.  Making Deep Neural Networks Robust to Label Noise: A Loss Correction Approach , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[26]  J. Koerts Generalized linear models: Monographs on Statistics and Applied Probability, Chapman and Hall, London, 1983, xiii + 261 pages, £15.00 , 1984 .

[27]  M. Verleysen,et al.  Classification in the Presence of Label Noise: A Survey , 2014, IEEE Transactions on Neural Networks and Learning Systems.

[28]  Farshad Fotouhi,et al.  Adaptive Indexing in Very Large Databases , 1995 .

[29]  Pascal Vincent,et al.  Adding noise to the input of a model trained with a regularized objective , 2011, ArXiv.

[30]  Junchi Yan,et al.  Self-Paced MultiTask Learning , 2017 .

[31]  Qi Tian,et al.  Fine-Grained Image Classification via Low-Rank Sparse Coding With General and Class-Specific Codebooks , 2017, IEEE Transactions on Neural Networks and Learning Systems.

[32]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[33]  Carla E. Brodley,et al.  Identifying Mislabeled Training Data , 1999, J. Artif. Intell. Res..

[34]  Jian Sun,et al.  Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).

[35]  Benoît Frénay,et al.  A comprehensive introduction to label noise , 2014, ESANN.

[36]  Rocco A. Servedio,et al.  Random classification noise defeats all convex potential boosters , 2008, ICML '08.

[37]  Yuandong Tian,et al.  Scale-invariant learning and convolutional networks , 2015, ArXiv.

[38]  Christopher M. Bishop,et al.  Current address: Microsoft Research, , 2022 .

[39]  Johannes Gehrke,et al.  Cayuga: A General Purpose Event Monitoring System , 2007, CIDR.

[40]  Chao Zhang,et al.  Discrete Hypergraph Matching , 2015 .

[41]  Qi Tian,et al.  Incremental Codebook Adaptation for Visual Representation and Categorization , 2018, IEEE Transactions on Cybernetics.

[42]  Qi Tian,et al.  Image-Specific Classification With Local and Global Discriminations , 2018, IEEE Transactions on Neural Networks and Learning Systems.

[43]  Fouad Bahrpeyma,et al.  Improving the Accuracy of Active Learning Method via Noise Injection for Estimating Hydraulic Flow Units: An Example from a Heterogeneous Carbonate Reservoir , 2015 .

[44]  Qi Tian,et al.  Multiview Label Sharing for Visual Representations and Classifications , 2018, IEEE Transactions on Multimedia.

[45]  Nitish Srivastava,et al.  Improving neural networks by preventing co-adaptation of feature detectors , 2012, ArXiv.

[46]  Margaret Mitchell,et al.  VQA: Visual Question Answering , 2015, International Journal of Computer Vision.

[47]  Dong Xu,et al.  Visual recognition by learning from web data: A weakly supervised domain generalization approach , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[48]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[49]  Qi Tian,et al.  Beyond Explicit Codebook Generation: Visual Representation Using Implicitly Transferred Codebooks , 2015, IEEE Transactions on Image Processing.

[50]  Armand Joulin,et al.  Unsupervised Learning by Predicting Noise , 2017, ICML.

[51]  Jacob Goldberger,et al.  Training deep neural-networks using a noise adaptation layer , 2016, ICLR.

[52]  Hongyuan Zha,et al.  A Short Survey of Recent Advances in Graph Matching , 2016, ICMR.

[53]  Yoshua Bengio,et al.  Show, Attend and Tell: Neural Image Caption Generation with Visual Attention , 2015, ICML.

[54]  Moon Gi Kang,et al.  Super-resolution image reconstruction: a technical overview , 2003, IEEE Signal Process. Mag..

[55]  Qi Tian,et al.  Image-level classification by hierarchical structure learning with visual and semantic similarities , 2018, Inf. Sci..

[56]  Lu Wang,et al.  Risk Minimization in the Presence of Label Noise , 2016, AAAI.

[57]  Qingshan Liu,et al.  A Self-Paced Regularization Framework for Multilabel Learning , 2016, IEEE Transactions on Neural Networks and Learning Systems.

[58]  Kilian Q. Weinberger,et al.  Deep Networks with Stochastic Depth , 2016, ECCV.

[59]  Frank Nielsen,et al.  Loss factorization, weakly supervised learning and label noise robustness , 2016, ICML.

[60]  Abhinav Gupta,et al.  Learning from Noisy Large-Scale Datasets with Minimal Supervision , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[61]  Qi Tian,et al.  Image classification by search with explicitly and implicitly semantic representations , 2017, Inf. Sci..

[62]  Jian Sun,et al.  Identity Mappings in Deep Residual Networks , 2016, ECCV.

[63]  Sina Honari,et al.  Learning to Generate Samples from Noise through Infusion Training , 2017, ICLR.

[64]  Sergey Ioffe,et al.  Rethinking the Inception Architecture for Computer Vision , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[65]  Qi Tian,et al.  Bundled Local Features for Image Representation , 2018, IEEE Transactions on Circuits and Systems for Video Technology.

[66]  Alex Krizhevsky,et al.  Learning Multiple Layers of Features from Tiny Images , 2009 .

[67]  Qingshan Liu,et al.  Dynamic Structure Embedded Online Multiple-Output Regression for Streaming Data , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[68]  Qingshan Liu,et al.  Max-Margin-Based Discriminative Feature Learning , 2016, IEEE Transactions on Neural Networks and Learning Systems.

[69]  Dumitru Erhan,et al.  Going deeper with convolutions , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[70]  Cheng Soon Ong,et al.  Learning from Corrupted Binary Labels via Class-Probability Estimation , 2015, ICML.

[71]  Dit-Yan Yeung,et al.  Learning a Deep Compact Image Representation for Visual Tracking , 2013, NIPS.