More Discriminative CNN with Inter Loss for Classification

Recently years, convolutional neural networks (CNN) has been a hot spot in various areas such as object detection, classification. As deep study in CNN, its performance is almost human-competitive. We find that the test accuracy largely depends on the relationship of samples in feature space. Softmax loss is widely used in many deep learning algorithms. However, it cannot directly reflect this kind of relationship. In this paper, we design a new loss function, named inter loss. This inter loss function can maximizes the distance between different classes, analogous to maximizing margin in SVM. By integrating inter loss and softmax loss, larger inter-class distance and smaller intra-class distance can be obtained. In this way, we can significantly improve the accuracy in classification. Impressive results is obtained in SVHN and CIFAR-10 datasets. However, our main goal is to introduce a novel loss function tasks rather than beating the state-of-the-art. In our experiments, other forms of loss functions based on inter and intra class distance is also considered as to demonstrate the effectiveness of inter loss.

[1]  Huimin Lu,et al.  FDCNet: filtering deep convolutional network for marine organism classification , 2018, Multimedia Tools and Applications.

[2]  Yu Qiao,et al.  A Discriminative Feature Learning Approach for Deep Face Recognition , 2016, ECCV.

[3]  Peter Christoffersen,et al.  Série Scientifique Scientific Series the Importance of the Loss Function in Option Valuation the Importance of the Loss Function in Option Valuation , 2022 .

[4]  Thomas Serre,et al.  Object recognition with features inspired by visual cortex , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05).

[5]  Xiang Zhang,et al.  OverFeat: Integrated Recognition, Localization and Detection using Convolutional Networks , 2013, ICLR.

[6]  Yoram Singer,et al.  Adaptive Subgradient Methods for Online Learning and Stochastic Optimization , 2011, J. Mach. Learn. Res..

[7]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[8]  Jian Sun,et al.  Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[9]  Dumitru Erhan,et al.  Going deeper with convolutions , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[10]  Yiannis Aloimonos,et al.  LightNet: A Versatile, Standalone Matlab-based Environment for Deep Learning , 2016, ACM Multimedia.

[11]  Lorenzo Rosasco,et al.  Are Loss Functions All the Same? , 2004, Neural Computation.

[12]  Trevor Darrell,et al.  Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation , 2013, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[13]  Yan Wang,et al.  DeepContour: A deep convolutional feature learned by positive-sharing loss for contour detection , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[14]  Jitendra Malik,et al.  Learning Rich Features from RGB-D Images for Object Detection and Segmentation , 2014, ECCV.

[15]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[16]  Bin Li,et al.  Wound intensity correction and segmentation with convolutional neural networks , 2017, Concurr. Comput. Pract. Exp..

[17]  Yoav Freund,et al.  A decision-theoretic generalization of on-line learning and an application to boosting , 1997, EuroCOLT.

[18]  Nader Ebrahimi,et al.  Bayesian approach to life testing and reliability estimation using asymmetric loss function , 1991 .

[19]  Yoshua Bengio,et al.  Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.

[20]  Sergey Ioffe,et al.  Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.

[21]  Varun Ramakrishna,et al.  Convolutional Pose Machines , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[22]  Kaiming He,et al.  Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[23]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[24]  Nitish Srivastava,et al.  Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..

[25]  Camille Couprie,et al.  Learning Hierarchical Features for Scene Labeling , 2013, IEEE Transactions on Pattern Analysis and Machine Intelligence.