A New Knowledge Distillation for Incremental Object Detection

Nowadays, the Convolutional Neural Network is successfully applied to the images object detection. When new classes of object emerges, it is popular to adapt the convolutional neural network based detection model through a retraining process with the new classes of samples. Unfortunately, the adapted model can only detect the new classes of objects, but cannot identify the old classes of objects, which is called catastrophic forgetting, also occurring in incremental classification tasks. Knowledge distillation has achieved good results in incremental learning for classification tasks. Due to the dual tasks within object detection, object classification and location at the same time, a straightforward migration of knowledge distillation method cannot provide a satisfactory result in incremental learning for object detection tasks. Hence, this paper propose a new knowledge distillation for incremental object detection, which introduces a new object detection distillation loss, a loss not only for classification results but also for location results of the predicted bounding boxes, not only for all final detected regions of interest but also for all intermediate regions proposal. Furthermore, to avoid forgetting learned knowledge from old datasets, this paper not only employs hint learning to retain the characteristic information of the initial model, but also innovatively uses confidence loss to extract the confidence information of the initial model. A series of experiment results on the PASCAL VOC 2007 dataset verify the effectiveness of the proposed method.

[1]  Geoffrey E. Hinton,et al.  Distilling the Knowledge in a Neural Network , 2015, ArXiv.

[2]  Martin Jägersand,et al.  Incremental learning for robot perception through HRI , 2017, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[3]  Faisal Shafait,et al.  Distillation Techniques for Pseudo-rehearsal Based Incremental Learning , 2018, ArXiv.

[4]  Ross B. Girshick,et al.  Fast R-CNN , 2015, 1504.08083.

[5]  Jian Sun,et al.  Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[6]  Heiko Wersing,et al.  Combining offline and online classifiers for life-long learning , 2015, 2015 International Joint Conference on Neural Networks (IJCNN).

[7]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[8]  Barbara Hammer,et al.  Incremental learning algorithms and applications , 2016, ESANN.

[9]  Trevor Darrell,et al.  Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation , 2013, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[10]  Luc Van Gool,et al.  The Pascal Visual Object Classes (VOC) Challenge , 2010, International Journal of Computer Vision.

[11]  Kaiming He,et al.  Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[12]  Yoshua Bengio,et al.  FitNets: Hints for Thin Deep Nets , 2014, ICLR.

[13]  Tony X. Han,et al.  Learning Efficient Object Detection Models with Knowledge Distillation , 2017, NIPS.

[14]  Kui-Ting Chen,et al.  A Convolutional Neural Network with Incremental Learning , 2017 .