AdaBoostNet: An Efficient Hierarchical Neural Network for Image Classification

Hierarchical Neural Networks [e.g., deep neural networks (DNNs)] have recently gained increasing attention for image classification task. Most of the previous works devote to stack multiple levels of feature learning modules with the hope that higher level modules can represent more abstract semantics of the image data. However, training of such hierarchical network is typically cast as a time-consuming non-convex optimization problem, and its effectiveness for image feature representation critically depends on expertise in parameter tuning with various ad hoc tricks. To address these issues, we advocate a biologically-inspired hierarchical neural network. One key philosophy is that the higher-level modules in the network should correct misclassification induced by the lower ones. Given this idea, a sequential stacking strategy of basic feature learning modules is presented. In practical, an efficient layer wise learning and decision aggregation method is applied to boost the hierarchical network, which is different from naive layers cascading and end-to-end finetuning, and we term it as AdaboostNet. Moreover, the basic feature learning module is set as extreme learning machine (ELM), an effective and cheaply-optimized model. Experiments are conducted on benchmark datasets to evaluate our claims. It shows that the proposed network achieves comparable or better image classification accuracy and training efficiency than traditional DNNs and hierarchical ELMs.

[1]  Guang-Bin Huang,et al.  Extreme Learning Machine for Multilayer Perceptron , 2016, IEEE Transactions on Neural Networks and Learning Systems.

[2]  D. Hubel,et al.  Receptive fields of single neurones in the cat's striate cortex , 1959, The Journal of physiology.

[3]  Geoffrey E. Hinton,et al.  Learning representations by back-propagating errors , 1986, Nature.

[4]  Chee Kheong Siew,et al.  Extreme learning machine: Theory and applications , 2006, Neurocomputing.

[5]  Guang-Bin Huang,et al.  Trends in extreme learning machines: A review , 2015, Neural Networks.

[6]  Sameer A. Nene,et al.  Columbia Object Image Library (COIL100) , 1996 .

[7]  Dumitru Erhan,et al.  Going deeper with convolutions , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[8]  L. C. Kasun,et al.  Representational Learning with Extreme Learning Machine for Big Data Liyanaarachchi , 2022 .

[9]  John C. Duchi,et al.  Learning Kernels with Random Features , 2016, NIPS.

[10]  Geoffrey E. Hinton,et al.  Reducing the Dimensionality of Data with Neural Networks , 2006, Science.

[11]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[12]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.

[13]  Yee Whye Teh,et al.  A Fast Learning Algorithm for Deep Belief Nets , 2006, Neural Computation.

[14]  D. Hubel,et al.  Early Exploration of the Visual Cortex , 1998, Neuron.

[15]  Yoav Freund,et al.  A decision-theoretic generalization of on-line learning and an application to boosting , 1995, EuroCOLT.

[16]  Bernt Schiele,et al.  Analyzing appearance and contour based methods for object categorization , 2003, 2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2003. Proceedings..

[17]  M. Wibral,et al.  The Faces of Predictive Coding , 2015, The Journal of Neuroscience.

[18]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[19]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[20]  Trevor Hastie,et al.  Multi-class AdaBoost ∗ , 2009 .