A feasibility study of an autoencoder meta-model for improving generalization capabilities on training sets of small sizes

A meta-model approach for autoencoders is presented.New modification of autoencoder's learning algorithm is proposed.Developed meta-model provides faster training in terms of consumed training samples. The problem of training autoencoders (with logistic regression as the classification layer) on sets of small sizes is considered on the example of image classification and scene categorization tasks. Conventional autoencoders with uniform priors usually fail to learn useful features from few samples. A possibility to overcome this difficulty is considered on the example of a proposed meta-model generating autoencoders from a vector of meta-parameters of much smaller dimension than the number of autoencoder parameters. Handcrafted image features designed for the task of scene classification are used as the baseline for comparison. Unbiased autoencoders showed the worst results on tiny training sets and seem to require much larger sets to outperform other methods. On the other hand, the developed biased autoencoders work better on the training sets of small sizes, but have much less discriminative power, since the considered meta-model assigns strictly zero probabilities to a large subset of solutions. General possibility of increasing training speed (in term of consumed training patterns) of autoencoders is confirmed.

[1]  David J. Field,et al.  Sparse coding with an overcomplete basis set: A strategy employed by V1? , 1997, Vision Research.

[2]  Justus H. Piater,et al.  Distinctive Features Should Be Learned , 2000, Biologically Motivated Computer Vision.

[3]  Jürgen Schmidhuber,et al.  Multi-column deep neural network for traffic sign classification , 2012, Neural Networks.

[4]  Luca Maria Gambardella,et al.  Deep, Big, Simple Neural Nets for Handwritten Digit Recognition , 2010, Neural Computation.

[5]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[6]  Trevor Darrell,et al.  Caffe: Convolutional Architecture for Fast Feature Embedding , 2014, ACM Multimedia.

[7]  Alexey Potapov,et al.  Limited Generalization Capabilities of Autoencoders with Logistic Regression on Training Sets of Small Sizes , 2014, AIAI.

[8]  Robert M. Haralick,et al.  Textural Features for Image Classification , 1973, IEEE Trans. Syst. Man Cybern..

[9]  Rachid Deriche,et al.  Using Canny's criteria to derive a recursively implemented optimal edge detector , 1987, International Journal of Computer Vision.

[10]  Matthijs C. Dorst Distinctive Image Features from Scale-Invariant Keypoints , 2011 .

[11]  Luca Maria Gambardella,et al.  Deep Big Simple Neural Nets Excel on Handwritten Digit Recognition , 2010, ArXiv.

[12]  Andrew Zisserman,et al.  Smooth object retrieval using a bag of boundaries , 2011, 2011 International Conference on Computer Vision.

[13]  Antonio Torralba,et al.  Recognizing indoor scenes , 2009, CVPR.

[14]  Manik Varma,et al.  Character Recognition in Natural Images , 2009, VISAPP.

[15]  Alex Krizhevsky,et al.  Learning Multiple Layers of Features from Tiny Images , 2009 .

[16]  Guy M. Wallis,et al.  Using Spatio-temporal Correlations to Learn Invariant Object Recognition , 1996, Neural Networks.

[17]  Anil K. Jain,et al.  On image classification: city images vs. landscapes , 1998, Pattern Recognit..

[18]  Thomas Hofmann,et al.  Greedy Layer-Wise Training of Deep Networks , 2007 .

[19]  Yoshua Bengio,et al.  Extracting and composing robust features with denoising autoencoders , 2008, ICML '08.

[20]  Jürgen Schmidhuber,et al.  Deep learning in neural networks: An overview , 2014, Neural Networks.

[21]  Yann LeCun,et al.  The mnist database of handwritten digits , 2005 .

[22]  Yee Whye Teh,et al.  A Fast Learning Algorithm for Deep Belief Nets , 2006, Neural Computation.

[23]  Charles Kemp,et al.  How to Grow a Mind: Statistics, Structure, and Abstraction , 2011, Science.

[24]  Ray J. Solomonoff,et al.  Algorithmic Probability, Heuristic Programming and AGI , 2010, AGI 2010.

[25]  Marc'Aurelio Ranzato,et al.  Efficient Learning of Sparse Representations with an Energy-Based Model , 2006, NIPS.