A comparison between shallow and deep architecture classifiers on small dataset

Many machine learning algorithms have been introduced to solve different types of problem. Recently, many of these algorithms have been applied to deep architecture model and showed very impressive performance. In general, deep architecture model suffers from over-fitting problem when there is a small number of training data. In this paper, we attempted to remedy this problem in deep architecture with regularization techniques including overlap pooling, flipped-image augmentation and dropout, and we also compared a deep structure model (convolutional neural network (CNN)) with shallow structure models (support vector machine and artificial neural network with one hidden layer) on a small dataset. It was statistically confirmed that the shallow models achieved better performance than the deep model that did not use a regularization technique. However, a deep model augmented with a regularization technique-CNN with dropout technique-was competitive to the shallow models.

[1]  Ah Chung Tsoi,et al.  Face recognition: a convolutional neural-network approach , 1997, IEEE Trans. Neural Networks.

[2]  Stefan Winkler,et al.  Deep Learning for Emotion Recognition on Small Datasets using Transfer Learning , 2015, ICMI.

[3]  Josef van Genabith,et al.  A Comparative Evaluation of Deep and Shallow Approaches to the Automatic Detection of Common Grammatical Errors , 2007, EMNLP.

[4]  Kitsuchart Pasupa,et al.  Sparse multinomial kernel discriminant analysis (sMKDA) , 2009, Pattern Recognit..

[5]  Wisuwat Sunhem,et al.  An approach to face shape classification for hairstyle recommendation , 2016, 2016 Eighth International Conference on Advanced Computational Intelligence (ICACI).

[6]  Saiful Anwar,et al.  Robustness Analysis of Artificial Neural Networks and Support Vector Machine in Making Prediction , 2011, 2011 IEEE Ninth International Symposium on Parallel and Distributed Processing with Applications.

[7]  Nitish Srivastava,et al.  Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..

[8]  Wisuwat Sunhem,et al.  Hairstyle recommendation system for women , 2016, 2016 Fifth ICT International Student Project Conference (ICT-ISPC).

[9]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[10]  Kunihiko Fukushima,et al.  Neocognitron: A self-organizing neural network model for a mechanism of pattern recognition unaffected by shift in position , 1980, Biological Cybernetics.

[11]  Keum-Shik Hong,et al.  Comparison of artificial neural network and support vector machine classifications for fNIRS-based BCI , 2015, 2015 15th International Conference on Control, Automation and Systems (ICCAS).

[12]  Peter Norvig,et al.  Artificial Intelligence: A Modern Approach , 1995 .

[13]  Jose Jesus Castro-Schez,et al.  A Recommender System Based on a Machine Learning Algorithm for B2C Portals , 2009, 2009 IEEE/WIC/ACM International Joint Conference on Web Intelligence and Intelligent Agent Technology.

[14]  D. Inman Machine learning applied to recognising hand-written Japanese , 1995, Proceedings 4th IEEE International Workshop on Robot and Human Communication.

[15]  Abhilasha Sharma,et al.  Image understanding using decision tree based machine learning , 2011, ICIMU 2011 : Proceedings of the 5th international Conference on Information Technology & Multimedia.

[16]  Sven Behnke,et al.  Evaluation of Pooling Operations in Convolutional Architectures for Object Recognition , 2010, ICANN.

[17]  Reda Alhajj,et al.  Text summarization techniques: SVM versus neural networks , 2009, iiWAS.

[18]  Rich Caruana,et al.  An empirical comparison of supervised learning algorithms , 2006, ICML.