The appropriateness of k-Sparse autoencoders in sparse coding

Learning of representations usually happens in different ways. Sometimes it persuades sparsity thus enhances performance through the task categorization. The sparse elements entail the learning algorithms that relate to the sparse-coding. Sometimes the algorithms have neural training networks with sparsity penalties and fines. The k-sparse autoencoder (KSA) model appears linear. The appropriateness of the model in sparse coding forms the foundation of this paper. Most important, the model appears speedily encoded and easily trained. Given these advantages, the model is suited for solving large-size issues or problems. We used openly available Mixed National Institute of Standard and Technology Database (MINST) and NYU Object Recognition Benchmark (NORB) dataset in supervisory and un-supervisory learning tasks to validate the hypothesis. The result of the paper shows that the traditional algorithms cannot resolve large size problems for sparse coding as the k-Sparse autoencoder model. Keywords—k-sparse autoencoder (KSA), Sparsity, algorithms, Sparse-coding

[1]  Pierre Baldi,et al.  Autoencoders, Unsupervised Learning, and Deep Architectures , 2011, ICML Unsupervised and Transfer Learning.

[2]  Yoshua Bengio,et al.  Why Does Unsupervised Pre-training Help Deep Learning? , 2010, AISTATS.

[3]  Josef Kittler,et al.  Intelligent Science and Intelligent Data Engineering , 2012, Lecture Notes in Computer Science.

[4]  Yong Peng,et al.  Marginalized Denoising Autoencoder via Graph Regularization for Domain Adaptation , 2013, ICONIP.

[5]  Carl E. Rasmussen,et al.  Gaussian processes for machine learning , 2005, Adaptive computation and machine learning.

[6]  Thomas Hofmann,et al.  Greedy Layer-Wise Training of Deep Networks , 2007 .

[7]  Andrew Y. Ng,et al.  The Importance of Encoding Versus Training with Sparse Coding and Vector Quantization , 2011, ICML.

[8]  David J. Field,et al.  Sparse coding with an overcomplete basis set: A strategy employed by V1? , 1997, Vision Research.

[9]  Jason Weston,et al.  Large-scale kernel machines , 2007 .

[10]  Geoffrey E. Hinton,et al.  3D Object Recognition with Deep Belief Nets , 2009, NIPS.

[11]  Ajith Abraham,et al.  Hybrid information systems , 2002 .

[12]  James H. McMillan,et al.  Research in Education: Evidence Based Inquiry , 2005 .

[13]  Michael I. Jordan,et al.  Advances in Neural Information Processing Systems 30 , 1995 .

[14]  Sergios Theodoridis,et al.  Signal processing theory and machine learning , 2014 .

[15]  Hugo Larochelle,et al.  Efficient Learning of Deep Boltzmann Machines , 2010, AISTATS.

[16]  Changyin Sun,et al.  Intelligent Science and Intelligent Data Engineering , 2011, Lecture Notes in Computer Science.

[17]  Michael Elad,et al.  K-SVD : DESIGN OF DICTIONARIES FOR SPARSE REPRESENTATION , 2005 .

[18]  Marc'Aurelio Ranzato,et al.  Fast Inference in Sparse Coding Algorithms with Applications to Object Recognition , 2010, ArXiv.

[19]  Yoshua Bengio,et al.  Scaling learning algorithms towards AI , 2007 .

[20]  Honglak Lee,et al.  An Analysis of Single-Layer Networks in Unsupervised Feature Learning , 2011, AISTATS.

[21]  Jyrki Kivinen Algorithmic learning theory : 22nd International Conference, ALT 2011, Espoo, Finland, October 5-7, 2011 : proceedings , 2011 .

[22]  Akira Hirose,et al.  Neural Information Processing , 2016, Lecture Notes in Computer Science.

[23]  Honglak Lee,et al.  Sparse deep belief net model for visual area V2 , 2007, NIPS.

[24]  Nahid Golafshani,et al.  Understanding Reliability and Validity in Qualitative Research , 2003 .

[25]  Joel A. Tropp,et al.  Signal Recovery From Random Measurements Via Orthogonal Matching Pursuit , 2007, IEEE Transactions on Information Theory.

[26]  Yann LeCun,et al.  Learning Fast Approximations of Sparse Coding , 2010, ICML.

[27]  Glyn Winter A Comparative Discussion of the Notion of 'Validity' in Qualitative and Quantitative Research , 2000 .

[28]  Yoshua Bengio,et al.  Deep Learning of Representations for Unsupervised and Transfer Learning , 2011, ICML Unsupervised and Transfer Learning.

[29]  Kjersti Engan,et al.  Method of optimal directions for frame design , 1999, 1999 IEEE International Conference on Acoustics, Speech, and Signal Processing. Proceedings. ICASSP99 (Cat. No.99CH36258).

[30]  Pascal Vincent,et al.  A Connection Between Score Matching and Denoising Autoencoders , 2011, Neural Computation.

[31]  Brendan J. Frey,et al.  k-Sparse Autoencoders , 2013, ICLR.

[32]  Cor J. Veenman,et al.  Kernel Codebooks for Scene Categorization , 2008, ECCV.

[33]  Simon Haykin,et al.  Correlative Learning: A Basis for Brain and Adaptive Systems (Adaptive and Learning Systems for Signal Processing, Communications and Control Series) , 2007 .

[34]  C. Villet,et al.  Introduction to Educational Research , 2011 .

[35]  Y. LeCun,et al.  Learning methods for generic object recognition with invariance to pose and lighting , 2004, Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2004. CVPR 2004..