K-Means Clustering Optimizing Deep Stacked Sparse Autoencoder

Because of the large structure and long training time, the development cycle of the common depth model is prolonged. How to speed up training is a problem deserving of study. In order to accelerate training, K-means clustering optimizing deep stacked sparse autoencoder (K-means sparse SAE) is presented in this paper. First, the input features are divided into K small subsets by K-means clustering, then each subset is input into corresponding autoencoder model for training, which only has fewer nodes in the hidden layer than traditional models. After training, each autoencoder’s trained weights and biases is merged to obtain the next layer’s input features by feedforward network. The above steps are repeated till the softmax layer, then fine-tuning is carried out. Using MNIST-Rotation datasets to train the network that has three hidden layers and each layer has 800 nodes, the improved model has higher classification accuracy and shorter training time when K = 10. With K increasing, the training time is reduced to almost the same as the fine-tuning time but the recognition ability is descended. Compared with the recently stacked denoising sparse autoencoder, the recognition accuracy is improved by 1%, not only the noise factor is not selected but also the training speed is significantly increased. The trained filters from the improved model is also used to train convolutional autoencoder, and it performs better than traditional models. We find that pre-training stage doesn’t need large samples simultaneously, and small samples parallel training reduces the probability of falling into the local minimum.

[1]  Jochen J. Steil,et al.  Efficient online learning of a non-negative sparse autoencoder , 2010, ESANN.

[2]  Tetsuya Sakurai,et al.  Parallel Implementation of the Nonlinear Semi-NMF Based Alternating Optimization Method for Deep Neural Networks , 2017, Neural Processing Letters.

[3]  Yongsheng Ding,et al.  Thematic issue on “bio-inspired learning for data analysis” , 2017, Memetic Comput..

[4]  Hong Zhang,et al.  Facial expression recognition via learning deep sparse autoencoders , 2018, Neurocomputing.

[5]  Boris Vrdoljak,et al.  Statistical hierarchical clustering algorithm for outlier detection in evolving data streams , 2020, Mach. Learn..

[6]  B. Chandra,et al.  Adaptive Noise Schedule for Denoising Autoencoder , 2014, ICONIP.

[7]  Brendan J. Frey,et al.  k-Sparse Autoencoders , 2013, ICLR.

[8]  Nan Zhang,et al.  Research of stacked denoising sparse autoencoder , 2016, Neural Computing and Applications.

[9]  Marc'Aurelio Ranzato,et al.  Building high-level features using large scale unsupervised learning , 2011, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.

[10]  Pascal Vincent,et al.  Contractive Auto-Encoders: Explicit Invariance During Feature Extraction , 2011, ICML.

[11]  Angshul Majumdar,et al.  Majorization Minimization Technique for Optimally Solving Deep Dictionary Learning , 2017, Neural Processing Letters.

[12]  Wang Baoyi,et al.  Research on Personnel Information Collaborative Sensing Method of Intelligent Building Based on CPS , 2019 .

[13]  Haruhiko Nishimura,et al.  Evolving neural networks with iterative learning scheme for associative memory , 2005, Neural Processing Letters.

[14]  Yoshua Bengio,et al.  What regularized auto-encoders learn from the data-generating distribution , 2012, J. Mach. Learn. Res..

[15]  Juan-Manuel Torres-Moreno,et al.  Characterization of the Sonar Signals Benchmark , 1998, Neural Processing Letters.

[16]  Weihua Liu,et al.  The effect of whitening transformation on pooling operations in convolutional autoencoders , 2015, EURASIP J. Adv. Signal Process..

[17]  Angshul Majumdar,et al.  Imposing Class-Wise Feature Similarity in Stacked Autoencoders by Nuclear Norm Regularization , 2017, Neural Processing Letters.

[18]  Cong Hu,et al.  Structure Preserving Sparse Coding for Data Representation , 2018, Neural Processing Letters.

[19]  Rik Das,et al.  Partition selection with sparse autoencoders for content based image classification , 2017, Neural Computing and Applications.

[20]  Paul J. Kennedy,et al.  Relational autoencoder for feature extraction , 2017, 2017 International Joint Conference on Neural Networks (IJCNN).

[21]  Yoshua Bengio,et al.  Deep Sparse Rectifier Neural Networks , 2011, AISTATS.

[22]  Nathalie Japkowicz,et al.  Manifold-based synthetic oversampling with manifold conformance estimation , 2018, Machine Learning.

[23]  Zhiwen Yu,et al.  Three-dimensional image-based human pose recovery with hypergraph regularized autoencoders , 2017, Multimedia Tools and Applications.

[24]  Bo Li,et al.  First-Order Sensitivity Analysis for Hidden Neuron Selection in Layer-Wise Training of Networks , 2017, Neural Processing Letters.

[25]  Fuchun Sun,et al.  Denoising deep extreme learning machine for sparse representation , 2016, Memetic Computing.

[26]  Yee Whye Teh,et al.  A Fast Learning Algorithm for Deep Belief Nets , 2006, Neural Computation.