Due to the development of internet, plentiful different data appear rapidly. The amounts of features also increase when the technique for collecting data becomes mature. Observation of different data is usually not an easy task because it needs some background related to data pre-processing. Therefore, dimensionality reduction (DR) becomes a familiar method to reduce the amount of features and keep the critical information. However, the loss of information during the processing of dimensionality reduction is unavoidable. When the targeted dimension is far lower than original dimension, the loss is too high to be endurable. To solve this problem, we use the encoder structure from autoencoder to compare to some common dimensionality reduction methods. We use the simplest autoencoder structure as the preprocessing of Support Vector Machine (SVM) to see the result.
[1]
Retantyo Wardoyo,et al.
Time Complexity Analysis of Support Vector Machines (SVM) in LibSVM
,
2015
.
[2]
Eric O. Postma,et al.
Dimensionality Reduction: A Comparative Review
,
2008
.
[3]
H. Hotelling.
Analysis of a complex of statistical variables into principal components.
,
1933
.
[4]
S T Roweis,et al.
Nonlinear dimensionality reduction by locally linear embedding.
,
2000,
Science.
[5]
H. Bourlard,et al.
Auto-association by multilayer perceptrons and singular value decomposition
,
1988,
Biological Cybernetics.