Nonlinear Discriminative Dimensionality Reduction of Multiple Datasets

Dimensionality reduction (DR) is critical to many machine learning and signal processing tasks involving high-dimensional large-scale data. Standard DR tools such as principal component analysis (PCA) deal with a single dataset at a time. In diverse practical settings however, one is often tasked with learning the discriminant subspace such that one dataset of particular interest (a.k.a., target data) lies on, whereas the other dataset(s) (a.k.a., control data) do not. This is what is known as discriminative DR. Building on but considerably generalizing existing linear variants, this contribution puts forth a novel nonlinear approach for discriminative DR of multiple datasets through kernel-based learning. Interestingly, its solution can be provided analytically in terms of a generalized eigenvalue decomposition problem, for which various efficient solvers are available. Numerical experiments using synthetic and real data showcase the merits of the proposed nonlinear discriminative DR approach relative to state-of-the-art alternatives.

[1]  Jia Chen,et al.  Adaptive regularized canonical correlations in clustering sensor data , 2014, 2014 48th Asilomar Conference on Signals, Systems and Computers.

[2]  Gang Wang,et al.  Randomized Block Frank–Wolfe for Convergent Large-Scale Learning , 2016, IEEE Transactions on Signal Processing.

[3]  Karl Pearson F.R.S. LIII. On lines and planes of closest fit to systems of points in space , 1901 .

[4]  R. Fisher THE USE OF MULTIPLE MEASUREMENTS IN TAXONOMIC PROBLEMS , 1936 .

[5]  Michael I. Jordan,et al.  Multiple kernel learning, conic duality, and the SMO algorithm , 2004, ICML.

[6]  Gang Wang,et al.  Graph Multiview Canonical Correlation Analysis , 2018, IEEE Transactions on Signal Processing.

[7]  Gang Wang,et al.  Dpca: Dimensionality Reduction for Discriminative Analytics of Multiple Large-Scale Datasets , 2017, 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[8]  Gang Wang,et al.  Nonlinear Dimensionality Reduction for Discriminative Analytics of Multiple Datasets , 2018, IEEE Transactions on Signal Processing.

[9]  N. Aronszajn Theory of Reproducing Kernels. , 1950 .

[10]  S T Roweis,et al.  Nonlinear dimensionality reduction by locally linear embedding. , 2000, Science.

[11]  Gang Wang,et al.  Going beyond linear dependencies to unveil connectivity of meshed grids , 2017, 2017 IEEE 7th International Workshop on Computational Advances in Multi-Sensor Adaptive Processing (CAMSAP).

[12]  Ignacio Rojas,et al.  Design, implementation and validation of a novel open framework for agile development of mobile health applications , 2015, BioMedical Engineering OnLine.

[13]  Gang Wang,et al.  Canonical Correlation Analysis of Datasets With a Common Source Graph , 2018, IEEE Transactions on Signal Processing.

[14]  H. Hotelling Analysis of a complex of statistical variables into principal components. , 1933 .

[15]  Grace X. Y. Zheng,et al.  Massively parallel digital transcriptional profiling of single cells , 2016, Nature Communications.

[16]  Nello Cristianini,et al.  Kernel Methods for Pattern Analysis , 2003, ICTAI.

[17]  Vivek Kumar Bagaria,et al.  Contrastive Principal Component Analysis , 2017, ArXiv.

[18]  Bernhard Schölkopf,et al.  Kernel Principal Component Analysis , 1997, ICANN.

[19]  Georgios B. Giannakis,et al.  Topology Identification and Learning over Graphs: Accounting for Nonlinearities and Dynamics , 2018, Proceedings of the IEEE.