Toward an optimal analysis of hyperspectral data

In hyperspectral data materials of practical interest usually exist in a number of states and are observed in a number of conditions of illumination. It is thus necessary to characterize them not with a single spectral response but with a family of responses. One of the most versatile means for representing such a family of responses quantitatively is to model each by a normal distribution, as this makes possible classification by assigning the class to a sample based on Bayes rule. This leads to competitive performance only under special circumstances. The purpose of this dissertation is to improve the quantitative definitions of classes by replacing the classical normal model with more flexible and powerful alternatives and thereby investigate the relationship between class definition precision and classification accuracy. One other factor that directly affects the precision of class definitions is the number of labeled samples available for training. Characterizing class data with a limited set of labeled samples may have severe consequences. In a typical setting a remote sensing analyst should either sacrifice from the classifier performance by confining himself to the already available labeled data set or commit more time and effort to acquire more labeled samples both of which comes at a price. The present study also addresses this problem by proposing a semi-supervised binary classifier which seeks to incorporate unlabeled data into the training in order to improve the quantitative definitions of classes and hence improve the classifier performance at no extra cost.

[1]  Philip H. Swain,et al.  Remote Sensing: The Quantitative Approach , 1981, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[2]  David A. Landgrebe,et al.  Covariance estimation with limited training samples , 1999, IEEE Trans. Geosci. Remote. Sens..

[3]  Bor-Chen Kuo,et al.  A covariance estimator for small sample size classification problems and its application to feature extraction , 2002, IEEE Trans. Geosci. Remote. Sens..

[4]  David A. Landgrebe On the relationship between class definition precision and classification accuracy in hyperspectral analysis , 2000, IGARSS 2000. IEEE 2000 International Geoscience and Remote Sensing Symposium. Taking the Pulse of the Planet: The Role of Remote Sensing in Managing the Environment. Proceedings (Cat. No.00CH37120).

[5]  Bor-Chen Kuo,et al.  Hyperspectral data classification using nonparametric weighted feature extraction , 2002, IEEE International Geoscience and Remote Sensing Symposium.

[6]  R. Tibshirani,et al.  Discriminant Analysis by Gaussian Mixtures , 1996 .

[7]  David A. Landgrebe,et al.  The effect of unlabeled samples in reducing the small sample size problem and mitigating the Hughes phenomenon , 1994, IEEE Trans. Geosci. Remote. Sens..

[8]  J. Modified Kernel-based Nonlinear Feature Extraction , 2004 .

[9]  Vladimir Vapnik,et al.  Statistical learning theory , 1998 .

[10]  G. Wahba Support vector machines, reproducing kernel Hilbert spaces, and randomized GACV , 1999 .

[11]  Vittorio Castelli,et al.  The relative value of labeled and unlabeled samples in pattern recognition with an unknown mixing parameter , 1996, IEEE Trans. Inf. Theory.

[12]  G. F. Hughes,et al.  On the mean accuracy of statistical pattern recognizers , 1968, IEEE Trans. Inf. Theory.

[13]  Thorsten Joachims,et al.  Transductive Inference for Text Classification using Support Vector Machines , 1999, ICML.

[14]  Adrian E. Raftery,et al.  Model-Based Clustering, Discriminant Analysis, and Density Estimation , 2002 .

[15]  Robert W. Basedow,et al.  HYDICE: an airborne system for hyperspectral imaging , 1993, Defense, Security, and Sensing.

[16]  Tong Zhang,et al.  The Value of Unlabeled Data for Classification Problems , 2000, ICML 2000.

[17]  D. Landgrebe,et al.  A method for estimating the number of components in a normal mixture density function , 2000, IGARSS 2000. IEEE 2000 International Geoscience and Remote Sensing Symposium. Taking the Pulse of the Planet: The Role of Remote Sensing in Managing the Environment. Proceedings (Cat. No.00CH37120).

[18]  G. McLachlan,et al.  The EM algorithm and extensions , 1996 .

[19]  Qiong Jackson,et al.  An adaptive classifier design for high-dimensional data analysis with a limited training data set , 2001, IEEE Trans. Geosci. Remote. Sens..

[20]  Sebastian Thrun,et al.  Text Classification from Labeled and Unlabeled Documents using EM , 2000, Machine Learning.

[21]  Ayhan Demiriz,et al.  Semi-Supervised Support Vector Machines , 1998, NIPS.

[22]  Adrian E. Raftery,et al.  How Many Clusters? Which Clustering Method? Answers Via Model-Based Cluster Analysis , 1998, Comput. J..

[23]  B. Scholkopf,et al.  Fisher discriminant analysis with kernels , 1999, Neural Networks for Signal Processing IX: Proceedings of the 1999 IEEE Signal Processing Society Workshop (Cat. No.98TH8468).

[24]  Nello Cristianini,et al.  Further results on the margin distribution , 1999, COLT '99.

[25]  Anthony Widjaja,et al.  Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond , 2003, IEEE Transactions on Neural Networks.

[26]  Gunnar Rätsch,et al.  Input space versus feature space in kernel-based methods , 1999, IEEE Trans. Neural Networks.

[27]  O. Nelles,et al.  An Introduction to Optimization , 1996, IEEE Antennas and Propagation Magazine.

[28]  Vittorio Castelli,et al.  On the exponential value of labeled samples , 1995, Pattern Recognit. Lett..

[29]  Michael E. Tipping Sparse Kernel Principal Component Analysis , 2000, NIPS.

[30]  O. Mangasarian,et al.  Semi-Supervised Support Vector Machines for Unlabeled Data Classification , 2001 .

[31]  Robert F. Cromp,et al.  Support Vector Machine Classifiers as Applied to AVIRIS Data , 1999 .

[32]  David A. Landgrebe,et al.  Signal Theory Methods in Multispectral Remote Sensing , 2003 .

[33]  Fabio Gagliardi Cozman,et al.  Unlabeled Data Can Degrade Classification Performance of Generative Classifiers , 2002, FLAIRS.

[34]  D. Rubin,et al.  Maximum likelihood from incomplete data via the EM - algorithm plus discussions on the paper , 1977 .

[35]  Keinosuke Fukunaga,et al.  A Branch and Bound Clustering Algorithm , 1975, IEEE Transactions on Computers.

[36]  Bernhard Schölkopf,et al.  Estimating a Kernel Fisher Discriminant in the Presence of Label Noise , 2001, ICML.

[37]  J. Anthony Gualtieri,et al.  Support vector machines for hyperspectral remote sensing classification , 1999, Other Conferences.

[38]  David A. Landgrebe,et al.  Covariance Matrix Estimation and Classification With Limited Training Data , 1996, IEEE Trans. Pattern Anal. Mach. Intell..

[39]  Federico Girosi,et al.  An improved training algorithm for support vector machines , 1997, Neural Networks for Signal Processing VII. Proceedings of the 1997 IEEE Signal Processing Society Workshop.

[40]  Avrim Blum,et al.  The Bottleneck , 2021, Monopsony Capitalism.