Complexity of Magnetic Resonance Spectrum Classification

We use several data complexity measures to explain the differences in classification accuracy using various sets of features selected from samples of magnetic resonance spectra for two-class discrimination. Results suggest that for this typical problem with sparse samples in a high-dimensional space, even robust classifiers like random decision forests can benefit from sophisticated feature selection procedures, and the improvement can be explained by the more favorable characteristics in the class geometry given by the resultant feature sets.

[1]  C E Mountford,et al.  Diagnosis and prognosis of breast cancer by magnetic resonance spectroscopy of fine‐needle aspirates analysed using a statistical classification strategy , 2001, The British journal of surgery.

[2]  Robert P. W. Duin,et al.  Bagging, Boosting and the Random Subspace Method for Linear Classifiers , 2002, Pattern Analysis & Applications.

[3]  R L Somorjai,et al.  Near‐optimal region selection for feature space reduction: novel preprocessing methods for classifying MR spectra , 1998, NMR in biomedicine.

[4]  R. Somorjai,et al.  Distinguishing normal from rejecting renal allografts: application of a three—stage classification strategy to MR and IR spectra of urine , 2002 .

[5]  Richard Baumgartner,et al.  Class prediction and discovery using gene microarray and proteomics mass spectroscopy data: curses, caveats, cautions , 2003, Bioinform..

[6]  R. Somorjai,et al.  Rapid Identification of Candida Species by Using Nuclear Magnetic Resonance Spectroscopy and a Statistical Classification Strategy , 2003, Applied and Environmental Microbiology.

[7]  Tin Kam Ho,et al.  The Random Subspace Method for Constructing Decision Forests , 1998, IEEE Trans. Pattern Anal. Mach. Intell..

[8]  Giorgio Valentini,et al.  Bio-molecular cancer prediction with random subspace ensembles of support vector machines , 2005, Neurocomputing.

[9]  Ray L. Somorjai,et al.  Accurate diagnosis and prognosis of human cancers by proton MRS and a three-stage classification strategy , 2002 .

[10]  E. M. Kleinberg,et al.  Stochastic discrimination , 1990, Annals of Mathematics and Artificial Intelligence.

[11]  Anil K. Jain,et al.  Small Sample Size Effects in Statistical Pattern Recognition: Recommendations for Practitioners , 1991, IEEE Trans. Pattern Anal. Mach. Intell..

[12]  Tin Kam Ho,et al.  Complexity Measures of Supervised Classification Problems , 2002, IEEE Trans. Pattern Anal. Mach. Intell..

[13]  Horst Bunke,et al.  Feature selection algorithms for the generation of multiple classifier systems and their application to handwritten word recognition , 2004 .

[14]  Tin Kam Ho,et al.  Nearest Neighbors in Random Subspaces , 1998, SSPR/SPR.