Minimum Risk Feature Transformations

We develop an approach for automatically learning the optimal feature transformation for a given classification problem. The approach is based on extending the principle of risk minimization (RM), commonly used for learning classifiers, to learning feature transformations that admit classifiers with minimum risk. This allows feature extraction and classification to proceed in a unified framework, both guided by the RM principle. The framework is applied to derive new algorithms for learning feature transformations. Our experiments demonstrate the ability of the resulting algorithms to learn good features for a variety of classification problems.

[1]  Loo-Nin Teow,et al.  Handwritten digit recognition with a novel vision model that extracts linearly separable features , 2000, Proceedings IEEE Conference on Computer Vision and Pattern Recognition. CVPR 2000 (Cat. No.PR00662).

[2]  Jason Weston,et al.  Vicinal Risk Minimization , 2000, NIPS.

[3]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.

[4]  Sayan Mukherjee,et al.  Feature Selection for SVMs , 2000, NIPS.

[5]  Sayan Mukherjee,et al.  Choosing Multiple Parameters for Support Vector Machines , 2002, Machine Learning.

[6]  Vladimir N. Vapnik,et al.  The Nature of Statistical Learning Theory , 2000, Statistics for Engineering and Information Science.

[7]  Vladimir Vapnik,et al.  Statistical learning theory , 1998 .

[8]  Jonathan Baxter,et al.  A Model of Inductive Bias Learning , 2000, J. Artif. Intell. Res..

[9]  Shaul Markovitch,et al.  Feature Generation Using General Constructor Functions , 2002, Machine Learning.