Pattern Classification via Single Spheres

Previous sphere-based classification algorithms usually need a number of spheres in order to achieve good classification performance. In this paper, inspired by the support vector machines for classification and the support vector data description method, we present a new method for constructing single spheres that separate data with the maximum separation ratio. In contrast to previous methods that construct spheres in the input space, the new method constructs separating spheres in the feature space induced by the kernel. As a consequence, the new method is able to construct a single sphere in the feature space to separate patterns that would otherwise be inseparable when using a sphere in the input space. In addition, by adjusting the ratio of the radius of the sphere to the separation margin, it can provide a series of solutions ranging from spherical to linear decision boundaries, effectively encompassing both the support vector machines for classification and the support vector data description method. Experimental results show that the new method performs well on both artificial and real-world datasets.

[1]  Federico Girosi,et al.  Support Vector Machines: Training and Applications , 1997 .

[2]  Bernhard E. Boser,et al.  A training algorithm for optimal margin classifiers , 1992, COLT '92.

[3]  Paul W. Cooper,et al.  The Hypersphere in Pattern Recognition , 1962, Inf. Control..

[4]  Bernhard Schölkopf,et al.  Extracting Support Data for a Given Task , 1995, KDD.

[5]  D. L. Reilly,et al.  A neural model for category learning , 1982, Biological Cybernetics.

[6]  Vladimir Vapnik,et al.  Statistical learning theory , 1998 .

[7]  Vladimir Vapnik,et al.  Estimation of Dependences Based on Empirical Data: Springer Series in Statistics (Springer Series in Statistics) , 1982 .

[8]  Corinna Cortes,et al.  Support-Vector Networks , 1995, Machine Learning.

[9]  Catherine Blake,et al.  UCI Repository of machine learning databases , 1998 .

[10]  V. Vapnik Estimation of Dependences Based on Empirical Data , 2006 .

[11]  John C. Platt,et al.  Fast training of support vector machines using sequential minimal optimization, advances in kernel methods , 1999 .

[12]  Bernhard Schölkopf,et al.  Estimating the Support of a High-Dimensional Distribution , 2001, Neural Computation.

[13]  Paul W. Cooper,et al.  A Note on an Adaptive Hypersphere Decision Boundary , 1966, IEEE Trans. Electron. Comput..

[14]  Dana Z. Anderson,et al.  Neural Information Processing Systems , 1988 .

[16]  F. Glineur Pattern separation via ellipsoids and conic programming , 1998 .

[17]  Robert P. W. Duin,et al.  Support vector domain description , 1999, Pattern Recognit. Lett..

[18]  John Shawe-Taylor,et al.  The Set Covering Machine , 2003, J. Mach. Learn. Res..

[19]  Leon N. Cooper,et al.  Pattern Class Degeneracy in an Unrestricted Storage Density Memory , 1987, NIPS.

[20]  Bruce G. Batchelor,et al.  Practical approach to pattern classification , 1974 .