The authors consider the problem of learning a linear threshold function (a halfspace in n dimensions, also called a "perceptron"). Methods for solving this problem generally fall into two categories. In the absence of noise, this problem can be formulated as a linear program and solved in polynomial time with the ellipsoid algorithm (or interior point methods). On the other hand, simple greedy algorithms such as the perceptron algorithm seem to work well in practice and can be made noise tolerant; but, their running time depends on a separation parameter (which quantifies the amount of "wiggle room" available) and can be exponential in the description length of the input. They show how simple greedy methods can be used to find weak hypotheses (hypotheses that classify noticeably more than half of the examples) in polynomial time, without dependence on any separation parameter. This results in a polynomial-time algorithm for learning linear threshold functions in the PAC model in the presence of random classification noise. The algorithm is based on a new method for removing outliers in data. Specifically, for any set S of points in R/sup n/, each given to b bits of precision, they show that one can remove only a small fraction of S so that in the remaining set T, for every vector v, max/sub x/spl epsiv/T/(v/spl middot/x)/sup 2//spl les/poly(n,b)|T|/sup -1//spl Sigma//sub x/spl epsiv/T/(v/spl middot/x)/sup 2/. After removing these outliers, they are able to show that a modified version of the perceptron learning algorithm works in polynomial time, even in the presence of random classification noise.
[1]
Robert E. Schapire,et al.
The strength of weak learnability
,
1990,
Mach. Learn..
[2]
S. Agmon.
The Relaxation Method for Linear Inequalities
,
1954,
Canadian Journal of Mathematics.
[3]
Tom Bylander.
Polynomial learnability of linear threshold approximations
,
1993,
COLT '93.
[4]
Tom Bylander,et al.
Learning linear threshold functions in the presence of classification noise
,
1994,
COLT '94.
[5]
Yoav Freund,et al.
An improved boosting algorithm and its implications on learning complexity
,
1992,
COLT '92.
[6]
A. A. Mullin,et al.
Principles of neurodynamics
,
1962
.
[7]
Michael Kearns,et al.
Efficient noise-tolerant learning from statistical queries
,
1993,
STOC.