This paper describes error-correction adjustment procedures for determining the weight vector of linear pattern classifiers under general pattern distribution. It is mainly aimed at clarifying theoretically the performance of adaptive pattern classifiers. In the case where the loss depends on the distance between a pattern vector and a decision boundary and where the average risk function is unimodal, it is proved that, by the procedures proposed here, the weight vector converges to the optimal one even under nonseparable pattern distributions. The speed and the accuracy of convergence are analyzed, and it is shown that there is an important tradeoff between speed and accuracy of convergence. Dynamical behaviors, when the probability distributions of patterns are changing, are also shown. The theory is generalized and made applicable to the case with general discriminant functions, including piecewise-linear discriminant functions.
[1]
W. Highleyman.
Linear Decision Functions, with Application to Pattern Recognition
,
1962,
Proceedings of the IRE.
[2]
A. Albert.
A Mathematical Theory of Pattern Recognition
,
1963
.
[3]
Thomas M. Cover,et al.
Geometrical and Statistical Properties of Systems of Linear Inequalities with Applications in Pattern Recognition
,
1965,
IEEE Trans. Electron. Comput..
[4]
Richard O. Duda,et al.
Pattern Classification by Iteratively Determined Linear and Piecewise Linear Discriminant Functions
,
1966,
IEEE Trans. Electron. Comput..
[5]
J. S. Koford,et al.
The use of an adaptive threshold element to design a linear optimal pattern classifier
,
1966,
IEEE Trans. Inf. Theory.