If f(x) and g(x) are the densities for the N-dimensional measurement vector x, conditioned on the classes c1 and c2, and if finite sets of samples from the two classes are available, then a decision function based on estimates f(x) and ¿(x) can be used to classify future observations. In general, however, when the measurement complexity (the dimensionality N) is increased arbitrarily and the sets of training samples remain finite, a ``peaking phenomenon'' of the following kind is observed: classification accuracy improves at first, peaks at a finite value of N, called the optimum measurement complexity, and starts deteriorating thereafter. We derive, for the case of statistically independent measurements, general conditions under which it can be guaranteed that the peaking phenomenon will not occur, and the correct classification probability will keep increasing to value unity as N ¿ ¿. Several applications are considered which together indicate, contrary to general belief, that independence of measurements alone does not guarantee the absence of the peaking phenomenon.
[1]
S. Silvey.
A Note on Maximum‐Likelihood in the Case of Dependent Random Variables
,
1961
.
[2]
G. F. Hughes,et al.
On the mean accuracy of statistical pattern recognizers
,
1968,
IEEE Trans. Inf. Theory.
[3]
B. Chandrasekaran,et al.
Comments on "On the mean accuracy of statistical pattern recognizers" by Hughes, G. F
,
1969,
IEEE Trans. Inf. Theory.
[4]
B. Chandrasekaran,et al.
On dimensionality and sample size in statistical pattern classification
,
1971,
Pattern Recognit..
[5]
B. Chandrasekaran,et al.
Independence of measurements and the mean recognition accuracy
,
1971,
IEEE Trans. Inf. Theory.
[6]
B. Chandrasekaran,et al.
Quantization Complexity and Independent Measurements
,
1974,
IEEE Transactions on Computers.