The 2-NN Rule for More Accurate NN Risk Estimation

By proper design of a nearest-neighbor (NN) rule, it is possible to reduce effects of sample size in NN risk estimation. The 2-NN rule for the two-class problem eliminates the first-order effects of sample size. Since its asymptotic value is exactly half that of the 1-NN rule, it is possible to substitute the 2-NN rule for the 1-NN rule with a resultant increase in accuracy. For further stabilization of the risk estimate with respect to sample size, 2-NN polarization is suggested. Examples are included. The 2-NN approach is extended to M-class and 2k-NN.

[1]  Keinosuke Fukunaga,et al.  The optimal distance measure for nearest neighbor classification , 1981, IEEE Trans. Inf. Theory.

[2]  Peter E. Hart,et al.  Nearest neighbor pattern classification , 1967, IEEE Trans. Inf. Theory.

[3]  Josef Kittler,et al.  Pattern recognition : a statistical approach , 1982 .

[4]  Keinosuke Fukunaga,et al.  An Optimal Global Nearest Neighbor Metric , 1984, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[5]  Martin E. Hellman,et al.  The Nearest Neighbor Classification Rule with a Reject Option , 1970, IEEE Trans. Syst. Sci. Cybern..