On Not Making Dissimilarities Euclidean

Non-metric dissimilarity measures may arise in practice e.g. when objects represented by sensory measurements or by structural de- scriptions are compared. It is an open issue whether such non-metric measures should be corrected in some way to be metric or even Eu- clidean. The reason for such corrections is the fact that pairwise metric distances are interpreted in metric spaces, while Euclidean distances can be embedded into Euclidean spaces. Hence, traditional learning methods can be used. The k-nearest neighbor rule is usually applied to dissimilarities. In our earlier study (12, 13), we proposed some alternative approaches to general dissimilarity representations (DRs). They rely either on an embedding to a pseudo-Euclidean space and building classifiers there or on construct- ing classifiers on the representation directly. In this paper, we investigate ways of correcting DRs to make them more Euclidean (metric) either by adding a proper constant or by some concave transformations. Classi- fication experiments conducted on five dissimilarity data sets indicate that non-metric dissimilarity measures can be more beneficial than their corrected Euclidean or metric counterparts. The discriminating power of the measure itself is more important than its Euclidean (or metric) properties.

[1]  C. K. Chow,et al.  On optimum recognition error and reject tradeoff , 1970, IEEE Trans. Inf. Theory.

[2]  N. JARDINE,et al.  A New Approach to Pattern Recognition , 1971, Nature.

[3]  Gerald W. Kimble,et al.  Information and Computer Science , 1975 .

[4]  Azriel Rosenfeld,et al.  Progress in pattern recognition , 1985 .

[5]  J. Gower,et al.  Metric and Euclidean properties of dissimilarity coefficients , 1986 .

[6]  Anil K. Jain,et al.  A modified Hausdorff distance for object matching , 1994, Proceedings of 12th International Conference on Pattern Recognition.

[7]  Anil K. Jain,et al.  Representation and Recognition of Handwritten Digits Using Deformable Templates , 1997, IEEE Trans. Pattern Anal. Mach. Intell..

[8]  Catherine Blake,et al.  UCI Repository of machine learning databases , 1998 .

[9]  Daphna Weinshall,et al.  Classification with Nonmetric Distances: Image Retrieval and Class Representation , 2000, IEEE Trans. Pattern Anal. Mach. Intell..

[10]  Robert P. W. Duin,et al.  A Generalized Kernel Approach to Dissimilarity-based Classification , 2002, J. Mach. Learn. Res..

[11]  Horst Bunke,et al.  Towards Bridging the Gap between Statistical and Structural Pattern Recognition: Two New Concepts in Graph Matching , 2001, ICAPR.

[12]  Sameer Singh,et al.  Advances in Pattern Recognition — ICAPR 2001 , 2001, Lecture Notes in Computer Science.

[13]  Robert P. W. Duin,et al.  Dissimilarity representations allow for building good classifiers , 2002, Pattern Recognit. Lett..

[14]  Joachim M. Buhmann,et al.  Going Metric: Denoising Pairwise Data , 2002, NIPS.

[15]  Pierre Courrieu,et al.  Straight monotonic embedding of data sets in Euclidean spaces , 2002, Neural Networks.

[16]  Luisa Micó,et al.  A modification of the LAESA algorithm for approximated k-NN classification , 2003, Pattern Recognit. Lett..