K-Mean Algorithm with a Distance Based on the Characteristic of Differences

A non-metric distance measure for similarity estimation based on the characteristic of differences is presented. This kind of distance is implemented in the well-known k-means clustering algorithm. To demonstrate the effectiveness of the distance we proposed, the performance of this kind of distance and the Euclidean and Manhattan distances were compared by clustering Iris dataset from the UCI repository. Experiment results show that the new distance measure can provide a more accurate feature model than the classical Euclidean and Manhattan distances.

[1]  R. Fisher THE USE OF MULTIPLE MEASUREMENTS IN TAXONOMIC PROBLEMS , 1936 .

[2]  Moshe Zakai General error criteria (Corresp.) , 1964, IEEE Trans. Inf. Theory.

[3]  G. Gates,et al.  The reduced nearest neighbor rule (Corresp.) , 1972, IEEE Trans. Inf. Theory.

[4]  Belur V. Dasarathy,et al.  Nosing Around the Neighborhood: A New System Structure and Classification Rule for Recognition in Partially Exposed Environments , 1980, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[5]  Anil K. Jain,et al.  Algorithms for Clustering Data , 1988 .

[6]  G. Gates The Reduced Nearest Neighbor Rule , 1998 .

[7]  Chien-Hsing Chou,et al.  Short Papers , 2001 .

[8]  Ernst Bonek,et al.  Improving clustering performance using multipath component distance , 2006 .

[9]  A.S. Ali,et al.  Unique distance measure approach for K-means (UDMA-Km) clustering algorithm , 2007, TENCON 2007 - 2007 IEEE Region 10 Conference.

[10]  Nicu Sebe,et al.  Distance Learning for Similarity Estimation , 2008, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[11]  Vincent Kanade,et al.  Clustering Algorithms , 2021, Wireless RF Energy Transfer in the Massive IoT Era.