Spontaneous Clustering via Minimum Gamma-Divergence

We propose a new method for clustering based on local minimization of the gamma-divergence, which we call spontaneous clustering. The greatest advantage of the proposed method is that it automatically detects the number of clusters that adequately reflect the data structure. In contrast, existing methods, such as K-means, fuzzy c-means, or model-based clustering need to prescribe the number of clusters. We detect all the local minimum points of the gamma-divergence, by which we define the cluster centers. A necessary and sufficient condition for the gamma-divergence to have local minimum points is also derived in a simple setting. Applications to simulated and real data are presented to compare the proposed method with existing ones.

[1]  Jianhong Wu,et al.  Projective Clustering Using Neural Networks with Adaptive Delay and Signal Transmission Loss , 2011, Neural Computation.

[2]  G. Nickless,et al.  THE ANALYSIS OF ROMANO-BRITISH POTTERY BY ATOMIC ABSORPTION SPECTROPHOTOMETRY , 1980 .

[3]  Han-Ming Wu,et al.  On biological validity indices for soft clustering algorithms for gene expression data , 2011, Comput. Stat. Data Anal..

[4]  Ting-Li Chen,et al.  $\gamma$-SUP: A clustering algorithm for cryo-electron microscopy images of asymmetric particles , 2012, 1205.2034.

[5]  Anil K. Jain,et al.  Data clustering: a review , 1999, CSUR.

[6]  Shogo Kato,et al.  Entropy and Divergence Associated with Power Function and the Statistical Application , 2010, Entropy.

[7]  Radford M. Neal Pattern Recognition and Machine Learning , 2007, Technometrics.

[8]  Yizong Cheng,et al.  Mean Shift, Mode Seeking, and Clustering , 1995, IEEE Trans. Pattern Anal. Mach. Intell..

[9]  Christopher M. Bishop,et al.  Pattern Recognition and Machine Learning (Information Science and Statistics) , 2006 .

[10]  Le Thi Hoai An,et al.  The DC (Difference of Convex Functions) Programming and DCA Revisited with DC Models of Real World Nonconvex Optimization Problems , 2005, Ann. Oper. Res..

[11]  Alan L. Yuille,et al.  The Concave-Convex Procedure , 2003, Neural Computation.

[12]  Jochen Einbeck,et al.  Bandwidth Selection for Mean-shift based Unsupervised Learning Techniques: a Unified Approach via Self-coverage , 2011 .

[13]  Rui Xu,et al.  Survey of clustering algorithms , 2005, IEEE Transactions on Neural Networks.

[14]  S. Eguchi,et al.  Robust parameter estimation with a small bias against heavy contamination , 2008 .

[15]  Shinto Eguchi,et al.  Duality of Maximum Entropy and Minimum Divergence , 2014, Entropy.

[16]  P. Deb Finite Mixture Models , 2008 .

[17]  Su-Yun Huang,et al.  On the weak convergence and Central Limit Theorem of blurring and nonblurring processes with application to robust location estimation , 2014, J. Multivar. Anal..

[18]  Shinto Eguchi,et al.  Robust Extraction of Local Structures by the Minimum β-Divergence Method , 2007 .

[19]  T. Caliński,et al.  A dendrite method for cluster analysis , 1974 .

[20]  Robert Tibshirani,et al.  The Elements of Statistical Learning: Data Mining, Inference, and Prediction, 2nd Edition , 2001, Springer Series in Statistics.

[21]  Bin Li,et al.  A New Clustering Approach on the Basis of Dynamical Neural Field , 2011, Neural Computation.

[22]  De Helguero,et al.  SUI MASSIMI DELLE CURVE DIMORFICHE , 1904 .

[23]  P. J. Green,et al.  Density Estimation for Statistics and Data Analysis , 1987 .

[24]  Robert Tibshirani,et al.  Estimating the number of clusters in a data set via the gap statistic , 2000 .

[25]  Shinto Eguchi,et al.  Detection of Heterogeneous Structures on the Gaussian Copula Model Using Projective Power Entropy , 2013 .