With the boom of web and social network, the amount of generated text data has increased enormously. On one hand, although text clustering methods are applicable to classify text data and facilitate data mining work such as information retrieval and recommendation, inadequate aspects are still evident. Especially, most existing text clustering methods provide either a hard partitioned or a hierarchical result, which cannot describe the data from various perspectives. On the other hand, multiple clustering approaches, which are proposed to classify data with various perspectives, meet several challenges such as high time complexity and incomprehensible results while applied to text documents. In this paper, we propose a frequent term-based multiple clustering approach for text documents. Our approach classifies text documents with various perspectives and provides a semantic explanation for each cluster. Through a series of experiments, we prove that our method is more scalable and provides more comprehensible results than traditional multiple clustering methods such as OSCLU and ASCLU while applied to text documents. In addition, we also found that our approach achieves a better clustering quality than existing text clustering approaches like FTC.
[1]
George A. Miller,et al.
WordNet: A Lexical Database for English
,
1995,
HLT.
[2]
D.M. Mount,et al.
An Efficient k-Means Clustering Algorithm: Analysis and Implementation
,
2002,
IEEE Trans. Pattern Anal. Mach. Intell..
[3]
Ira Assent,et al.
Evaluating Clustering in Subspace Projections of High Dimensional Data
,
2009,
Proc. VLDB Endow..
[4]
Ramakrishnan Srikant,et al.
Fast algorithms for mining association rules
,
1998,
VLDB 1998.
[5]
Emmanuel Müller,et al.
Detection of orthogonal concepts in subspaces of high dimensional data
,
2009,
CIKM.
[6]
David W. Conrath,et al.
Semantic Similarity Based on Corpus Statistics and Lexical Taxonomy
,
1997,
ROCLING/IJCLCLP.
[7]
Ramakrishnan Srikant,et al.
Fast Algorithms for Mining Association Rules in Large Databases
,
1994,
VLDB.