A New Cross Clustering Algorithm for Improving Performance of Supervised Learning

In this paper, a new clustering algorithm is proposed based on cross clusters without using membership functions. In light of the cross clustering data transformation, the spatial distribution of data is changed while the original data dimension simultaneously is maintained. Combining with the performance index and visual technology, an explanation of the performance improvement of the classification model is presented in accordance with the proposed algorithm. This approach was evaluated on UCR time series datasets, the experiments showed that the algorithm can improve not only the accuracy and the performance of the fully convolutional network and nearest neighbor algorithm, but also the time complexity in time series classification model. It is worth well to apply this method to further research and popularization.

[1]  Geoffrey I. Webb,et al.  Dynamic Time Warping Averaging of Time Series Allows Faster and More Accurate Classification , 2014, 2014 IEEE International Conference on Data Mining.

[2]  Delbert Dueck,et al.  Clustering by Passing Messages Between Data Points , 2007, Science.

[3]  Ying Wah Teh,et al.  Time-series clustering - A decade review , 2015, Inf. Syst..

[4]  Eamonn J. Keogh,et al.  Exact indexing of dynamic time warping , 2002, Knowledge and Information Systems.

[5]  Patrick Schäfer,et al.  Scalable time series classification , 2016, Data Mining and Knowledge Discovery.

[6]  Sungzoon Cho,et al.  Clustering-Based Reference Set Reduction for k-Nearest Neighbor , 2007, ISNN.

[7]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[8]  Li Wei,et al.  Experiencing SAX: a novel symbolic representation of time series , 2007, Data Mining and Knowledge Discovery.

[9]  Samina Kausar,et al.  Integration of Data Mining Clustering Approach in the Personalized E-Learning System , 2018, IEEE Access.

[10]  Germain Forestier,et al.  Transfer learning for time series classification , 2018, 2018 IEEE International Conference on Big Data (Big Data).

[11]  Jason Lines,et al.  Time-Series Classification with COTE: The Collective of Transformation-Based Ensembles , 2015, IEEE Transactions on Knowledge and Data Engineering.

[12]  Jason Lines,et al.  Time series classification with ensembles of elastic distance measures , 2015, Data Mining and Knowledge Discovery.

[13]  Chih-Ping Wei,et al.  Nearest-neighbor-based approach to time-series classification , 2012, Decis. Support Syst..

[14]  Heyuan Shi,et al.  Lithium-Ion Cell Screening With Convolutional Neural Networks Based on Two-Step Time-Series Clustering and Hybrid Resampling for Imbalanced Data , 2018, IEEE Access.

[15]  Han Xin-jie The Fuzzy C-Means Clustering Algorithm and Its Application in the Fault Diagnosis of Ships , 2007 .

[16]  J. Tenenbaum,et al.  A global geometric framework for nonlinear dimensionality reduction. , 2000, Science.

[17]  Philip S. Yu,et al.  Top 10 algorithms in data mining , 2007, Knowledge and Information Systems.

[18]  J. Bezdek,et al.  FCM: The fuzzy c-means clustering algorithm , 1984 .

[19]  Samee Ullah Khan,et al.  MapReduce-based fast fuzzy c-means algorithm for large-scale underwater image segmentation , 2016, Future Gener. Comput. Syst..

[20]  George C. Runger,et al.  A Bag-of-Features Framework to Classify Time Series , 2013, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[21]  Patrick Schäfer The BOSS is concerned with time series classification in the presence of noise , 2014, Data Mining and Knowledge Discovery.

[22]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[23]  Kuangrong Hao,et al.  An integrated algorithm for multi-agent fault-tolerant scheduling based on MOEA , 2019, Future Gener. Comput. Syst..

[24]  Xinyu Luo,et al.  Cost-Sensitive Convolution based Neural Networks for Imbalanced Time-Series Classification , 2018, ArXiv.

[25]  Tim Oates,et al.  Time series classification from scratch with deep neural networks: A strong baseline , 2016, 2017 International Joint Conference on Neural Networks (IJCNN).

[26]  Sean Hughes,et al.  Clustering by Fast Search and Find of Density Peaks , 2016 .

[27]  Anil K. Jain Data clustering: 50 years beyond K-means , 2008, Pattern Recognit. Lett..

[28]  Simon Fong,et al.  Clustering big IoT data by metaheuristic optimized mini-batch and parallel partition-based DGC in Hadoop , 2018, Future Gener. Comput. Syst..

[29]  Sergey Ioffe,et al.  Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.

[30]  Wei Lu,et al.  Dynamic Background Subtraction Using Histograms Based on Fuzzy C-Means Clustering and Fuzzy Nearness Degree , 2019, IEEE Access.