Machine Learning Classifiers Based Classification For IRIS Recognition

Classification is the most widely applied machine learning problem today, with implementations in face recognition, flower classification, clustering, and other fields. The goal of this paper is to organize and identify a set of data objects. The study employs K-nearest neighbors, decision tree (j48), and random forest algorithms, and then compares their performance using the IRIS dataset. The results of the comparison analysis showed that the K-nearest neighbors outperformed the other classifiers. Also, the random forest classifier worked better than the decision tree (j48). Finally, the best result obtained by this study is 100% and there is no error rate for the classifier that was obtained. Keywords—Data Mining, Classification, Decision Tree, Random Forest, K-nearest neighbors

[1]  Gaurav Gupta,et al.  A self explanatory review of decision tree classifiers , 2014, International Conference on Recent Advances and Innovations in Engineering (ICRAIE-2014).

[2]  B. Pradhan,et al.  Landslide Susceptibility Assessment in Vietnam Using Support Vector Machines, Decision Tree, and Naïve Bayes Models , 2012 .

[3]  Muhammd Jawad Hamid Mughal Data Mining: Web Data Mining Techniques, Tools and Algorithms: An Overview , 2018 .

[4]  Adnan Mohsin Abdulazeez,et al.  The Role of Machine Learning Algorithms for Diagnosing Diseases , 2021 .

[5]  Liliya Demidova,et al.  Defining the Ranges Boundaries of the Optimal Parameters Values for the Random Forest Classifier , 2019, 2019 1st International Conference on Control Systems, Mathematical Modelling, Automation and Energy Efficiency (SUMMA).

[6]  Diyar Qader Zeebaree,et al.  Robust watermarking scheme based LWT and SVD using artificial bee colony optimization , 2021 .

[7]  Habibollah Haron,et al.  Gene Selection and Classification of Microarray Data Using Convolutional Neural Network , 2018, 2018 International Conference on Advanced Science and Engineering (ICOASE).

[8]  Roseline Oluwaseun Ogundokun,et al.  Diagnosis of Long Sightedness Using Neural Network and Decision Tree Algorithms , 2021, Journal of Physics: Conference Series.

[9]  Marek Kretowski,et al.  Decision tree underfitting in mining of gene expression data. An evolutionary multi-test tree approach , 2019, Expert Syst. Appl..

[10]  Maad M. Mijwil,et al.  Utilizing the Genetic Algorithm to Pruning the C4.5 Decision Tree Algorithm , 2021, Asian Journal of Applied Sciences.

[11]  Hen-Geul Yeh,et al.  Feature Extraction, Selection, and K-Nearest Neighbors Algorithm for Shark Behavior Classification Based on Imbalanced Dataset , 2021, IEEE Sensors Journal.

[12]  Subhan,et al.  Facial recognition using two-dimensional principal component analysis and k-nearest neighbor: a case analysis of facial images , 2020, Journal of Physics: Conference Series.

[13]  Bofeng Zhang,et al.  Information Gain with Weight Based Decision Tree for the Employment Forecasting of Undergraduates , 2013, 2013 IEEE International Conference on Green Computing and Communications and IEEE Internet of Things and IEEE Cyber, Physical and Social Computing.

[14]  Kajal Rai,et al.  Decision Tree Based Algorithm for Intrusion Detection , 2016 .

[15]  Dan A. Simovici,et al.  Functional entropy and decision trees , 1998, Proceedings. 1998 28th IEEE International Symposium on Multiple- Valued Logic (Cat. No.98CB36138).

[16]  Adel Sabry Eesa,et al.  A novel feature-selection approach based on the cuttlefish optimization algorithm for intrusion detection systems , 2015, Expert Syst. Appl..

[17]  A. Grigoriadis,et al.  Random Forest Modelling of High-Dimensional Mixed-Type Data for Breast Cancer Classification , 2021, Cancers.

[18]  Qin Ding,et al.  k-nearest Neighbor Classification on Spatial Data Streams Using P-trees , 2002, PAKDD.

[19]  Habibollah Haron,et al.  Machine learning and Region Growing for Breast Cancer Segmentation , 2019, 2019 International Conference on Advanced Science and Engineering (ICOASE).

[20]  Adel Sabry Eesa,et al.  A DIDS Based on The Combination of Cuttlefish Algorithm and Decision Tree , 2017 .

[21]  Sultan Moulay Slimane,et al.  Comparative Analysis of Random Forest and J48 Classifiers for 1 ”IRIS” Variety Prediction 2 , 2020 .

[22]  Sunita B Aher,et al.  Data Mining in Educational System using WEKA , 2011 .

[23]  Sharath Pankanti,et al.  Privacy Enhanced Decision Tree Inference , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).

[24]  Y. Zhao,et al.  Comparison of decision tree methods for finding active objects , 2007, 0708.4274.

[25]  Yasin N. Silva,et al.  Pivot-based approximate k-NN similarity joins for big high-dimensional data , 2020, Inf. Syst..

[26]  Jianping Gou,et al.  A Novel Weighted Voting for K-Nearest Neighbor Rule , 2011, J. Comput..

[27]  Chenglin Wen,et al.  Fault Detection Using Random Projections and k-Nearest Neighbor Rule for Semiconductor Manufacturing Processes , 2015, IEEE Transactions on Semiconductor Manufacturing.

[28]  Chaoyu Gong,et al.  Cumulative belief peaks evidential K-nearest neighbor clustering , 2020, Knowl. Based Syst..

[29]  Shiju Sathyadevan,et al.  Comparative Analysis of Decision Tree Algorithms: ID3, C4.5 and Random Forest , 2015, CI 2015.

[30]  Nawzat Sadiq Ahmed,et al.  Clarify of the Random Forest Algorithm in an Educational Field , 2018, 2018 International Conference on Advanced Science and Engineering (ICOASE).

[31]  Bahzad Charbuty,et al.  Classification Based on Decision Tree Algorithm for Machine Learning , 2021, Journal of Applied Science and Technology Trends.

[32]  Donald E. Brown,et al.  Text Classification Algorithms: A Survey , 2019, Inf..

[33]  Ali Kashif Bashir,et al.  COVID-19 Patient Health Prediction Using Boosted Random Forest Algorithm , 2020, Frontiers in Public Health.

[34]  Zeki Bilgin,et al.  Explaining Inaccurate Predictions of Models through k-Nearest Neighbors , 2021, ICAART.

[35]  Haza Nuzly Abdul Hamed,et al.  Improved Threshold Based and Trainable Fully Automated Segmentation for Breast Cancer Boundary and Pectoral Muscle in Mammogram Images , 2020, IEEE Access.

[36]  Michael E. Houle,et al.  NN-Descent on High-Dimensional Data , 2018, WIMS.

[37]  Ian H. Witten,et al.  WEKA - Experiences with a Java Open-Source Project , 2010, J. Mach. Learn. Res..

[38]  Andy Liaw,et al.  Classification and Regression by randomForest , 2007 .

[39]  Roberto Tagliaferri,et al.  Data Mining: Accuracy and Error Measures for Classification and Prediction , 2019, Encyclopedia of Bioinformatics and Computational Biology.

[40]  Habibollah Haron,et al.  Trainable Model Based on New Uniform LBP Feature to Identify the Risk of the Breast Cancer , 2019, 2019 International Conference on Advanced Science and Engineering (ICOASE).

[41]  Adnan Mohsin Abdulazeez,et al.  Machine Learning Classification Based on Radom Forest Algorithm: A Review , 2021 .

[42]  Diyar Qader Zeebaree,et al.  A Comprehensive Review of Dimensionality Reduction Techniques for Feature Selection and Feature Extraction , 2020, Journal of Applied Science and Technology Trends.

[43]  Sahar Shahaf,et al.  Probabilistic Random Forest: A Machine Learning Algorithm for Noisy Data Sets , 2018, The Astronomical Journal.

[44]  Dong Liang,et al.  Feature Comparison and Optimization for 30-M Winter Wheat Mapping Based on Landsat-8 and Sentinel-2 Data Using Random Forest Algorithm , 2019, Remote. Sens..

[45]  Habibollah Haron,et al.  Multi-Level Fusion in Ultrasound for Cancer Detection based on Uniform LBP Features , 2021 .

[46]  Wlodzislaw Duch,et al.  Comparison of Shannon, Renyi and Tsallis Entropy Used in Decision Trees , 2006, ICAISC.

[47]  Rohit Tyagi,et al.  Comparison of ML classifiers for Image Data , 2020 .

[48]  Ilia Nouretdinov,et al.  Prediction with Confidence Based on a Random Forest Classifier , 2010, AIAI.

[49]  Sebastian Raschka,et al.  Naive Bayes and Text Classification I - Introduction and Theory , 2014, ArXiv.

[50]  Gerard B. M. Heuvelink,et al.  Sampling design optimization for soil mapping with random forest , 2019 .

[51]  Priyanka,et al.  Decision tree classifier: a detailed survey , 2020, Int. J. Inf. Decis. Sci..

[52]  A Qualitative Performance Comparison Of Supervised Machine Learning Algorithms For Iris Recognition , 2020 .

[53]  Mukesh Patel,et al.  A Comparison of Haar Wavelets and Kekre’s Wavelets for Storing Colour Information in a Greyscale Image , 2010 .

[54]  Dastan Hussen Maulud,et al.  A Review on Linear Regression Comprehensive in Machine Learning , 2020 .

[55]  Amelec Viloria,et al.  Integration of Data Mining Techniques to PostgreSQL Database Manager System , 2019, Procedia Computer Science.

[56]  Amirsina Torfi Nearest Neighbor Classifier – From Theory to Practice , 2020 .

[57]  Priyanka Mishra,et al.  Flower Classification using Supervised Learning , 2020 .

[58]  S. P. Shantharajah,et al.  A Study on NSL-KDD Dataset for Intrusion Detection System Based on Classification Algorithms , 2015 .

[59]  Jianping Gou,et al.  A new distance-weighted k-nearest neighbor classifier , 2012 .

[60]  Raj Kumar,et al.  Classification Algorithms for Data Mining: a Survey , 2022 .