Effect of Irrelevant Variables on Faulty Wafer Detection in Semiconductor Manufacturing

Machine learning has been applied successfully for faulty wafer detection tasks in semiconductor manufacturing. For the tasks, prediction models are built with prior data to predict the quality of future wafers as a function of their precedent process parameters and measurements. In real-world problems, it is common for the data to have a portion of input variables that are irrelevant to the prediction of an output variable. The inclusion of many irrelevant variables negatively affects the performance of prediction models. Typically, prediction models learned by different learning algorithms exhibit different sensitivities with regard to irrelevant variables. Algorithms with low sensitivities are preferred as a first trial for building prediction models, whereas a variable selection procedure is necessarily considered for highly sensitive algorithms. In this study, we investigate the effect of irrelevant variables on three well-known representative learning algorithms that can be applied to both classification and regression tasks: artificial neural network, decision tree (DT), and k-nearest neighbors (k-NN). We analyze the characteristics of these learning algorithms in the presence of irrelevant variables with different model complexity settings. An empirical analysis is performed using real-world datasets collected from a semiconductor manufacturer to examine how the number of irrelevant variables affects the behavior of prediction models trained with different learning algorithms and model complexity settings. The results indicate that the prediction accuracy of k-NN is highly degraded, whereas DT demonstrates the highest robustness in the presence of many irrelevant variables. In addition, a higher model complexity of learning algorithms leads to a higher sensitivity to irrelevant variables.

[1]  Anil K. Jain,et al.  Feature Selection: Evaluation, Application, and Small Sample Performance , 1997, IEEE Trans. Pattern Anal. Mach. Intell..

[2]  S. C. Neoh,et al.  A Micro-GA Embedded PSO Feature Selection Approach to Intelligent Facial Emotion Recognition , 2017, IEEE Transactions on Cybernetics.

[3]  Ping Zhang,et al.  Feature selection considering the composition of feature relevancy , 2018, Pattern Recognit. Lett..

[4]  Anil K. Jain,et al.  Small Sample Size Effects in Statistical Pattern Recognition: Recommendations for Practitioners , 1991, IEEE Trans. Pattern Anal. Mach. Intell..

[5]  Hyoungjoo Lee,et al.  A virtual metrology system for semiconductor manufacturing , 2009, Expert Syst. Appl..

[6]  Isabelle Guyon,et al.  An Introduction to Variable and Feature Selection , 2003, J. Mach. Learn. Res..

[7]  Wei-Yin Loh,et al.  Fifty Years of Classification and Regression Trees , 2014 .

[8]  Chaouki Khammassi,et al.  A GA-LR wrapper approach for feature selection in network intrusion detection , 2017, Comput. Secur..

[9]  Inci Batmaz,et al.  A review of data mining applications for quality improvement in manufacturing industry , 2011, Expert Syst. Appl..

[10]  Jerome R. Busemeyer,et al.  The effect of "irrelevant" variables on decision making: Criterion shifts in preferential choice? , 1992 .

[11]  Cheng-Ching Yu,et al.  Control relevant issues in semiconductor manufacturing : Overview with some new results , 2007 .

[12]  Katharina Morik,et al.  Quality Prediction in Interlinked Manufacturing Processes based on Supervised & Unsupervised Machine Learning , 2013 .

[13]  Andrew P. Bradley,et al.  The use of the area under the ROC curve in the evaluation of machine learning algorithms , 1997, Pattern Recognit..

[14]  David W. Aha,et al.  Tolerating Noisy, Irrelevant and Novel Attributes in Instance-Based Learning Algorithms , 1992, Int. J. Man Mach. Stud..

[15]  John Mingers,et al.  An Empirical Comparison of Pruning Methods for Decision Tree Induction , 1989, Machine Learning.

[16]  Vladislav Maxim,et al.  Embedded vision equipment of industrial robot for inline detection of product errors by clustering–classification algorithms , 2016 .

[17]  Leslie S. Smith,et al.  Feature subset selection in large dimensionality domains , 2010, Pattern Recognit..

[18]  Thomas Blaschke,et al.  Evaluation of Feature Selection Methods for Object-Based Land Cover Mapping of Unmanned Aerial Vehicle Imagery Using Random Forest and Support Vector Machine Classifiers , 2017, ISPRS Int. J. Geo Inf..

[19]  Maria Cláudia Reis Cavalcanti,et al.  Automatic feature selection for supervised learning in link prediction applications: a comparative study , 2017, Knowledge and Information Systems.

[20]  L. Nelson Sanchez-Pinto,et al.  Comparison of variable selection methods for clinical predictive modeling , 2018, Int. J. Medical Informatics.

[21]  Gaël Varoquaux,et al.  Scikit-learn: Machine Learning in Python , 2011, J. Mach. Learn. Res..

[22]  Sungzoon Cho,et al.  Efficient Feature Selection-Based on Random Forward Search for Virtual Metrology Modeling , 2016, IEEE Transactions on Semiconductor Manufacturing.

[23]  António Pacheco,et al.  Theoretical foundations of forward feature selection methods based on mutual information , 2017, Neurocomputing.

[24]  Sotiris B. Kotsiantis,et al.  Decision trees: a recent overview , 2011, Artificial Intelligence Review.

[25]  Youngjae Chang,et al.  Variable Selection via Regression Trees in the Presence of Irrelevant Variables , 2013, Commun. Stat. Simul. Comput..

[26]  Yong Zhang,et al.  A PSO-based multi-objective multi-label feature selection method in classification , 2017, Scientific Reports.

[27]  Franco Scarselli,et al.  On the Complexity of Neural Network Classifiers: A Comparison Between Shallow and Deep Architectures , 2014, IEEE Transactions on Neural Networks and Learning Systems.

[28]  Holger R. Maier,et al.  Non-linear variable selection for artificial neural networks using partial mutual information , 2008, Environ. Model. Softw..

[29]  J DhaliaSweetlin,et al.  Feature selection using ant colony optimization with tandem-run recruitment to diagnose bronchitis from CT scan images , 2017, Comput. Methods Programs Biomed..

[30]  Yong Xia,et al.  GA-SVM based feature selection and parameter optimization in hospitalization expense modeling , 2019, Appl. Soft Comput..

[31]  Ping Zhang,et al.  Class-specific mutual information variation for feature selection , 2018, Pattern Recognit..

[32]  Charles X. Ling,et al.  Using AUC and accuracy in evaluating learning algorithms , 2005, IEEE Transactions on Knowledge and Data Engineering.

[33]  Jin Wang,et al.  Fault Detection Using the k-Nearest Neighbor Rule for Semiconductor Manufacturing Processes , 2007, IEEE Transactions on Semiconductor Manufacturing.

[34]  Claudio De Stefano,et al.  A GA-based feature selection approach with an application to handwritten character recognition , 2014, Pattern Recognit. Lett..

[35]  Majdi M. Mafarja,et al.  Hybrid Whale Optimization Algorithm with simulated annealing for feature selection , 2017, Neurocomputing.

[36]  Li Xiu,et al.  Application of data mining techniques in customer relationship management: A literature review and classification , 2009, Expert Syst. Appl..

[37]  Cheng-Lung Huang,et al.  A GA-based feature selection and parameters optimizationfor support vector machines , 2006, Expert Syst. Appl..

[38]  Reha Uzsoy,et al.  A REVIEW OF PRODUCTION PLANNING AND SCHEDULING MODELS IN THE SEMICONDUCTOR INDUSTRY PART I: SYSTEM CHARACTERISTICS, PERFORMANCE EVALUATION AND PRODUCTION PLANNING , 1992 .

[39]  Nasser R. Sabar,et al.  An Exponential Monte-Carlo algorithm for feature selection problems , 2014, Comput. Ind. Eng..

[40]  Yaochu Jin,et al.  Feature selection for high-dimensional classification using a competitive swarm optimizer , 2016, Soft Computing.

[41]  Thomas B. Fomby,et al.  Loss of efficiency in regression analysis due to irrelevant variables: A generalization , 1981 .

[42]  Chen-Fu Chien,et al.  Semiconductor fault detection and classification for yield enhancement and manufacturing intelligence , 2012, Flexible Services and Manufacturing Journal.

[43]  Ferat Sahin,et al.  A survey on feature selection methods , 2014, Comput. Electr. Eng..

[44]  Hyoungjoo Lee,et al.  Machine learning-based novelty detection for faulty wafer detection in semiconductor manufacturing , 2012, Expert Syst. Appl..