Aggregating predictions of multi-models for the small dataset learning tasks in the TFT-LCD process

Compressing new product development schedules has become an important strategy for firms to earn more market share in recent years, because of the globalizing competition. Nevertheless, engineers often have to face the issue of small data learning in this context, since only few pilot runs are allowed to avoid lowering the yield rates of mass production. Bagging has shown to be an effective approach to deal with small data. However, when adopting multiple models in bagging to learn numeric forecasting tasks without the prior weights, bagging often takes the averages of the predictions of the models as the compromised ones, which are very probably affected by extreme values and thus become less precise. Accordingly, this study develops a systematic procedure to generate weights for prediction aggregation by employing box-and-whisker plots to build the possible distributions of the predictive errors of the algorithms with membership functions. In the experiment, a real case taken from a thin film transistor liquid crystal display maker is studied with four distinct algorithms and bagging; and the results show the errors of the predictions aggregated by the proposed weights are significantly lower than those of the four algorithms and the averaging method in bagging in general.

[1]  Francis K. H. Quek,et al.  Attribute bagging: improving accuracy of classifier ensembles by using random feature subsets , 2003, Pattern Recognit..

[2]  Richard J. Roiger,et al.  Data Mining: A Tutorial Based Primer , 2002 .

[3]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.

[4]  Gavin Brown,et al.  Random Ordinality Ensembles: Ensemble methods for multi-valued categorical data , 2015, Inf. Sci..

[5]  Ian H. Witten,et al.  Induction of model trees for predicting continuous classes , 1996 .

[6]  Yunqian Ma,et al.  Comparison of Model Selection for Regression , 2003, Neural Computation.

[7]  Padhraic Smyth,et al.  From Data Mining to Knowledge Discovery in Databases , 1996, AI Mag..

[8]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[9]  Ning Li,et al.  Force-based tool condition monitoring for turning process using v-support vector regression , 2017 .

[10]  Ronald R. Yager,et al.  Building ensemble classifiers using belief functions and OWA operators , 2008, Soft Comput..

[11]  Joaquim Ciurana,et al.  Random Forest ensemble prediction of stent dimensions in microfabrication processes , 2017 .

[12]  Alexander J. Smola,et al.  Support Vector Regression Machines , 1996, NIPS.

[13]  Zhi-Hua Zhou,et al.  Exploratory Under-Sampling for Class-Imbalance Learning , 2006, Sixth International Conference on Data Mining (ICDM'06).

[14]  Saso Dzeroski,et al.  Combining Multiple Models with Meta Decision Trees , 2000, PKDD.

[15]  John W. Tukey,et al.  Exploratory Data Analysis. , 1979 .

[16]  R. Tibshirani,et al.  An introduction to the bootstrap , 1993 .

[17]  R. Tibshirani Regression Shrinkage and Selection via the Lasso , 1996 .

[18]  Yu Ding,et al.  A classification procedure for highly imbalanced class sizes , 2010 .

[19]  Angappa Gunasekaran,et al.  A knowledge management approach for managing uncertainty in manufacturing , 2006, Ind. Manag. Data Syst..

[20]  Thomas G. Dietterich Multiple Classifier Systems , 2000, Lecture Notes in Computer Science.

[21]  Bin Lin,et al.  Path planning for support heads in mirror-milling machining system , 2017 .

[22]  Deniz Erdogmus,et al.  Subject-specific abnormal region detection in traumatic brain injury using sparse model selection on high dimensional diffusion data , 2017, Medical Image Anal..