Novel algorithm to measure consistency between extracted models from big dataset and predicting applicability of rule extraction

Many advancement is made in recent days and number of techniques are proposed by different researchers for processing and extracting knowledge from big data. But to evaluate the consistency in extracted model is always questionable. In this paper we are presenting two techniques for measuring the consistency between extracted model and predicting their applicability. In this paper, Meta learning based approach using characteristics of dataset is designed through which it can be identified whether the rule extraction technique will going to produce a better model as compare to conventional algorithm. Meta learning is concerned to identify the relationship between learning techniques and different big datasets. The proposed model is very generic and can be used in many different problems.

[1]  Susan Craw,et al.  CONSULTANT: providing advice for the machine learning toolbox , 1993 .

[2]  Tobias Scheffer Average-Case Analysis of Classification Algorithms for Boolean Functions and Decision Trees , 2000, ALT.

[3]  Larry A. Rendell,et al.  Layered Concept-Learning and Dynamically Variable Bias Management , 1987, IJCAI.

[4]  Peter A. Flach,et al.  Decision Tree-Based Data Characterization for Meta-Learning , 2002 .

[5]  Kate Smith-Miles,et al.  Cross-disciplinary perspectives on meta-learning for algorithm selection , 2009, CSUR.

[6]  Ricardo Vilalta,et al.  Metalearning - Applications to Data Mining , 2008, Cognitive Technologies.

[7]  Abraham Bernstein,et al.  An Intelligent Assistant for the Knowledge Discovery Process , 2001, IJCAI 2001.

[8]  Lior Rokach,et al.  Meta-learning for Selecting a Multi-label Classification Algorithm , 2011, 2011 IEEE 11th International Conference on Data Mining Workshops.

[9]  Abraham Bernstein,et al.  Toward intelligent assistance for a data mining process: an ontology-based approach for cost-sensitive classification , 2005, IEEE Transactions on Knowledge and Data Engineering.

[10]  Alexandros Kalousis,et al.  NOEMON: Design, implementation and performance results of an intelligent assistant for classifier selection , 1999, Intell. Data Anal..

[11]  Kate Smith-Miles,et al.  Meta-learning for data summarization based on instance selection method , 2010, IEEE Congress on Evolutionary Computation.

[12]  Asma Abu Salah,et al.  Meta-Learning Evolutionary Artificial Neural Networks: by Means of Cellular Automata , 2005, International Conference on Computational Intelligence for Modelling, Control and Automation and International Conference on Intelligent Agents, Web Technologies and Internet Commerce (CIMCA-IAWTIC'06).

[13]  Kate Smith-Miles Meta-Learning: From Classification to Forecasting, to Optimization, and Beyond , 2007, 6th IEEE/ACIS International Conference on Computer and Information Science (ICIS 2007).

[14]  João Gama,et al.  On Data and Algorithms: Understanding Inductive Performance , 2004, Machine Learning.

[15]  Norbert Jankowski,et al.  Versatile and Efficient Meta-Learning Architecture: Knowledge Representation and Management in Computational Intelligence , 2007, 2007 IEEE Symposium on Computational Intelligence and Data Mining.

[16]  Antonio F. Gómez-Skarmeta,et al.  METALA: a J2EE Technology Based Framework for Web Mining , 2004, Rev. Colomb. de Computación.

[17]  Pat Langley,et al.  Tractable Average-Case Analysis of Naive Bayesian Classifiers , 1999, ICML.

[18]  Ricardo Vilalta,et al.  A Characterization of Difficult Problems in Classification , 2002, ICMLA.

[19]  D. Wolpert The Supervised Learning No-Free-Lunch Theorems , 2002 .

[20]  Thomas G. Dietterich What is machine learning? , 2020, Archives of Disease in Childhood.

[21]  Maarten van Someren Model Class Selection and Construction: Beyond the Procrustean Approach to Machine Learning Applications , 2001, Machine Learning and Its Applications.

[22]  David J. Spiegelhalter,et al.  Machine Learning, Neural and Statistical Classification , 2009 .

[23]  Alexander Schliep,et al.  Ranking and selecting clustering algorithms using a meta-learning approach , 2008, 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence).

[24]  Saso Dzeroski,et al.  Ranking with Predictive Clustering Trees , 2002, ECML.

[25]  Abraham Bernstein,et al.  A survey of intelligent assistants for data analysis , 2013, CSUR.

[26]  Pavel Brazdil,et al.  An Iterative Process for Building Learning Curves and Predicting Relative Performance of Classifiers , 2007, EPIA Workshops.

[27]  Paul E. Utgoff,et al.  Shift of bias for inductive concept learning , 1984 .

[28]  Rodica Potolea,et al.  Evolutional meta-learning framework for automatic classifier selection , 2009, 2009 IEEE 5th International Conference on Intelligent Computer Communication and Processing.

[29]  Christophe G. Giraud-Carrier,et al.  The data mining advisor: meta-learning at the service of practitioners , 2005, Fourth International Conference on Machine Learning and Applications (ICMLA'05).

[30]  Hilan Bensusan,et al.  Meta-Learning by Landmarking Various Learning Algorithms , 2000, ICML.

[31]  Henrik Jacobsson,et al.  Rule extraction from recurrent neural networks , 2006 .

[32]  Bart Baesens,et al.  A new approach for measuring rule set consistency , 2007, Data Knowl. Eng..