A Comparative Evaluation of Voting and Meta-learning on Partitioned Data

Abstract Much of the research in inductive learning concentrates on problems with relatively small amounts of data. With the coming age of very large network computing, it is likely that orders of magnitude more data in databases will be available for various learning problems of real world importance. Some learning algorithms assume that the entire data set fits into main memory, which is not feasible for massive amounts of data. One approach to handling a large data set is to partition the data set into subsets, run the learning algorithm on each of the subsets, and combine the results. In this paper we evaluate different techniques for learning from partitioned data. Our meta-learning approach is empirically compared with techniques in the literature that aim to combine multiple evidence to arrive at one prediction.

[1]  Salvatore J. Stolfo,et al.  Integrating multiple learned models for improving and scaling machine learning algorithms , 1996 .

[2]  Adam Krzyżak,et al.  Methods of combining multiple classifiers and their applications to handwriting recognition , 1992, IEEE Trans. Syst. Man Cybern..

[3]  Salvatore J. Stolfo,et al.  The ALEXSYS Mortgage Pool Allocation Expert System : A Case Study of Speeding Up Rule-based Programs , 1990 .

[4]  Salvatore J. Stolfo,et al.  Report on Workshop on High Performance Computing and Communications for Grand Challenge Applications: Computer Vision, Speech and Natural Language Processing, and Artificial Intelligence , 1993, IEEE Trans. Knowl. Data Eng..

[5]  Wray L. Buntine,et al.  Introduction in IND and recursive partitioning , 1991 .

[6]  Salvatore J. Stolfo,et al.  Toward Multi-Strategy Parallel & Distributed Learning in Sequence Analysis , 1993, ISMB.

[7]  Jason Catlett,et al.  Experiments on the Costs and Benefits of Windowing in ID3 , 1988, ML.

[8]  Salvatore J. Stolfo,et al.  Learning Arbiter and Combiner Trees from Partitioned Data for Scaling Machine Learning , 1995, KDD.

[9]  Salvatore J. Stolfo,et al.  Toward parallel and distributed learning by meta-learning , 1993 .

[10]  Jude W. Shavlik,et al.  Learning to Represent Codons: A Challenge Problem for Constructive Induction , 1993, IJCAI.

[11]  Philip K. Chan,et al.  Learning Patterns from Unix Process Execution Traces for Intrusion Detection , 1997 .

[12]  J. Mesirov,et al.  Hybrid system for protein secondary structure prediction. , 1992, Journal of molecular biology.

[13]  Jude Shavlik,et al.  Refinement ofApproximate Domain Theories by Knowledge-Based Neural Networks , 1990, AAAI.

[14]  Salvatore J. Stolfo,et al.  Experiments on multistrategy learning by meta-learning , 1993, CIKM '93.

[15]  Salvatore J. Stolfo,et al.  An extensible meta-learning approach for scalable and accurate inductive learning , 1996 .

[16]  Philip K. Chan,et al.  Machine Learning in Molecular Biology Sequence Analysis , 1991 .

[17]  David W. Aha,et al.  Noise-Tolerant Instance-Based Learning Algorithms , 1989, IJCAI.

[18]  J. R. Quinlan,et al.  Induction over large data bases , 1979 .

[19]  David H. Wolpert,et al.  Stacked generalization , 1992, Neural Networks.

[20]  Jill P. Mesirov,et al.  An Efficient Implementation of the Back-propagation Algorithm on the Connection Machine CM-2 , 1989, NIPS.

[21]  Manfred K. Warmuth,et al.  The Weighted Majority Algorithm , 1994, Inf. Comput..

[22]  Philip K. Chan,et al.  Inductive Learning with BCT , 1989, ML.

[23]  Jason Catlett,et al.  Megainduction: A Test Flight , 1991, ML.

[24]  P. Chan,et al.  Toward multistrategy parallel and distributed learning in sequence analysis. , 1993, Proceedings. International Conference on Intelligent Systems for Molecular Biology.