This paper reports the implementation of DRAPH-GP an extension of the decision graph algorithm DGRAPH-OW using the AdaBoost algorithm. This algorithm, which we call 1-Stage Boosting, is shown to improve the accuracy of decision graphs, along with another technique which we combine with AdaBoost and call 2-Stage Boosting which shows greater improvement. Empirical tests demonstrate that both 1-Stage and 2-Stage Boosting techniques perform better than the boosted C4.5 algorithm (C5.0). The boosting has shown itself competitive for NLP tasks with a high disjunction of attribute space against memory based methods, and potentially better if part of an Hierarchical Multi-Method Classifier. An explanation for the effectiveness of boosting due to a poor choice of prior probabilities is presented. 1. INTRODUCTION In a wide variety of classification problems, boosting techniques have proven to be an effective method to significantly reduce the error of weak learning algorithms. While the AdaBoost algorithm (Freund & Schapire, 1995) has been used to improve the accuracy of a decision tree algorithm (Quinlan & Rivest, 1989), which uses the Minimum Description Length Principle (MDL), little is known about it's effectiveness on the decision graphs.
[1]
J. Ross Quinlan,et al.
C4.5: Programs for Machine Learning
,
1992
.
[2]
Jonathan J. Oliver.
Decision Graphs - An Extension of Decision Trees
,
1993
.
[3]
C. S. Wallace,et al.
An Information Measure for Classification
,
1968,
Comput. J..
[4]
Xavier Carreras,et al.
Boosting trees for clause splitting
,
2001,
CoNLL.
[5]
Hervé Déjean,et al.
Introduction to the CoNLL-2001 shared task: clause identification
,
2001,
CoNLL.
[6]
Erik F. Tjong Kim Sang,et al.
Memory-based clause identification
,
2001,
CoNLL.
[7]
Ronald L. Rivest,et al.
Inferring Decision Trees Using the Minimum Description Length Principle
,
1989,
Inf. Comput..
[8]
James Hammerton.
Clause identification with long short-term memory
,
2001,
CoNLL.
[9]
Yoav Freund,et al.
A decision-theoretic generalization of on-line learning and an application to boosting
,
1995,
EuroCOLT.
[10]
Hervé Déjean.
Using ALLiS for clausing
,
2001,
CoNLL.
[11]
Walter Daelemans,et al.
Forgetting Exceptions is Harmful in Language Learning
,
1998,
Machine Learning.
[12]
C. S. Wallace,et al.
Coding Decision Trees
,
1993,
Machine Learning.