We introduce a significant improvement for a relatively new machine learning method called Transformation-Based Learning. By applying a Monte Carlo strategy to randomly sample from the space of rules, rather than exhaustively analyzing all possible rules, we drastically reduce the memory and time costs of the algorithm, without compromising accuracy on unseen data. This enables Transformation-Based Learning to apply to a wider range of domains, as it can effectively consider a larger number of different features and feature interactions in the data. In addition, the Monte Carlo improvement decreases the labor demands on the human developer, who no longer needs to develop a minimal set of rule templates to maintain tractability.
[1]
Giorgio Satta,et al.
String Transformation Learning
,
1997,
ACL.
[2]
Mitchell P. Marcus,et al.
Exploring the Statistical Derivation of Transformational Rule Sequences for Part-of-Speech Tagging
,
1994,
ArXiv.
[3]
Ken Samuel,et al.
Computing Dialogue Acts from Features with Transformation-Based Learning
,
1998,
ArXiv.
[4]
Eric Brill,et al.
Transformation-Based Error-Driven Learning and Natural Language Processing: A Case Study in Part-of-Speech Tagging
,
1995,
CL.
[5]
Eric Brill,et al.
An Overview of Empirical Natural Language Processing
,
1997,
AI Mag..