We describe a language model employing a new headed-disjuncts formulation of Lafferty et al.'s (1992) probabilistic link grammar, together with (1) an EM training method for estimating the probabilities, and (2) a procedure for learning some simple lexicalized grammar structures. The model in its simplest form is a generalization of n-gram models, but in its general form possesses context-free expressiveness. Unlike the original experiments on probabilistic link grammars, we assume that no hand-coded grammar is initially available (as with n-gram models). We employ untyped links to concentrate the learning on lexical dependencies, and our formulation uses the lexical identities of heads to influence the structure of the parse graph. After learning, the language model consists of grammatical rules in the form of a set of simple disjuncts for each word, plus several sets of probability parameters. The formulation extends cleanly toward learning more powerful context-free grammars. Several issues relating to generalization bias, linguistic constraints, and parameter smoothing are considered. Preliminary experimental results on small artificial corpora are supportive of our approach.
[1]
Steve Young,et al.
Applications of stochastic context-free grammars using the Inside-Outside algorithm
,
1990
.
[2]
John Lafferty,et al.
Grammatical Trigrams: A Probabilistic Model of Link Grammar
,
1992
.
[3]
J. Baker.
Trainable grammars for speech recognition
,
1979
.
[4]
Daniel Dominic Sleator,et al.
Parsing English with a Link Grammar
,
1995,
IWPT.
[5]
D. Rubin,et al.
Maximum likelihood from incomplete data via the EM - algorithm plus discussions on the paper
,
1977
.
[6]
Frederick Jelinek,et al.
Basic Methods of Probabilistic Context Free Grammars
,
1992
.
[7]
L. Baum,et al.
An inequality and associated maximization technique in statistical estimation of probabilistic functions of a Markov process
,
1972
.
[8]
Glenn Carroll,et al.
Learn-ing probaballstic dependency grammars from labelled text
,
1992
.
[9]
Robert L. Mercer,et al.
An Estimate of an Upper Bound for the Entropy of English
,
1992,
CL.