Beam search is used to maintain tractability in large search spaces at the expense of completeness and optimality. We study supervised learning of linear ranking functions for controlling beam search. The goal is to learn ranking functions that allow for beam search to perform nearly as well as unconstrained search while gaining computational efficiency. We first study the computational complexity of the learning problem, showing that even for exponentially large search spaces the general consistency problem is in NP. We also identify tractable and intractable subclasses of the learning problem. Next, we analyze the convergence of recently proposed and modified online learning algorithms. We first provide a counter-example to an existing convergence result and then introduce alternative notions of "margin" that do imply convergence. Finally, we study convergence properties for ambiguous training data.
[1]
Thomas G. Dietterich,et al.
Solving the Multiple Instance Problem with Axis-Parallel Rectangles
,
1997,
Artif. Intell..
[2]
David S. Johnson,et al.
Computers and Intractability: A Guide to the Theory of NP-Completeness
,
1978
.
[3]
Albert B Novikoff,et al.
ON CONVERGENCE PROOFS FOR PERCEPTRONS
,
1963
.
[4]
Alan Fern,et al.
Discriminative Learning of Beam-Search Heuristics for Planning
,
2007,
IJCAI.
[5]
Daniel Marcu,et al.
Learning as search optimization: approximate large margin methods for structured prediction
,
2005,
ICML.
[6]
Dan Roth,et al.
Learnability of Bipartite Ranking Functions
,
2005,
COLT.
[7]
Hans Ulrich Simon,et al.
Robust Trainability of Single Neurons
,
1995,
J. Comput. Syst. Sci..