This paper addresses the problem of learning boolean functions in query and mistake-bound models in the presence of irrelevant attributes. In learning a concept, a learner may observe a great many more attributes than those that the concept depends upon, and in some sense the presence of extra, irrelevant attributes does not change the underlying concept being learned. Because of this, we are interested not only in the learnability of concept classes, but also in whether the classes can be learned by an algorithm that is attribute-efficient in that the dependence of the mistake bound (or number of queries) on the number of irrelevant attributes is low. The results presented here apply to projection and embedding-closed (p.e.c.) concept classes. We show that if a p.e.c. class is learnable attribute-efficiently in the mistake-bound model, then it is learnable in the infinite-attribute mistake-bound model as well. We show in addition how to convert any algorithm that learns a p.e.c. dass in the mistake-bound model with membership queries into an algorithm that learns the class attribute-efficiently in that model, or even in the infinite attribute version. In the membership query only model we show that learnability does not always imply attribute-efficient learnability for deterministic algorithms. However, we describe a large class of functions, including the set of monotone functions, for which learnability does imply attribute-efficient learnability in this model.
[1]
Marek Karpinski,et al.
Learning read-once formulas with queries
,
1993,
JACM.
[2]
Wolfgang Maass,et al.
On the complexity of learning from counterexamples and membership queries
,
1990,
Proceedings [1990] 31st Annual Symposium on Foundations of Computer Science.
[3]
Leonard Pitt,et al.
On the necessity of Occam algorithms
,
1990,
STOC '90.
[4]
Avrim Blum,et al.
Learning boolean functions in an infinite attribute space
,
1990,
STOC '90.
[5]
David Haussler,et al.
Learnability and the Vapnik-Chervonenkis dimension
,
1989,
JACM.
[6]
N. Littlestone.
Learning Quickly When Irrelevant Attributes Abound: A New Linear-Threshold Algorithm
,
1987,
28th Annual Symposium on Foundations of Computer Science (sfcs 1987).
[7]
David Haussler,et al.
Classifying learnable geometric concepts with the Vapnik-Chervonenkis dimension
,
1986,
STOC '86.
[8]
Temple F. Smith.
Occam's razor
,
1980,
Nature.