Improving the Union Bound: a Distribution Dependent Approach

Statistical Learning Theory deals with the problem of estimating the performance of a learning procedure. Any learning procedure implies making choices and this choices imply a risk. When the number of choices is finite, the state-of-the-art tool for evaluating the total risk of all the choice made is the Union Bound. The problem of the Union Bound is that it is very loose in practice if no a-priori information is available. In fact, the Union Bound considers all choices equally plausible while, as a matter of fact, a learning procedure targets just particular choices disregarding the others. In this work we will show that it is possible to improve the Union Bound based results using a distribution dependent weighting strategy of the true risks associated to each choice. Then we will prove that our proposal outperforms or, in the worst case, it degenerate in the Union Bound.

[1]  John Langford,et al.  Computable Shell Decomposition Bounds , 2000, J. Mach. Learn. Res..

[2]  Vladimir Vapnik,et al.  Statistical learning theory , 1998 .

[3]  P. Bartlett,et al.  Local Rademacher complexities , 2005, math/0508275.

[4]  Luca Oneto,et al.  Model Selection and Error Estimation in a Nutshell , 2020, Modeling and Optimization in Science and Technologies.

[5]  Eric R. Ziegel,et al.  The Elements of Statistical Learning , 2003, Technometrics.

[6]  John Langford,et al.  Quantitatively tight sample complexity bounds , 2002 .

[7]  Andrew W. Moore,et al.  The Racing Algorithm: Model Selection for Lazy Learners , 1997, Artificial Intelligence Review.

[8]  Larry Wasserman,et al.  Genome-Wide Significance Levels and Weighted Hypothesis Testing. , 2009, Statistical science : a review journal of the Institute of Mathematical Statistics.

[9]  Shai Ben-David,et al.  Understanding Machine Learning: From Theory to Algorithms , 2014 .

[10]  W. Hoeffding Probability Inequalities for sums of Bounded Random Variables , 1963 .