Fast and Flexible Monotonic Functions with Ensembles of Lattices

For many machine learning problems, there are some inputs that are known to be positively (or negatively) related to the output, and in such cases training the model to respect that monotonic relationship can provide regularization, and makes the model more interpretable. However, flexible monotonic functions are computationally challenging to learn beyond a few features. We break through this barrier by learning ensembles of monotonic calibrated interpolated look-up tables (lattices). A key contribution is an automated algorithm for selecting feature subsets for the ensemble base models. We demonstrate that compared to random forests, these ensembles produce similar or better accuracy, while providing guaranteed monotonicity consistent with prior knowledge, smaller model size and faster evaluation.

[1]  Francis R. Bach,et al.  Learning with Submodular Functions: A Convex Optimization Perspective , 2011, Found. Trends Mach. Learn..

[2]  Ponnuthurai N. Suganthan,et al.  Random Forests with ensemble of feature spaces , 2014, Pattern Recognit..

[3]  Yaser S. Abu-Mostafa,et al.  A Method for Learning From Hints , 1992, NIPS.

[4]  Tin Kam Ho,et al.  The Random Subspace Method for Constructing Decision Forests , 1998, IEEE Trans. Pattern Anal. Mach. Intell..

[5]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.

[6]  Rudolf Winter-Ebmer,et al.  A Meta-Analysis of the International Gender Wage Gap , 2003, SSRN Electronic Journal.

[7]  Yoshua Bengio,et al.  Incorporating Functional Knowledge in Neural Networks , 2009, J. Mach. Learn. Res..

[8]  Yunming Ye,et al.  Stratified sampling for feature subspace selection in random forests for high dimensional data , 2013, Pattern Recognit..

[9]  Senén Barro,et al.  Do we need hundreds of classifiers to solve real world classification problems? , 2014, J. Mach. Learn. Res..

[10]  Joann M. Taylor,et al.  Digital Color Imaging Handbook , 2004 .

[11]  H. Wan,et al.  Least Squares Isotonic Regression in Two Dimensions , 2003 .

[12]  Maya R. Gupta,et al.  A Light Touch for Heavily Constrained SGD , 2015, COLT.

[13]  Yung-Seop Lee,et al.  Enriched random forests , 2008, Bioinform..

[14]  Wojciech Kotlowski,et al.  Rule learning with monotonicity constraints , 2009, ICML '09.

[15]  Marina Velikova,et al.  Monotone and Partially Monotone Neural Networks , 2010, IEEE Transactions on Neural Networks.

[16]  Bao-Gang Hu,et al.  Generalized Constraint Neural Network Regression Model Subject to Linear Priors , 2011, IEEE Transactions on Neural Networks.

[17]  Sanjeev Khudanpur,et al.  Modeling phonetic context with non-random forests for speech recognition , 2015, INTERSPEECH.

[18]  Maya R. Gupta,et al.  Optimized Regression for Efficient Function Evaluation , 2012, IEEE Transactions on Image Processing.

[19]  Catherine Blake,et al.  UCI Repository of machine learning databases , 1998 .

[20]  Maya R. Gupta,et al.  Monotonic Calibrated Interpolated Look-Up Tables , 2015, J. Mach. Learn. Res..

[21]  Tony Jebara,et al.  Learning Monotonic Transformations for Classification , 2007, NIPS.

[22]  Maya R. Gupta,et al.  Lattice Regression , 2009, NIPS.