Towards scalable quantile regression trees

We provide an algorithm to build quantile regression trees in O(N log N) time, where N is the number of instances in the training set. Quantile regression trees are regression trees that model conditional quantiles of the response variable, rather than the conditional expectation as in standard regression trees. We build quantile regression trees by using the quantile loss function in our node splitting criterion. The performance of our algorithm stems from new online update procedures for both the quantile function and the quantile loss function. We test the quantile tree algorithm in three ways, comparing its running time against implementations of standard regression trees, demonstrating its ability to recover a known set of nonlinear quantile functions, and showing that quantile trees yield smaller test set errors (computed using mean absolute deviation) than standard regression trees. The tests include training sets with up to 16 million instances. Overall, our results enable future use of quantile regression trees for large-scale data mining.

[1]  W. Steiger,et al.  Least Absolute Deviations: Theory, Applications and Algorithms , 1984 .

[2]  R. Pace,et al.  Sparse spatial autoregressions , 1997 .

[3]  L. Torgo,et al.  Inductive learning of tree-based regression models , 1999 .

[4]  J. Friedman Greedy function approximation: A gradient boosting machine. , 2001 .

[5]  W. Loh,et al.  Nonparametric estimation of conditional quantiles using quantile regression trees ∗ ( Published in Bernoulli ( 2002 ) , 8 , 561 – 576 ) , 2008 .

[6]  Alok Baveja,et al.  Computing , Artificial Intelligence and Information Technology A data-driven software tool for enabling cooperative information sharing among police departments , 2002 .

[7]  T. E. Dielman,et al.  Least absolute value regression: recent contributions , 2005 .

[8]  R. Koenker Quantile Regression: Name Index , 2005 .

[9]  Nicolai Meinshausen,et al.  Quantile Regression Forests , 2006, J. Mach. Learn. Res..

[10]  Bianca Zadrozny,et al.  High-quantile modeling for customer wallet estimation and other applications , 2007, KDD '07.

[11]  Paulo Cortez,et al.  Using Data Mining for Wine Quality Assessment , 2009, Discovery Science.

[12]  R. Koenker Additive models for quantile regression: Model selection and confidence bandaids , 2010 .

[13]  Conrad Sanderson,et al.  Armadillo: An Open Source C++ Linear Algebra Library for Fast Prototyping and Computationally Intensive Experiments , 2010 .

[14]  Weining Wang,et al.  Local Quantile Regression , 2010, 1208.5384.

[15]  Douglas W. Nychka,et al.  Fast Nonparametric Quantile Regression With Arbitrary Smoothing Methods , 2011 .

[16]  Gabriela Ciuperca,et al.  Penalized least absolute deviations estimation for nonlinear model with change-points , 2011 .

[17]  Alex J. Cannon Quantile regression neural networks: Implementation in R and application to precipitation downscaling , 2011, Comput. Geosci..

[18]  Wei-Yin Loh,et al.  Classification and regression trees , 2011, WIREs Data Mining Knowl. Discov..

[19]  Marilena Pillati,et al.  Notes on the Robustness of Regression Trees Against Skewed and Contaminated Errors , 2011 .

[20]  Gabriela Ciuperca,et al.  Estimating nonlinear regression with and without change-points by the LAD method , 2011 .

[21]  Ying-hui Zhou,et al.  Quantile Regression via the EM Algorithm , 2014, Commun. Stat. Simul. Comput..

[22]  Youngjae Chang,et al.  Multi-step quantile regression tree , 2014 .

[23]  Harish S. Bhat,et al.  Forecasting retained earnings of privately held companies with PCA and L1 regression , 2014 .

[24]  Conrad Sanderson,et al.  RcppArmadillo: Accelerating R with high-performance C++ linear algebra , 2014, Comput. Stat. Data Anal..

[25]  Wei-Yin Loh,et al.  Fifty Years of Classification and Regression Trees , 2014 .

[26]  B. Ripley,et al.  Recursive Partitioning and Regression Trees , 2015 .

[27]  D. Pacák Least Absolute Deviations , 2017 .