DDAC-SpAM: A Distributed Algorithm for Fitting High-dimensional Sparse Additive Models with Feature Division and Decorrelation*

Distributed statistical learning has become a popular technique for large-scale data analysis. Most existing work in this area focuses on dividing the observations, but we propose a new algorithm, DDAC-SpAM, which divides the features under a high-dimensional sparse additive model. Our approach involves three steps: divide, decorrelate, and conquer. The decorrelation operation enables each local estimator to recover the sparsity pattern for each additive component without imposing strict constraints on the correlation structure among variables. The effectiveness and efficiency of the proposed algorithm are demonstrated through theoretical analysis and empirical results on both synthetic and real data. The theoretical results include both the consistent sparsity pattern recovery as well as statistical inference for each additive functional component. Our approach provides a practical solution for fitting sparse additive models, with promising applications in a wide range of domains.

[1]  D Zeng,et al.  Equity-specific effects of interventions to promote physical activity among middle-aged and older adults: results from applying a novel equity-specific re-analysis strategy , 2021, International Journal of Behavioral Nutrition and Physical Activity.

[2]  Anru R. Zhang,et al.  Sparse Group Lasso: Optimal Sample Complexity, Convergence Rate, and Statistical Inference , 2019, IEEE Transactions on Information Theory.

[3]  N. Simon,et al.  Generalized Sparse Additive Models , 2019, J. Mach. Learn. Res..

[4]  Daniela Witten,et al.  Data‐adaptive additive modeling , 2018, Statistics in medicine.

[5]  R. Tibshirani,et al.  Additive models with trend filtering , 2017, The Annals of Statistics.

[6]  Chengchun Shi,et al.  A Massive Data Framework for M-Estimators with Cubic-Rate , 2016, Journal of the American Statistical Association.

[7]  Qianchuan He,et al.  Sparse meta-analysis with high-dimensional data. , 2016, Biostatistics.

[8]  Yi Yang,et al.  A fast unified algorithm for solving group-lasso penalize learning problems , 2014, Statistics and Computing.

[9]  F. Liang,et al.  A split‐and‐merge Bayesian variable selection approach for ultrahigh dimensional regression , 2015 .

[10]  Jianqing Fan,et al.  Distributed Estimation and Inference with Statistical Guarantees , 2015, 1509.05457.

[11]  Ashley Petersen,et al.  Fused Lasso Additive Model , 2014, Journal of computational and graphical statistics : a joint publication of American Statistical Association, Institute of Mathematical Statistics, Interface Foundation of North America.

[12]  Johannes Gehrke,et al.  Sparse Partially Linear Additive Models , 2014, ArXiv.

[13]  Minge Xie,et al.  A Split-and-Conquer Approach for Analysis of Extraordinarily Large Data , 2014 .

[14]  Robert Tibshirani,et al.  STANDARDIZATION AND THE GROUP LASSO PENALTY. , 2012, Statistica Sinica.

[15]  Stephen P. Boyd,et al.  Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers , 2011, Found. Trends Mach. Learn..

[16]  S. Wood Fast stable restricted maximum likelihood and marginal likelihood estimation of semiparametric generalized linear models , 2011 .

[17]  M. Drton,et al.  Exact block-wise optimization in group lasso and sparse group lasso for linear regression , 2010, 1010.3320.

[18]  Martin J. Wainwright,et al.  Minimax-Optimal Rates For Sparse Additive Models Over Kernel Classes Via Convex Programming , 2010, J. Mach. Learn. Res..

[19]  Trevor Hastie,et al.  Regularization Paths for Generalized Linear Models via Coordinate Descent. , 2010, Journal of statistical software.

[20]  Martin J. Wainwright,et al.  Sharp Thresholds for High-Dimensional and Noisy Sparsity Recovery Using $\ell _{1}$ -Constrained Quadratic Programming (Lasso) , 2009, IEEE Transactions on Information Theory.

[21]  P. Bühlmann,et al.  The group lasso for logistic regression , 2008 .

[22]  J. Lafferty,et al.  Sparse additive models , 2007, 0711.4555.

[23]  M. Yuan,et al.  Model selection and estimation in regression with grouped variables , 2006 .

[24]  P. Massart,et al.  Adaptive estimation of a quadratic functional by model selection , 2000 .

[25]  H. H. Thodberg,et al.  Optimal minimal neural interpretation of spectra , 1992 .

[26]  T. Hastie,et al.  Generalized Additive Models , 1986 .

[27]  P. Speckman Spline Smoothing and Optimal Rates of Convergence in Nonparametric Regression Models , 1985 .

[28]  C. J. Stone,et al.  Additive Regression and Other Nonparametric Models , 1985 .

[29]  C. Khatri,et al.  Some Results on the Non-Central Multivariate Beta Distribution and Moments of Traces of Two Matrices , 1965 .

[30]  Karl Rohe,et al.  Preconditioning the Lasso for sign consistency , 2015 .

[31]  Min‐ge Xie,et al.  A split-and-conquer approach for analysis of , 2014 .

[32]  S. Weisberg,et al.  Discussion of " Least Angle Regression " by Efron Et Al , 2004 .

[33]  Y. Chikuse Statistics on special manifolds , 2003 .

[34]  R. Tibshirani Regression Shrinkage and Selection via the Lasso , 1996 .

[35]  D. Flinn Orientation Statistics , 1967, Nature.