Bayesian Regularized Neural Networks for Small n Big p Data

Artificial neural networks (ANN) mimic the function of the human brain and they have the capability to implement massively parallel computations for mapping, function approximation, classification, and pattern recognition processing. ANN can capture the highly nonlinear associations between inputs (predictors) and target (responses) variables and can adaptively learn the complex functional forms. Like other parametric and nonparametric methods, such as kernel regression and smoothing splines, ANNs can introduce overfitting (in particular with highlydimensional data, such as genome wide association -GWAS-, microarray data etc.) and resulting predictions can be outside the range of the training data. Regularization (shrinkage) in ANN allows bias of parameter estimates towards what are considered to be probable. Most common techniques of regularizations techniques in ANN are the Bayesian regularization (BR) and the early stopping methods. Early stopping is effectively limiting the used weights in the network and thus imposes regularization, effectively lowering the Vapnik-Chervonenkis dimension. In Bayesian regularized ANN (BRANN), the regularization techniques involve imposing certain prior distributions on the model parameters and penalizes large weights in anticipation of achieving smoother mapping.

[1]  Roberto Battiti,et al.  Using mutual information for selecting features in supervised neural net learning , 1994, IEEE Trans. Neural Networks.

[2]  Jouko Lampinen,et al.  Bayesian approach for neural networks--review and case studies , 2001, Neural Networks.

[3]  Stewart Bauck,et al.  Predicting expected progeny difference for marbling score in Angus cattle using artificial neural networks and Bayesian regression models , 2013, Genetics Selection Evolution.

[4]  Christopher M. Bishop,et al.  A Hierarchical Latent Variable Model for Data Visualization , 1998, IEEE Trans. Pattern Anal. Mach. Intell..

[5]  C. L. Philip Chen,et al.  Regularization parameter estimation for feedforward neural networks , 2003 .

[6]  D. M. Titterington,et al.  Bayesian Methods for Neural Networks and Related Models , 2004 .

[7]  Tshilidzi Marwala,et al.  Bayesian Training of Neural Networks Using Genetic Programming , 2006, The 2006 IEEE International Joint Conference on Neural Network Proceedings.

[8]  David J. C. MacKay,et al.  Information Theory, Inference, and Learning Algorithms , 2004, IEEE Transactions on Information Theory.

[9]  Daniel Gianola,et al.  Effect of genotype imputation on genome-enabled prediction of complex traits: an empirical study with mice data , 2014, BMC Genetics.

[10]  K. Weigel,et al.  Prediction of body mass index in mice using dense molecular markers and a regularized neural network. , 2011, Genetics research.

[11]  Randall Matignon,et al.  Data Mining Using SAS Enterprise Miner , 2007 .

[12]  Martin T. Hagan,et al.  Gauss-Newton approximation to Bayesian learning , 1997, Proceedings of International Conference on Neural Networks (ICNN'97).

[13]  Yuhui Qiu,et al.  Novel Approach for Promoting the Generalization Ability of Neural Networks , 2008 .

[14]  Gang Wei,et al.  Regularization Parameter Selection for Faulty Neural Networks , 2009 .

[15]  S. Hyakin,et al.  Neural Networks: A Comprehensive Foundation , 1994 .

[16]  K. Weigel,et al.  Predicting complex quantitative traits with Bayesian neural networks: a case study with Jersey cows and wheat , 2011, BMC Genetics.

[17]  Hayrettin Okut Evaluation of predictive ability of Cholesky factorization of genetic relationship matrix for additive and non-additive genetic effect using Bayesian regularized neural network , 2014 .