Fast and Accurate Uncertainty Estimation in Chemical Machine Learning.

We present a scheme to obtain an inexpensive and reliable estimate of the uncertainty associated with the predictions of a machine-learning model of atomic and molecular properties. The scheme is based on resampling, with multiple models being generated based on subsampling of the same training data. The accuracy of the uncertainty prediction can be benchmarked by maximum likelihood estimation, which can also be used to correct for correlations between resampled models and to improve the performance of the uncertainty estimation by a cross-validation procedure. In the case of sparse Gaussian Process Regression models, this resampled estimator can be evaluated at negligible cost. We demonstrate the reliability of these estimates for the prediction of molecular and materials energetics and for the estimation of nuclear chemical shieldings in molecular crystals. Extension to estimate the uncertainty in energy differences, forces, or other correlated predictions is straightforward. This method can be easily applied to other machine-learning schemes and will be beneficial to make data-driven predictions more reliable and to facilitate training-set optimization and active-learning strategies.

[1]  Shun-ichi Amari,et al.  Statistical Theory of Learning Curves under Entropic Loss Criterion , 1993, Neural Computation.

[2]  Jörg Behler,et al.  Automatic selection of atomic fingerprints and reference configurations for machine-learning potentials. , 2018, The Journal of chemical physics.

[3]  G. Schwarz Estimating the Dimension of a Model , 1978 .

[4]  Zhenwei Li,et al.  Molecular dynamics with on-the-fly machine learning of quantum-mechanical forces. , 2015, Physical review letters.

[5]  Volker Tresp,et al.  A Bayesian Committee Machine , 2000, Neural Computation.

[6]  Gábor Csányi,et al.  Gaussian approximation potentials: A brief tutorial introduction , 2015, 1502.01366.

[7]  J Behler,et al.  Representing potential energy surfaces by high-dimensional neural network potentials , 2014, Journal of physics. Condensed matter : an Institute of Physics journal.

[8]  Klaus Schulten,et al.  A Numerical Study on Learning Curves in Stochastic Multilayer Feedforward Networks , 1996, Neural Computation.

[9]  Carlo Cavazzoni,et al.  Advanced capabilities for materials modelling with Quantum ESPRESSO. , 2017, Journal of physics. Condensed matter : an Institute of Physics journal.

[10]  D. Chadi,et al.  Special points for Brillouin-zone integrations , 1977 .

[11]  Andrea Grisafi,et al.  Symmetry-Adapted Machine Learning for Tensorial Properties of Atomistic Systems. , 2017, Physical review letters.

[12]  Francesco Mauri,et al.  Calculation of NMR chemical shifts for extended systems using ultrasoft pseudopotentials , 2007 .

[13]  P. Massart,et al.  Gaussian model selection , 2001 .

[14]  H. Akaike A new look at the statistical model identification , 1974 .

[15]  Frederick R. Manby,et al.  Machine-learning approach for one- and two-body corrections to density functional theory: Applications to molecular and condensed water , 2013 .

[16]  A. Weigend,et al.  Estimating the mean and variance of the target probability distribution , 1994, Proceedings of 1994 IEEE International Conference on Neural Networks (ICNN'94).

[17]  Mikhail Langovoy,et al.  Data-driven goodness-of-fit tests , 2007, 0708.0169.

[18]  Yanli Wang,et al.  Quantum ESPRESSO: a modular and open-source software project for quantum simulations of materials , 2009 .

[19]  Carlo Cavazzoni,et al.  Enhancement of DFT-calculations at petascale: Nuclear Magnetic Resonance, Hybrid Density Functional Theory and Car-Parrinello calculations , 2013, Comput. Phys. Commun..

[20]  R. Kondor,et al.  Gaussian approximation potentials: the accuracy of quantum mechanics, without the electrons. , 2009, Physical review letters.

[21]  Robert Tibshirani,et al.  A Comparison of Some Error Estimates for Neural Network Models , 1996, Neural Computation.

[22]  David Hinkley,et al.  Bootstrap Methods: Another Look at the Jackknife , 2008 .

[23]  Gábor Csányi,et al.  Comparing molecules and solids across structural and alchemical space. , 2015, Physical chemistry chemical physics : PCCP.

[24]  Alok Choudhary,et al.  A General-Purpose Machine Learning Framework for Predicting Properties of Inorganic Materials , 2016 .

[25]  Michele Parrinello,et al.  Generalized neural-network representation of high-dimensional potential-energy surfaces. , 2007, Physical review letters.

[26]  Michele Parrinello,et al.  Demonstrating the Transferability and the Descriptive Power of Sketch-Map. , 2013, Journal of chemical theory and computation.

[27]  Aldo Glielmo,et al.  Efficient nonparametric n -body force fields from machine learning , 2018, 1801.04823.

[28]  Francesco Mauri,et al.  All-electron magnetic response with pseudopotentials: NMR chemical shifts , 2001 .

[29]  Ryo Kobayashi,et al.  Neural network potential for Al-Mg-Si alloys , 2017 .

[30]  I. Bruno,et al.  Cambridge Structural Database , 2002 .

[31]  Michael Wolf,et al.  Weak convergence of dependent empirical measures with application to subsampling in function spaces , 1999 .

[32]  Burke,et al.  Generalized Gradient Approximation Made Simple. , 1996, Physical review letters.

[33]  Pavlo O. Dral,et al.  Quantum chemistry structures and properties of 134 kilo molecules , 2014, Scientific Data.

[34]  Alireza Khorshidi,et al.  Addressing uncertainty in atomistic machine learning. , 2017, Physical chemistry chemical physics : PCCP.

[35]  Michael I. Jordan,et al.  Regression with input-dependent noise: A Gaussian process treatment , 1998 .

[36]  Michalis K. Titsias,et al.  Variational Learning of Inducing Variables in Sparse Gaussian Processes , 2009, AISTATS.

[37]  Joseph P. Romano,et al.  Large Sample Confidence Regions Based on Subsamples under Minimal Assumptions , 1994 .

[38]  Carl E. Rasmussen,et al.  A Unifying View of Sparse Approximate Gaussian Process Regression , 2005, J. Mach. Learn. Res..

[39]  Neil D. Lawrence,et al.  Fast Forward Selection to Speed Up Sparse Gaussian Process Regression , 2003, AISTATS.

[40]  I. A. Ibragimov,et al.  Gaussian Random Processes. Part 2 , 1977 .

[41]  J. Fox,et al.  Applied Regression Analysis and Generalized Linear Models , 2008 .

[42]  Olexandr Isayev,et al.  ANI-1, A data set of 20 million calculated off-equilibrium conformations for organic molecules , 2017, Scientific Data.

[43]  R. Kondor,et al.  On representing chemical environments , 2012, 1209.3140.

[44]  Christopher M. Bishop,et al.  Regression with Input-Dependent Noise: A Bayesian Treatment , 1996, NIPS.

[45]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[46]  M. Frisch,et al.  Ab Initio Calculation of Vibrational Absorption and Circular Dichroism Spectra Using Density Functional Force Fields , 1994 .

[47]  Stefano de Gironcoli,et al.  Projector augmented-wave and all-electron calculations across the periodic table: a comparison of structural and energetic properties , 2014, 1404.3015.

[48]  Noam Bernstein,et al.  Machine learning unifies the modeling of materials and molecules , 2017, Science Advances.

[49]  Tom Heskes,et al.  Practical Confidence and Prediction Intervals , 1996, NIPS.

[50]  J. Behler Perspective: Machine learning potentials for atomistic simulations. , 2016, The Journal of chemical physics.