Neural networks-based variationally enhanced sampling

Significance Atomistic-based simulations are one of the most widely used tools in contemporary science. However, in the presence of kinetic bottlenecks, their power is severely curtailed. In order to mitigate this problem, many enhanced sampling techniques have been proposed. Here, we show that by combining a variational approach with deep learning, much progress can be made in extending the scope of such simulations. Our development bridges the fields of enhanced sampling and machine learning and allows us to benefit from the rapidly growing advances in this area. Sampling complex free-energy surfaces is one of the main challenges of modern atomistic simulation methods. The presence of kinetic bottlenecks in such surfaces often renders a direct approach useless. A popular strategy is to identify a small number of key collective variables and to introduce a bias potential that is able to favor their fluctuations in order to accelerate sampling. Here, we propose to use machine-learning techniques in conjunction with the recent variationally enhanced sampling method [O. Valsson, M. Parrinello, Phys. Rev. Lett. 113, 090601 (2014)] in order to determine such potential. This is achieved by expressing the bias as a neural network. The parameters are determined in a variational learning scheme aimed at minimizing an appropriate functional. This required the development of a more efficient minimization technique. The expressivity of neural networks allows representing rapidly varying free-energy surfaces, removes boundary effects artifacts, and allows several collective variables to be handled.

[1]  Fu Jie Huang,et al.  A Tutorial on Energy-Based Learning , 2006 .

[2]  Michele Parrinello,et al.  Multithermal-Multibaric Molecular Simulations from a Variational Principle. , 2018, Physical review letters.

[3]  Michele Parrinello,et al.  Generalized neural-network representation of high-dimensional potential-energy surfaces. , 2007, Physical review letters.

[4]  M. Parrinello,et al.  Well-tempered metadynamics: a smoothly converging and tunable free-energy method. , 2008, Physical review letters.

[5]  Eric Moulines,et al.  Non-strongly-convex smooth stochastic approximation with convergence rate O(1/n) , 2013, NIPS.

[6]  M. Parrinello,et al.  Canonical sampling through velocity rescaling. , 2007, The Journal of chemical physics.

[7]  Massimiliano Bonomi,et al.  PLUMED 2: New feathers for an old bird , 2013, Comput. Phys. Commun..

[8]  M. Parrinello,et al.  Metadynamics with Adaptive Gaussians. , 2012, Journal of chemical theory and computation.

[9]  Wei Chen,et al.  Collective variable discovery and enhanced sampling using autoencoders: Innovations in network architecture and error function design. , 2018, The Journal of chemical physics.

[10]  Michele Parrinello,et al.  Variational approach to enhanced sampling and free energy calculations. , 2014, Physical review letters.

[11]  Hythem Sidky,et al.  Learning free energy landscapes using artificial neural networks. , 2017, The Journal of chemical physics.

[12]  V. Hornak,et al.  Comparison of multiple Amber force fields and development of improved protein backbone parameters , 2006, Proteins.

[13]  Pratyush Tiwary,et al.  Reweighted autoencoded variational Bayes for enhanced sampling (RAVE). , 2018, The Journal of chemical physics.

[14]  Yuji Sugita,et al.  Neural Network and Nearest Neighbor Algorithms for Enhancing Sampling of Molecular Dynamics. , 2017, Journal of chemical theory and computation.

[15]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[16]  Michele Parrinello,et al.  Phase diagrams from single molecular dynamics simulations , 2019 .

[17]  R. Kondor,et al.  Gaussian approximation potentials: the accuracy of quantum mechanics, without the electrons. , 2009, Physical review letters.

[18]  E Weinan,et al.  Reinforced dynamics for enhanced sampling in large atomic and molecular systems. I. Basic Methodology , 2017, The Journal of chemical physics.

[19]  Geoffrey E. Hinton,et al.  Deep Learning , 2015, Nature.

[20]  G. Torrie,et al.  Nonphysical sampling distributions in Monte Carlo free-energy estimation: Umbrella sampling , 1977 .

[21]  Michele Parrinello,et al.  Coarse graining from variationally enhanced sampling applied to the Ginzburg–Landau model , 2017, Proceedings of the National Academy of Sciences.

[22]  Michele Parrinello,et al.  Enhancing Important Fluctuations: Rare Events and Metadynamics from a Conceptual Viewpoint. , 2016, Annual review of physical chemistry.

[23]  N. Metropolis,et al.  Equation of State Calculations by Fast Computing Machines , 1953, Resonance.

[24]  Michele Parrinello,et al.  Bespoke Bias for Obtaining Free Energy Differences within Variationally Enhanced Sampling. , 2016, Journal of chemical theory and computation.

[25]  Noam Bernstein,et al.  Exploration, Sampling, And Reconstruction of Free Energy Surfaces with Gaussian Process Regression. , 2016, Journal of chemical theory and computation.

[26]  Michele Parrinello,et al.  Variationally Optimized Free-Energy Flooding for Rate Calculation. , 2015, Physical review letters.

[27]  J. Behler Perspective: Machine learning potentials for atomistic simulations. , 2016, The Journal of chemical physics.

[28]  A. Laio,et al.  Escaping free-energy minima , 2002, Proceedings of the National Academy of Sciences of the United States of America.

[29]  Jakub Rydzewski,et al.  Promoting transparency and reproducibility in enhanced molecular simulations , 2019, Nature Methods.

[30]  Klaus-Robert Müller,et al.  Efficient BackProp , 2012, Neural Networks: Tricks of the Trade.

[31]  Matthias Troyer,et al.  Solving the quantum many-body problem with artificial neural networks , 2016, Science.

[32]  Mohammad M. Sultan,et al.  Variational encoding of complex dynamics. , 2017, Physical review. E.

[33]  Michele Parrinello,et al.  Well-Tempered Variational Approach to Enhanced Sampling. , 2015, Journal of chemical theory and computation.

[34]  Gerrit Groenhof,et al.  GROMACS: Fast, flexible, and free , 2005, J. Comput. Chem..

[35]  Markus Schöberl,et al.  Predictive Collective Variable Discovery with Deep Bayesian Models , 2018, The Journal of chemical physics.

[36]  Michele Parrinello,et al.  Enhanced, targeted sampling of high-dimensional free-energy landscapes using variationally enhanced sampling, with an application to chignolin , 2016, Proceedings of the National Academy of Sciences.

[37]  Frank Noé,et al.  Time-lagged autoencoders: Deep learning of slow collective variables for molecular kinetics , 2017, The Journal of chemical physics.

[38]  Weber,et al.  Computer simulation of local order in condensed phases of silicon. , 1985, Physical review. B, Condensed matter.

[39]  Michele Parrinello,et al.  Making the Best of a Bad Situation: A Multiscale Approach to Free Energy Calculation. , 2019, Journal of chemical theory and computation.

[40]  Peter V. Coveney,et al.  Large-scale molecular dynamics simulation of DNA: implementation and validation of the AMBER98 force field in LAMMPS , 2004, Philosophical Transactions of the Royal Society of London. Series A: Mathematical, Physical and Engineering Sciences.

[41]  Michele Parrinello,et al.  Silicon Liquid Structure and Crystal Nucleation from Ab Initio Deep Metadynamics. , 2018, Physical review letters.

[42]  Steve Plimpton,et al.  Fast parallel algorithms for short-range molecular dynamics , 1993 .

[43]  M. Klein,et al.  Constant pressure molecular dynamics algorithms , 1994 .

[44]  P. Piaggi,et al.  A variational approach to nucleation simulation. , 2016, Faraday discussions.

[45]  Massimiliano Bonomi,et al.  Efficient Sampling of High-Dimensional Free-Energy Landscapes with Parallel Bias Metadynamics. , 2015, Journal of chemical theory and computation.

[46]  C. Svaneborg Large-scale Atomic/Molecular Massively Parallel Simulator , 2011 .

[47]  E Weinan,et al.  Deep Potential Molecular Dynamics: a scalable model with the accuracy of quantum mechanics , 2017, Physical review letters.

[48]  Michele Parrinello,et al.  Path collective variables without paths , 2018 .