Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2020 Aug 26;11(35):9459-9467.
doi: 10.1039/d0sc03635h.

Molecular latent space simulators

Affiliations

Molecular latent space simulators

Hythem Sidky et al. Chem Sci. .

Abstract

Small integration time steps limit molecular dynamics (MD) simulations to millisecond time scales. Markov state models (MSMs) and equation-free approaches learn low-dimensional kinetic models from MD simulation data by performing configurational or dynamical coarse-graining of the state space. The learned kinetic models enable the efficient generation of dynamical trajectories over vastly longer time scales than are accessible by MD, but the discretization of configurational space and/or absence of a means to reconstruct molecular configurations precludes the generation of continuous atomistic molecular trajectories. We propose latent space simulators (LSS) to learn kinetic models for continuous atomistic simulation trajectories by training three deep learning networks to (i) learn the slow collective variables of the molecular system, (ii) propagate the system dynamics within this slow latent space, and (iii) generatively reconstruct molecular configurations. We demonstrate the approach in an application to Trp-cage miniprotein to produce novel ultra-long synthetic folding trajectories that accurately reproduce atomistic molecular structure, thermodynamics, and kinetics at six orders of magnitude lower cost than MD. The dramatically lower cost of trajectory generation enables greatly improved sampling and greatly reduced statistical uncertainties in estimated thermodynamic averages and kinetic rates.

PubMed Disclaimer

Conflict of interest statement

A. L. F. is a consultant of Evozyne and a co-author of US Provisional Patents 62/853,919 and 62/900,420 and International Patent Application PCT/US2020/035206.

Figures

Fig. 1
Fig. 1. Schematic diagram of the latent space simulator (LSS) comprising three back-to-back deep neural networks. A state-free reversible VAMPnet (SRV) learns an encoding E of molecular configurations into a latent space spanned by the leading eigenfunctions of the transfer operator (eqn (1)). A mixture density network (MDN) learns a propagator P to sample transition probabilities pτ(ψt+τ|ψt) within the latent space. A conditional Wasserstein GAN (cWGAN) learns a generative decoding D of molecular configurations conditioned on the latent space coordinates. The trained LSS is used to generate ultra-long synthetic trajectories by projecting the initial configuration into the latent space using the SRV, sampling from the MDN to generate long latent space trajectories, and decoding to molecular configurations using the cWGAN.
Fig. 2
Fig. 2. Validation of the LSS in a 1D four-well potential. The MDN propagator predicts (a) a stationary distribution, (b) kinetic transitions, and (c and d) transition densities in excellent accord with analytical and Brownian dynamics results.
Fig. 3
Fig. 3. Free energy profiles for the MD and LSS trajectories projected into the slowest latent space coordinate ψ1. Shaded backgrounds represent standard errors estimated by five-fold block averaging. The profiles agree within a 0.91kBT root mean squared error. Ten representative structures from the MD and LSS ensembles are sampled from the folded (ψ1 ≈ 0), unfolded (ψ1 ≈ 0.9), and metastable (ψ1 ≈ 0.45) regions.
Fig. 4
Fig. 4. Free energy profiles of the MD and LSS trajectories projected into the leading three MD TICA coordinates.

Similar articles

Cited by

References

    1. Frenkel D. and Smit B., Understanding Molecular Simulation: From algorithms to applications, Academic Press, San Diego, 2002
    1. Stone J. E. Hardy D. J. Ufimtsev I. S. Schulten K. J. Mol. Graphics Modell. 2010;29:116–125. doi: 10.1016/j.jmgm.2010.06.010. - DOI - PMC - PubMed
    1. Shaw D. E., Grossman J. P., Bank J. A., Batson B., Butts J. A., Chao J. C., Deneroff M. M., Dror R. O., Even A., Fenton C. H., Forte A., Gagliardo J., Gill G., Greskamp B., Ho C. R., Ierardi D. J., Iserovich L., Kuskin J. S., Larson R. H., Layman T., Lee L. S., Lerer A. K., Li C., Killebrew D., Mackenzie K. M., Mok S. Y. H., Moraes M. A., Mueller R., Nociolo L. J., Peticolas J. L., Quan T., Ramot D., Salmon J. K., Scarpazza D. P., Ben Schafer U., Siddique N., Snyder C. W., Spengler J., Tang P. T. P., Theobald M., Toma H., Towles B., Vitale B., Wang S. C. and Young C., SC'14: Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis, 2014, pp. 41–53
    1. Phillips J. C. Braun R. Wang W. Gumbart J. Tajkhorshid E. Villa E. Chipot C. Skeel R. D. Kale L. Schulten K. J. Comput. Chem. 2005;26:1781–1802. doi: 10.1002/jcc.20289. - DOI - PMC - PubMed
    1. Chow E., Rendleman C. A., Bowers K. J., Dror R. O., Hughes D. H., Gullingsrud J., Sacerdoti F. D. and Shaw D. E., Desmond performance on a cluster of multicore processors (DESRES/TR–2008-01), DE Shaw Research Technical Report, 2008