Encyclopedia of Applied and Computational Mathematics

2015 Edition
| Editors: Björn Engquist

Large-Scale Computing for Molecular Dynamics Simulation

  • Aiichiro Nakano
  • Rajiv K. Kalia
  • Ken-ichi Nomura
  • Priya Vashishta
Reference work entry
DOI: https://doi.org/10.1007/978-3-540-70529-1_279

Mathematics Subject Classification

65Y05; 70F10; 81V55

Synonyms

High performance computing for atomistic simulation

Short Definition

Large-scale computing for molecular dynamics simulation combines advanced computing hardware and efficient algorithms for atomistic simulation to study material properties and processes encompassing large spatiotemporal scales.

Description

Material properties and processes are often dictated by complex dynamics of a large number of atoms. To understand atomistic mechanisms that govern macroscopic material behavior, large-scale molecular dynamics (MD) simulations [1] involving multibillion atoms are performed on parallel supercomputers consisting of over 105 processors [2]. In addition, special-purpose computers are built to enable long-time MD simulations extending millisecond time scales (or 1012 time steps using a time discretization unit of 10−15 s) [3] (for extending the time scale, see also  Transition Pathways, Rare Events and Related Questions)....

This is a preview of subscription content, log in to check access.

References

  1. 1.
    Rapaport, D.C.: The art of molecular dynamics simulation. 2nd edn. Cambridge University Press, Cambridge (2004)CrossRefMATHGoogle Scholar
  2. 2.
    Nakano, A., et al.: De novo ultrascale atomistic simulations on high-end parallel supercomputers. Int. J. High Perform. Comput. Appl. 22, 113 (2008)CrossRefGoogle Scholar
  3. 3.
    Shaw, D.E., et al.: Anton, a special-purpose machine for molecular dynamics simulation. Commun. ACM. 51, 91 (2008)CrossRefGoogle Scholar
  4. 4.
    Greengard, L., Rokhlin, V.: A fast algorithm for particle simulations. J. Comput. Phys. 73, 325 (1987)MathSciNetCrossRefMATHGoogle Scholar
  5. 5.
    Ogata S., et al.: Scalable and portable implementation of the fast multipole method on parallel computers. Comput. Phys. Commun. 153, 445 (2003)CrossRefGoogle Scholar
  6. 6.
    Martyna, G.J., et al.: Explicit reversible integrators for extended systems dynamics. Mol. Phys. 87, 1117 (1996)CrossRefGoogle Scholar
  7. 7.
    Schlick, T., et al.: Algorithmic challenges in computational molecular biophysics. J. Comput. Phys. 151, 9 (1999)MathSciNetCrossRefMATHGoogle Scholar
  8. 8.
    Nakano, A.: Parallel multilevel preconditioned conjugate-gradient approach to variable-charge molecular dynamics. Comput. Phys. Commun. 104, 59 (1997)CrossRefGoogle Scholar
  9. 9.
    Car, R., Parrinello, M.: Unified approach for molecular dynamics and density functional theory. Phys. Rev. Lett. 55, 2471 (1985)CrossRefGoogle Scholar
  10. 10.
    Goedecker, S.: Linear scaling electronic structure methods. Rev. Mod. Phys. 71, 1085 (1999)CrossRefGoogle Scholar
  11. 11.
    Bowler, D.R., et al.: Introductory remarks: Linear scaling methods - Preface. J. Phys. Condens. Matter 20, 290301 (2008)CrossRefGoogle Scholar
  12. 12.
    Kohn, W.: Density functional and density matrix method scaling linearly with the number of atoms. Phys. Rev. Lett. 76, 3168 (1996)CrossRefGoogle Scholar
  13. 13.
    Yang, W.: Direct calculation of electron-density in density-functional theory. Phys. Rev. Lett. 66, 1438 (1991)CrossRefGoogle Scholar
  14. 14.
    Goedecker, S., Scuseria, G.E.: Linear scaling electronic structure methods in chemistry and physics. Comput. Sci. Eng. 5, 14 (2003)CrossRefGoogle Scholar
  15. 15.
    Kitaura, K., et al.: Fragment molecular orbital method: an approximate computational method for large molecules. Chem. Phys. Lett. 313, 701 (1999)CrossRefGoogle Scholar
  16. 16.
    Grama, A., et al.: Introduction to parallel computing. 2nd edn. Addison Wesley, Harlow (2003)Google Scholar
  17. 17.
    Asanovic, K., et al.: The landscape of parallel computing research: A view from Berkeley. University of California, Berkeley (2006)Google Scholar
  18. 18.
    Plimpton, S.J.: Fast parallel algorithms for short-range molecular dynamics. J. Comput. Phys. 117, 1 (1995)CrossRefMATHGoogle Scholar
  19. 19.
    Kale, L., et al.: NAMD2: greater scalability for parallel molecular dynamics. J. Comput. Phys. 151, 283 (1999)MathSciNetCrossRefMATHGoogle Scholar
  20. 20.
    Shaw, D.E.: A fast, scalable method for the parallel evaluation of distance-limited pairwise particle interactions. J. Comput. Chem. 26, 1318 (2005)CrossRefGoogle Scholar
  21. 21.
    Devine, K.D., et al.: New challenges in dynamic load balancing. Appl. Num. Math. 52, 133 (2005)MathSciNetCrossRefMATHGoogle Scholar
  22. 22.
    Nomura, K., et al.: A metascalable computing framework for large spatiotemporal-scale atomistic simulations. Proceedings of International Parallel and Distributed Processing Symposium IPDPS 2009, IEEE, Rome (2009)Google Scholar
  23. 23.
    Gropp, W., Lusk, E., Skjellum, A.: Using MPI. 2nd edn. MIT, Cambridge (1999)MATHGoogle Scholar
  24. 24.
    Chapman, B., Jost, G., van der Pas, R.: Using OpenMP. MIT, Cambrige (2007)Google Scholar
  25. 25.
    Phillips, J.C., Stone, J.E.: Probing biomolecular machines with graphics processors. Commun. ACM. 52, 34 (2009)CrossRefGoogle Scholar
  26. 26.
    Peng, L., et al.: Exploiting hierarchical parallelisms for molecular dynamics simulation on multicore clusters. J. Supercomput. 57, 20 (2011)CrossRefGoogle Scholar
  27. 27.
    Glosli, J.N., et al.: Extending stability beyond CPU millennium: a micron-scale atomistic simulation of Kelvin-Helmholtz instability. Proceedings of Supercomputing (SC07), ACM, New York (2007)Google Scholar
  28. 28.
    Germann, T.C., Kadau, K.: Trillion-atom molecular dynamics becomes a reality. Int. J. Mod. Phys. C. 19, 1315 (2008)CrossRefMATHGoogle Scholar
  29. 29.
    Chen, H.P., et al.: Embrittlement of metal by solute segregation-induced amorphization. Phys. Rev. Lett. 104, 155502 (2010)CrossRefGoogle Scholar
  30. 30.
    Gygi, F., et al.: Large-scale first-principles molecular dynamics simulations on the BlueGene/L platform using the Qbox code. Proceedings of Supercomputing 2005 (SC05), ACM, Washington, DC (2005)Google Scholar
  31. 31.
    Shimojo, F., et al.: Enhanced reactivity of nanoenergetic materials: A first-principles molecular dynamics study based on divide-and-conquer density functional theory. Appl. Phys. Lett. 95, 043114 (2009)CrossRefGoogle Scholar
  32. 32.
    Ikeshoji, T., et al.: Fast-ionic conductivity of Li+ in LiBH4. Phys. Rev. B. 83, 144301 (2011)CrossRefGoogle Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2015

Authors and Affiliations

  • Aiichiro Nakano
    • 1
  • Rajiv K. Kalia
    • 2
  • Ken-ichi Nomura
    • 3
  • Priya Vashishta
    • 4
  1. 1.Department of Computer Science, Department of Physics and Astronomy, and Department of Chemical Engineering and Materials ScienceUniversity of Southern CaliforniaLos AngelesUSA
  2. 2.Department of Computer Science, Department of Physics and Astronomy, and Department of Chemical Engineering and Materials ScienceUniversity of Southern CaliforniaLos AngelesUSA
  3. 3.Department of Computer Science, Department of Physics and Astronomy, and Department of Chemical Engineering and Materials ScienceUniversity of Southern CaliforniaLos AngelesUSA
  4. 4.Department of Computer Science, Department of Physics and Astronomy, and Department of Chemical Engineering and Materials ScienceUniversity of Southern CaliforniaLos AngelesUSA