[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.1145/2335755.2335787acmotherconferencesArticle/Chapter ViewAbstractPublication PagesxsedeConference Proceedingsconference-collections
research-article

Running many molecular dynamics simulations on many supercomputers

Published: 16 July 2012 Publication History

Abstract

The challenges facing biomolecular simulations are many-fold. In addition to long time simulations of a single large system, an important challenge is the ability to run a large number of identical copies (ensembles) of the same system. Ensemble-based simulations are important for effective sampling. Due to the low-level of coupling between them, ensemble-based simulations are good candidates to utilize distributed cyberinfrastructure. The problem for the practitioner is thus effectively marshaling thousands if not millions of high-performance simulations on distributed cyberinfrastructure. Here we assess the ability of an interoperable and extensible pilot-job tool (BigJob), to support high-throughput simulations of high-performance molecular dynamics simulations across distributed supercomputing infrastructure. BigJob provides the capability to run hundreds or thousands of MPI ensembles concurrently. This is advantageous on large machines because it reduces the number of submissions to the queue, thereby reducing the overall waiting time in the queue. The wait time problem is further complicated by scheduling policies on some large XSEDE machines that prioritize large job requests over very small or single core job requests. Using a nucleosome positioning problem as an exemplar, we demonstrate how we have addressed this challenge on the TeraGrid/XSEDE. Specifically, we compute 336 independent trajectories of 20 ns each. Each trajectory is divided into twenty 1 ns long simulation tasks. A single task requires ≈ 42 MB of input, 9 hours of compute time on 32 cores, and generates 3.8 GB of data. In total we have 6,720 tasks (6.7 μs) and approximately 25 TB to manage. There is natural task-level concurrency, as these 6,720 tasks can be executed with 336-way task concurrency. This project requires approximately 2 million hours of CPU time and could be completed in just over 1 month on a dedicated supercomputer containing 3,000 cores. In practice, even such a modest supercomputer is a shared resource and our experience suggests that a simple scheme to automatically batch queue the tasks, might require several years to complete the project. In order to reduce the total time-to-completion, we need to scale-up, out and across various resources. Our approach is to aggregate many ensemble members into pilot-jobs, distribute pilot-jobs over multiple compute resources concurrently, and dynamically assign tasks across the available resources. Here we report the computational methodology employed in our study and refrain from analyzing the biological aspects of the simulations.

References

[1]
SAGA-BigJob. https://github.com/saga-project/BigJob/wiki.
[2]
M. Balman, I. Suslu, and T. Kosar. Distributed data management with petashare. In Proceedings of the 15th ACM Mardi Gras conference: From lightweight mash-ups to lambda grids: Understanding the spectrum of distributed computing requirements, applications, tools, infrastructures, interoperability, and the incremental adoption of key capabilities, MG '08, pages 19:1--19:1, New York, NY, USA, 2008. ACM.
[3]
C. Chipot and J. Henin. Exploring the free-energy landscape of a short peptide using an average force. The Journal of Chemical Physics, 123(24):244906, 2005.
[4]
Y. Cui, K. B. Olsen, T. H. Jordan, K. Lee, J. Zhou, P. Small, D. Roten, G. Ely, D. K. Panda, A. Chourasia, J. Levesque, S. M. Day, and P. Maechling. Scalable earthquake simulation on petascale supercomputers. In Proceedings of the 2010 ACM/IEEE International Conference for High Performance Computing, Networking, Storage and Analysis, SC '10, pages 1--20, Washington, DC, USA, 2010. IEEE Computer Society.
[5]
E. Darve and A. Pohorille. Calculating free energies using average force. The Journal of Chemical Physics, 115(20):9169--9183, 2001.
[6]
D. Earl and M. Deem. Parallel tempering: Theory, applications, and new perspectives. PHYSICAL CHEMISTRY CHEMICAL PHYSICS, 7(23):3910--3916, 2005.
[7]
F. Gagliardi, B. Jones, F. Grey, M.-E. Bégin, and M. Heikkurinen. Building an infrastructure for scientific grid computing: status and goals of the egee project. Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences, 363(1833):1729--1742, 2005.
[8]
M. Gower, J. Cohen, J. Phillips, R. Kufrin, and K. Schulten. Managing biomolecular simulations in a grid environment with namd-g. In Proceedings of the 2006 TeraGrid Conference, 2006., 2006.
[9]
U. Hansmann. Parallel tempering algorithm for conformational studies of biological molecules. CHEMICAL PHYSICS LETTERS, 281(1-3):140--150, DEC 19 1997.
[10]
N. Jacq, V. Breton, H.-Y. Chen, L.-Y. Ho, M. Hofmann, V. Kasam, H.-C. Lee, Y. Legré, S. C. Lin, A. Maaß, E. Medernach, I. Merelli, L. Milanesi, G. Rastelli, M. Reichstadt, J. Salzemann, H. Schwichtenberg, Y.-T. Wu, and M. Zimmermann. Virtual screening on large scale grids. Parallel Computing, 33(4-5):289--301, 2007. Large Scale Grids.
[11]
Y. E. Khamra and S. Jha. Title: Developing Autonomic Distributed Scientific Applications: A Case Study From History Matching Using Ensemble Kalman-Filters. In GMAC '09: Proceedings of the 6th international conference industry session on Grids meets autonomic computing, pages 19--28, New York, NY, USA, 2009. IEEE.
[12]
S.-H. Ko, N. Kim, J. Kim, A. Thota, and S. Jha. Efficient runtime environment for coupled multi-physics simulations: Dynamic resource allocation and load-balancing. In Proceedings of the 2010 10th IEEE/ACM International Conference on Cluster, Cloud and Grid Computing, CCGRID '10, pages 349--358, Washington, DC, USA, 2010. IEEE Computer Society.
[13]
S. M. Larson, C. D. Snow, M. Shirts, V. S. P, and V. S. Pande. Folding@home and genome@home: Using distributed computing to tackle previously intractable problems in computational biology.
[14]
H.-C. Lee, J. Salzemann, N. Jacq, H.-Y. Chen, L.-Y. Ho, I. Merelli, L. Milanesi, V. Breton, S. Lin, and Y.-T. Wu. Grid-enabled high-throughput in silico screening against influenza a neuraminidase. NanoBioscience, IEEE Transactions on, 5(4):288--295, dec. 2006.
[15]
A. Luckow, L. Lacinski, and S. Jha. SAGA BigJob: An Extensible and Interoperable Pilot-Job Abstraction for Distributed Applications and Systems. In The 10th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing, pages 135--144, 2010.
[16]
J. Phillips, R. Braun, W. Wang, J. Gumbart, E. Tajkhorshid, E. Villa, C. Chipot, R. Skeel, L. Kale, and K. Schulten. Scalable molecular dynamics with NAMD. Journal of Computational Chemistry, 26:1781--1802, 2005.
[17]
J. Roy and C. A. Laughton. Long-timescale molecular-dynamics simulations of the major urinary protein provide atomistic interpretations of the unusual thermodynamicsÂňâĂăof ligand binding. Biophysical Journal, 99(1):218--226, 2010.
[18]
Y. Sugita and Y. Okamoto. Replica-exchange molecular dynamics method for protein folding. CHEMICAL PHYSICS LETTERS, 314(1-2):141--151, NOV 26 1999.
[19]
A. Thota, A. Luckow, and S. Jha. Efficient Large-Scale Replica-Exchange Simulations on Production Infrastructure.

Cited By

View all

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Other conferences
XSEDE '12: Proceedings of the 1st Conference of the Extreme Science and Engineering Discovery Environment: Bridging from the eXtreme to the campus and beyond
July 2012
423 pages
ISBN:9781450316026
DOI:10.1145/2335755
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 16 July 2012

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. HPC
  2. MD
  3. NAMD
  4. XSEDE resources
  5. distributed computing
  6. experience
  7. large scale
  8. technology

Qualifiers

  • Research-article

Funding Sources

Conference

XSEDE12

Acceptance Rates

Overall Acceptance Rate 129 of 190 submissions, 68%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)11
  • Downloads (Last 6 weeks)2
Reflects downloads up to 13 Dec 2024

Other Metrics

Citations

Cited By

View all
  • (2019)TMB Library of Nucleosome SimulationsJournal of Chemical Information and Modeling10.1021/acs.jcim.9b0025259:10(4289-4299)Online publication date: 6-Sep-2019
  • (2016)Distinct Roles of Histone H3 and H2A Tails in Nucleosome StabilityScientific Reports10.1038/srep314376:1Online publication date: 16-Aug-2016
  • (2014)Progress in molecular modelling of DNA materialsMolecular Simulation10.1080/08927022.2014.91379240:10-11(777-783)Online publication date: 23-May-2014
  • (2014)Making campus bridging work for researchersConcurrency and Computation: Practice & Experience10.1002/cpe.326626:13(2141-2148)Online publication date: 10-Sep-2014
  • (2013)Scalable online comparative genomics of mononucleosomesProceedings of the Conference on Extreme Science and Engineering Discovery Environment: Gateway to Discovery10.1145/2484762.2484819(1-8)Online publication date: 22-Jul-2013
  • (2013)Atomistic simulations of nucleosomesWIREs Computational Molecular Science10.1002/wcms.11393:4(378-392)Online publication date: 31-Jan-2013
  • (2012)Distributed Application Runtime Environment (DARE)Journal of Grid Computing10.1007/s10723-012-9244-110:4(647-664)Online publication date: 1-Dec-2012

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media