Scalable Memory Use in MPI: A Case Study with MPICH2
One of the factors that can limit the scalability of MPI to exascale is the amount of memory consumed by the MPI implementation. In fact, some researchers believe that existing MPI implementations, if used unchanged, will themselves consume a large fraction of the available system memory at exascale. To investigate and address this issue, we undertook a study of the memory consumed by the MPICH2 implementation of MPI, with a focus on identifying parts of the code where the memory consumed per process scales linearly with the total number of processes. We report on the findings of this study and discuss ways to avoid the linear growth in memory consumption. We also describe specific optimizations that we implemented in MPICH2 to avoid this linear growth and present experimental results demonstrating the memory savings achieved and the impact on performance.
KeywordsMemory Consumption Message Size Node Leader Memory Saving Virtual Connection
Unable to display preview. Download preview PDF.
- 1.ASC Sequoia Benchmark Codes: AMG, https://asc.llnl.gov/sequoia/benchmarks/#amg (May 2011)
- 2.Bailey, D., Harris, T., Saphir, W., Van Der Wijngaart, R., Woo, A., Yarrow, M.: The NAS Parallel Benchmarks 2.0. NAS Technical Report NAS-95-020, NASA Ames Research Center, Moffett Field, CA (1995)Google Scholar
- 4.Barbay, J., Navarro, G.: Compressed representations of permutations, and applications. In: Proc. of 26th Int’l Symposium on Theoretical Aspects of Computer Science (STACS), pp. 111–122 (2009)Google Scholar
- 6.Kamal, H., Mirtaheri, S.M., Wagner, A.: Scalability of communicators and groups in MPI. In: Proc. of the ACM International Symposium on High Performance Distributed Computing, HPDC (2010)Google Scholar
- 8.Stevens, R., White, A.: Report of the workshop on architectures and technologies for extreme scale computing (December 2009), http://extremecomputing.labworks.org/hardware/report.stm
- 10.uthash (May 2011), http://uthash.sourceforge.net/