MilkyWay-2 supercomputer: system and application
- 504 Downloads
On June 17, 2013, MilkyWay-2 (Tianhe-2) supercomputer was crowned as the fastest supercomputer in the world on the 41th TOP500 list. This paper provides an overview of the MilkyWay-2 project and describes the design of hardware and software systems. The key architecture features of MilkyWay-2 are highlighted, including neo-heterogeneous compute nodes integrating commodity-off-the-shelf processors and accelerators that share similar instruction set architecture, powerful networks that employ proprietary interconnection chips to support the massively parallel message-passing communications, proprietary 16-core processor designed for scientific computing, efficient software stacks that provide high performance file system, emerging programming model for heterogeneous systems, and intelligent system administration. We perform extensive evaluation with wide-ranging applications from LINPACK and Graph500 benchmarks to massively parallel software deployed in the system.
KeywordsMilkyWay-2 supercomputer petaflops computing neo-heterogeneous architecture interconnect network heterogeneous programing model system management benchmark optimization performance evaluation
Unable to display preview. Download preview PDF.
- 2.Zhang H, Wang K, Zhang J, Wu N, Dai Y. A fast and fair shared buffer for high-radix router. Journal of Circuits, Systems, and Computers, 2013Google Scholar
- 3.Kirk D. Nvidia cuda software and GPU parallel computing architecture. In: Proceedings of the 6th International Symposium on Memory Management. 2007, 103–104Google Scholar
- 4.Sherlekar S. Tutorial: Intel many integrated core (MIC) architecture. In: Proceedings of the 18th IEEE International Conference on Parallel and Distributed Systems. 2012, 947Google Scholar
- 5.Gaster B, Howes L, Kaeli D R, Mistry P, Schaa D. Heterogeneous Computing with OpenCL. Morgan Kaufmann Publishers Inc., 2011Google Scholar
- 6.Lee S, Vetter J S. Early evaluation of directive-based GPU programming models for productive exascale computing. In: Proceedings of the 2012 International Conference for High Performance Computing, Networking, Storage and Analysis. 2012, 1–11Google Scholar
- 7.Wienke S, Springer P, Terboven C, Mey D. Openacc: first experiences with real-world applications. In: Proceedings of the 18th International Conference on Parallel Processing. 2012, 859–870Google Scholar
- 8.PGI Accelerator Compilers. Portland Group Inc, 2011Google Scholar
- 10.Dolbeau R, Bihan S, Bodin F. Hmpp: a hybrid multi-core parallel programming environment. In: Proceedings of the 2007 Workshop on General Purpose Processing on Graphics Processing Units. 2007, 1–5Google Scholar
- 11.Checconi F, Petrini F, Willcock J, Lumsdaine A, Choudhury A R, Sabharwal Y. Breaking the speed and scalability barriers for graph exploration on distributed-memory machines. In: Proceedings of the 2012 International Conference for High Performance Computing, Networking, Storage and Analysis. 2012, 1–12CrossRefGoogle Scholar
- 12.Beamer S, Buluç A, Asanovic K, Patterson D. Distributed memory breadth-first search revisited: enabling bottom-up search. In: Proceedings of the 27th IEEE International Symposium on Parallel and Distributed Processing Workshops and PhD Forum. 2013, 1618–1627Google Scholar
- 18.Yang C, Xue W, Fu H, Gan L, Li L, Xu Y, Lu Y, Sun J, Yang G, Zheng W. A peta-scalable CPU-GPU algorithm for global atmospheric simulations. In: Proceedings of the 18th ACM SIGPLAN Symposium on Principles and Practice of Parallel Programming. 2013, 1–12Google Scholar