Towards Exascale Co-design in a Runtime System

  • Thomas Sterling
  • Matthew AndersonEmail author
  • P. Kevin Bohan
  • Maciej Brodowicz
  • Abhishek Kulkarni
  • Bo Zhang
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 8759)


Achieving the performance potential of an Exascale machine depends on realizing both operational efficiency and scalability in high performance computing applications. This requirement has motivated the emergence of several new programming models which emphasize fine and medium grain task parallelism in order to address the aggravating effects of asynchrony at scale. The performance modeling of Exascale systems for these programming models requires the development of fundamentally new approaches due to the demands of both scale and complexity. This work presents a performance modeling case study of the Livermore Unstructured Lagrangian Explicit Shock Hydrodynamics (LULESH) proxy application where the performance modeling approach has been incorporated directly into a runtime system with two modalities of operation: computation and performance modeling simulation. The runtime system exposes performance sensitivies and projects operation to larger scales while also realizing the benefits of removing global barriers and extracting more parallelism from LULESH. Comparisons between the computation and performance modeling simulation results are presented.


Performance Modeling Discrete Event Simulator Execution Model Runtime System Communicate Sequential Process 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.



The authors acknowledge Benjamin Martin, Jackson DeBuhr, Ezra Kissel, Luke D’Alessandro, and Martin Swany for their technical assistance.


  1. 1.
    Hydrodynamics Challenge Problem, Lawrence Livermore National Laboratory. Technical Report LLNL-TR-490254Google Scholar
  2. 2.
    Livermore unstructured lagrangian explicit shock hydrodynamics (lulesh).
  3. 3.
    Alexandrov, A., Ionescu, M.F., Schauser, K.E., Scheiman, C.: LogGP: incorporating long messages into the LogP modelone step closer towards a realistic model for parallel computation. In: Proceedings of the Seventh Annual ACM Symposium on Parallel Algorithms and Architectures, SPAA 1995, pp. 95–105. ACM, New York, NY, USA (1995)Google Scholar
  4. 4.
    Anderson, M., Brodowicz, M., Kulkarni, A., Sterling, T.: Performance modeling of gyrokinetic toroidal simulations for amany-tasking runtime system. In: Jarvis, S.A., Wright, S.A., Hammond, S.D. (eds.) High Performance Computing Systems. Performance Modeling, Benchmarking and Simulation. LNCS, pp. 136–157. Springer, Heidelberg (2014)CrossRefGoogle Scholar
  5. 5.
    Carrington, L., Laurenzano, M., Tiwari, A.: Inferring large-scale computation behavior via trace extrapolation. In: Large-Scale Parallel Processing workshop (IPDPS 2013)Google Scholar
  6. 6.
    Carrington, L., Snavely, A., Gao, X., Wolter, N.: A performance prediction framework for scientific applications. In: ICCS Workshop on Performance Modeling and Analysis (PMA03), pp. 926–935 (2003)Google Scholar
  7. 7.
    Clauss, P.-N., Stillwell, M., Genaud, S., Suter, F., Casanova, H., Quinson, M.: Single node on-line simulation of MPI applications with SMPI. In: Parallel Distributed Processing Symposium (IPDPS), 2011 IEEE International, pp. 664–675 (2011)Google Scholar
  8. 8.
    Culler, D., Karp, R., Patterson, D., Sahay, A., Schauser, K.E., Santos, E., Subramonian, R., von Eicken, T.: LogP: towards a realistic model of parallel computation. In: Proceedings of the Fourth ACM SIGPLAN Symposium on Principles and Practice Of Parallel Programming, PPOPP 1993, pp. 1–12. ACM, New York, NY, USA (1993)Google Scholar
  9. 9.
    Gao, G., Sterling, T., Stevens, R., Hereld, M., Zhu, W.: ParalleX: a study of a new parallel computation model. In: Parallel and Distributed Processing Symposium. IPDPS 2007. IEEE International, pp. 1–6 (2007)Google Scholar
  10. 10.
    Hoefler, T., Schneider, T., Lumsdaine, A.: LogGOPSim - simulating large-scale applications in the LogGOPS Model. In: Proceedings of the 19th ACM International Symposium on High Performance Distributed Computing, pp. 597–604. ACM, June 2010Google Scholar
  11. 11.
    Janssen, C.L., Adalsteinsson, H., Cranford, S., Kenny, J.P., Pinar, A., Evensky, D.A., Mayo, J.: A simulator for large-scale parallel computer architectures. IJDST 1(2), 57–73 (2010)Google Scholar
  12. 12.
    Sottile, M., Dakshinamurthy, A., Hendry, G., Dechev, D.: Semi-automatic extraction of software skeletons for benchmarking large-scale parallel applications. In: Proceedings of the 2013 ACM SIGSIM Conference on Principles of Advanced Discrete Simulation, SIGSIM-PADS 2013, pp. 1–10. ACM, New York, NY, USA (2013)Google Scholar
  13. 13.
    Spafford, K.L., Vetter, J.S.: Aspen: a domain specific language for performance modeling. In: Proceedings of the International Conference on High Performance Computing, Networking, Storage and Analysis, SC 2012, pp. 84:1–84:11. IEEE Computer Society Press, Los Alamitos, CA, USA (2012)Google Scholar
  14. 14.
    Sterling, T.: Towards a ParalleX-enabled Exascale Architecture. Presentation to the DOE Architecture 2 Workshop, 10 August 2011Google Scholar
  15. 15.
    Totoni, E., Bhatele, A., Bohm, E., Jain, N., Mendes, C., Mokos, R., Zheng, G., Kale, L.: Simulation-based performance analysis and tuning for a two-level directly connected system. In: Proceedings of the 17th IEEE International Conference on Parallel and Distributed Systems, December 2011Google Scholar
  16. 16.
    Zheng, G., Wilmarth, T., Lawlor, O.S., Kalé, L.V., Adve, S., Padua, D., Geubelle, P.: Performance modeling and programming environments for Petaflops computers and the Blue Gene machine. In: NSF Next Generation Systems Program Workshop, 18th International Parallel and Distributed Processing Symposium(IPDPS), p. 197. IEEE Press, Santa Fe, New Mexico, April 2004Google Scholar

Copyright information

© Springer International Publishing Switzerland 2015

Authors and Affiliations

  • Thomas Sterling
    • 1
  • Matthew Anderson
    • 1
    Email author
  • P. Kevin Bohan
    • 1
  • Maciej Brodowicz
    • 1
  • Abhishek Kulkarni
    • 1
  • Bo Zhang
    • 1
  1. 1.Center for Research in Extreme Scale Technologies, School of Informatics and ComputingIndiana UniversityBloomingtonUSA

Personalised recommendations