Using Non-blocking I/O Operations in High Performance Computing to Reduce Execution Times

  • David Buettner
  • Julian Kunkel
  • Thomas Ludwig
Part of the Lecture Notes in Computer Science book series (LNCS, volume 5759)


As supercomputers become faster, the I/O part of applications can become a real problem in regard to overall execution times. System administrators and developers of hardware or software components reduce execution times by creating new and optimized parts for the supercomputers. While this helps a lot in the struggle to minimize I/O times, adjustment of the execution environment is not the only option to improve overall application behavior.

In this paper we examine if the application programmer can also contribute by making use of non-blocking I/O operations. After an analysis of non-blocking I/O operations and their potential for shortening execution times we present a benchmark which was created and run in order to see if the theoretical promises also hold in practice.


MPI-IO MPICH2 PVFS2 non-blocking I/O PIOviz benchmarking 


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.


  1. 1.
    Message Passing Interface Forum: MPI: A Message-Passing Interface Standard (May 1994),
  2. 2.
    Message Passing Interface Forum: MPI-2: Extensions to the Message-Passing Interface (June 1997),
  3. 3.
  4. 4.
    Zaki, O., Lusk, E., Gropp, W., Swider, D.: Toward Scalable Performance Visualization with Jumpshot. Int. J. High Perform. Comput. Appl. 13 (1999)Google Scholar
  5. 5.
    The Parallel Virtual File System – Version 2,
  6. 6.
    The PVFS2 Development Team: PVFS2 Internal Documentation included in the source code package (2006)Google Scholar
  7. 7.
    Thakur, R., Gropp, W.D.: Test Suite for Evaluating Performance of MPI Implementations That Support MPI_THREAD_MULTIPLE. In: Cappello, F., Herault, T., Dongarra, J. (eds.) PVM/MPI 2007. LNCS, vol. 4757, pp. 46–55. Springer, Heidelberg (2007)CrossRefGoogle Scholar
  8. 8. (April 2009),
  9. 9.
    Kunkel, J.: Performance Analysis of the PVFS2 Persistency Layer, Bachelor’s Thesis, Ruprecht-Karls-Universität Heidelberg, Germany (March 2006),
  10. 10.
    Büttner, D.: Benchmarking of Non-Blocking Input/Output on Compute Clusters, Bachelor’s Thesis, Ruprecht-Karls-Universität Heidelberg, Germany (April 2007),
  11. 11.
    Latham, R., Miller, N., Ross, R., Carns, P.: A Next-Generation Parallel File System for Linux Clusters. LinuxWorld Magazine, 56–59Google Scholar
  12. 12.
    Gropp, W., Lusk, E., Sterling, T.: Beowulf Cluster Computing with Linux, 2nd edn. MIT Press, Cambridge (2003)Google Scholar
  13. 13.
    Stallings, W.: Betriebssysteme, 4th edn. Pearson Studium, London (2003)Google Scholar
  14. 14.
    Rabenseifner, R., Koniges, A.E.: Effective Communication and File-I/O Bandwidth Benchmarks. In: Cotronis, Y., Dongarra, J. (eds.) PVM/MPI 2001. LNCS, vol. 2131, p. 24. Springer, Heidelberg (2001)CrossRefGoogle Scholar
  15. 15.
    Krietemayer, M., Versick, D., Tavangarian, D.: THE PRIOmark PARALLEL I/O-BENCHMARK. In: International Conference on Parallel and Distributed Computing and Networks, Insbruck, Austria (February 2005)Google Scholar
  16. 16.
    Lawry, W., Wilson, C., Maccabe, A.B., Brightwell, R.: COMB: A Portable Benchmark Suite for Assessing MPI Overlap. In: Proceedings of the 4th IEEE International Conference on Cluster Computing (September 2002)Google Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2009

Authors and Affiliations

  • David Buettner
    • 1
  • Julian Kunkel
    • 1
  • Thomas Ludwig
    • 1
  1. 1.Ruprecht-Karls-Universität HeidelbergHeidelbergGermany

Personalised recommendations