A Lightweight Message Logging Scheme for Fault Tolerant MPI

  • Inseon Lee
  • Heon Y. Yeom
  • Taesoon Park
  • Hyoungwoo Park
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 3019)


This paper presents a new lightweight logging scheme for MPI to provide fault tolerance. Checkpointing recovery is the most widely used fault tolerance scheme for the distributed systems. However, all the processes should be rolled back and restarted even for a single process failure to preserve consistency. Message logging can be used so that the other processes can proceed unaffected by the failure. However, logging all the messages tends to be prohivitively expensive. We note that the applications programmed using MPI follow certain rules and not all of the messages need to be logged. Our logging scheme is based on this observation and only the absolutely necessary information is logged or piggybacked. As a result, it is possible to greately reduce the logging overhead using our scheme and the experimental results matched well with the expectation.


Fault Tolerance Message Overhead Distribute Shared Memory Shared Memory System Vector Clock 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.


  1. 1.
    Foster, I., Kesselman, C.: The Grid: Blueprint for a Future Computing Infrastructure. Morgan Faufmann Publishers, San Francisco (1999)Google Scholar
  2. 2.
    Stellner, G.: CoCheck: Checkpointing and process migration for MPI. In: Proceedings of the International Parallel Processing Symposium, pp. 526–531 (1996)Google Scholar
  3. 3.
    Agbaria, A., Friedman, R.: Starfish: Fault-tolerant dynamic mpi programs on clusters of workstations. In: Proceedings of IEEE Symposium on High Performance Distributed Computing (1999)Google Scholar
  4. 4.
    Batchu, R., Skjellum, A., Cui, Z., Beddhu, M., Neelamegam, J.P., Dandass, Y., Apte, M.: MPI/FT:architecture and taxonomies for fault-tolerant, message-passing middleware for performance-portable parallel computing. In: 1st International Symposium on Cluster Computing and the Grid (2001)Google Scholar
  5. 5.
    Louca, S., Neophytou, N., Lachanas, A., Evripidou, P.: Portable fault tolerance scheme for MPI. Parallel Processing Letters 10, 371–382 (2000)CrossRefGoogle Scholar
  6. 6.
    Fagg, G.E., Dongarra, J.: FT-MPI: Fault tolerant MPI, supporting dynamic applications in a dynamic world. In: Dongarra, J., Kacsuk, P., Podhorszki, N. (eds.) PVM/MPI 2000. LNCS, vol. 1908, pp. 346–353. Springer, Heidelberg (2000)CrossRefGoogle Scholar
  7. 7.
    Bosilca, G., Bouteiller, A., Cappello, F., Djilali, S., Magniette, G.F., Néri, V., Selikhov, A.: MPICH-V: Toward a scalable fault tolerant MPI for volatile nodes. In: SuperComputing 2002 (2002)Google Scholar
  8. 8.
    Woo, N., Yeom, H.Y., Park, T., Park, H.: MPICH-GF, transparent checkpointing and rollback-recovery for grid-enabled mpi processes. In: Proceedings of the 2nd Workshop on Hardware/Software Support for High Performance Scientific and Engineering Computing (2003)Google Scholar
  9. 9.
    Alvisi, L., Hoppe, B., Marzullo, K.: Nonblocking and orphan-free message logging protocols. In: Symposium on Fault-Tolerant Computing, pp. 145–154 (1993)Google Scholar
  10. 10.
    Keleher, P.J., Cox, A.L., Zwaenepoel, W.: Lazy release consistency for software distributed shared memory. In: The 18th Annual International Symposium on Computer Architecture, pp. 13–21 (1992)Google Scholar
  11. 11.
    Yi, Y., Park, T., Yeom, H.Y.: A causal logging scheme for lazy release consistent distributed shared memory systems. In: Proceedings of the International Conference on Parallel and Distributed Systems, pp. 139–146 (1998)Google Scholar
  12. 12.
    Park, T., Lee, I., Yeom, H.Y.: An efficient causal logging scheme for recoverable distributed shared memory systems. Parallel Computing 28, 1549–1572 (2002)zbMATHCrossRefGoogle Scholar
  13. 13.
    Netzer, R.H.B., Miller, B.P.: Optimal tracing and replay for debugging messagepassing parallel programs. In: Proceedings of Supercomputing 1992, pp. 502–511 (1992)Google Scholar
  14. 14.
    Schlichting, R.D., Schneider, F.B.: Fail-stop processors: An approach to designing fault-tolerant computing systems. ACM Trans. on Computer Systems 1, 222–238 (1983)CrossRefGoogle Scholar
  15. 15.
    Foster, I., Kesselman, C.: The globus project: A status report. In: Proceedings of the Heterogeneous Computing Workshop, pp. 4–18 (1998)Google Scholar
  16. 16.
    NASA Ames Research Center: Nas parallel benchmarks. Technical report (1997),
  17. 17.
    Woo, S., Ohara, M., Torrie, E., Gupta, J.S., The SPLASH-2, A.: programs: Characterization and methodological considerations. In: Proceedings of the 22nd International Symposium on Computer Architectures, pp. 24–36 (1995)Google Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2004

Authors and Affiliations

  • Inseon Lee
    • 1
  • Heon Y. Yeom
    • 1
  • Taesoon Park
    • 2
  • Hyoungwoo Park
    • 3
  1. 1.School of Computer Science and EngineeringSeoul National UniversitySeoulKorea
  2. 2.Department of Computer EngineeringSejong UniversitySeoulKorea
  3. 3.Supercomputing CenterKISTITaejonKorea

Personalised recommendations