Skip to main content

Wait-Free Message Passing Protocol for Non-coherent Shared Memory Architectures

  • Conference paper
Recent Advances in the Message Passing Interface (EuroMPI 2012)

Part of the book series: Lecture Notes in Computer Science ((LNPSE,volume 7490))

Included in the following conference series:

  • 1427 Accesses

Abstract

The number of cores in future CPUs is expected to increase steadily. Balanced CPU designs scale hardware cache coherency functionality according to the number of cores, in order to minimize bottlenecks in parallel applications. An alternative approach is to do away with hardware coherence entirely; the Single-chip Cloud Computer (SCC), a 48 core experimental processor from Intel labs, does exactly that. A wait-free protocol for message passing on non-coherent buffers was introduced with the RCKMPI library, in order to support MPI on the SCC. In this work, the message passing performance of the protocol is modeled. Additionally, a port for symmetric multi-processors is introduced and used for comparison with MPICH2-Nemesis and Open MPI. Performance is analyzed based on statistics collected on a 4-dimensional space composed of source rank, target rank, message size and frequency.

The communication protocol presented here was developed in cooperation with Intel Labs Braunschweig for the RCKMPI library. RCKMPI is provided by Intel under an open-source license at the MARC community [1].

Support for this work was provided by the Transregional Collaborative Research Centre 89: Invasive Computing (InvasIC) [7].

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Intel’s Many-core Applications Research Community, http://communities.intel.com/community/marc

  2. KNEM: High-Performance Intra-Node MPI Communication, http://runtime.bordeaux.inria.fr/knem/

  3. Leibniz-Rechenzentrum (LRZ): SuperMUC Petascale System, http://www.lrz.de/services/compute/supermuc/systemdescription/

  4. MPICH2, http://www.mcs.anl.gov/research/projects/mpich2/

  5. Ohio State University (OSU) Micro-Benchmarks, http://mvapich.cse.ohio-state.edu/benchmarks/

  6. Open MPI, http://www.open-mpi.org/

  7. Transregional Research Center InvasIC, http://www.invasic.de

  8. Buntinas, D., Goglin, B., Goodell, D., Mercier, G., Moreaud, S.: Cache-Efficient, Intranode, Large-Message MPI Communication with MPICH2-Nemesis. In: Parallel Processing, ICPP 2009 (2009)

    Google Scholar 

  9. Chapman, K., Hussein, A., Hosking, A.L.: X10 on the Single-chip Cloud Computer: Porting and Preliminary Performance. In: Proceedings of the ACM SIGPLAN X10 Workshop (2011)

    Google Scholar 

  10. Christgau, S., Kiertscher, S., Schnor, B.: The Benefit of Topology Awareness of MPI Applications on the SCC. In: 3rd Many-core Applications Research Community (MARC) Symposium (2011)

    Google Scholar 

  11. Clauss, C., Lankes, S., Bemmerl, T.: Performance Tuning of SCC-MPICH by Means of the Proposed MPI-3.0 Tool Interface. In: Cotronis, Y., Danalis, A., Nikolopoulos, D.S., Dongarra, J. (eds.) EuroMPI 2011. LNCS, vol. 6960, pp. 318–320. Springer, Heidelberg (2011)

    Chapter  Google Scholar 

  12. Clauss, C., Lankes, S., Reble, P., Bemmerl, T.: Recent Advances and Future Prospects in iRCCE and SCC-MPICH. In: 3rd Many-core Applications Research Community (MARC) Symposium (2011)

    Google Scholar 

  13. Comprés Ureña, I.A., Gerndt, M.: Improved RCKMPI’s SCCMPB Channel: Scaling and Dynamic Processes Support. In: 4th Many-core Applications Research Community (MARC) Symposium (2011)

    Google Scholar 

  14. Comprés Ureña, I.A., Riepen, M., Konow, M.: RCKMPI – Lightweight MPI Implementation for Intel’s Single-chip Cloud Computer (SCC). In: Cotronis, Y., Danalis, A., Nikolopoulos, D.S., Dongarra, J. (eds.) EuroMPI 2011. LNCS, vol. 6960, pp. 208–217. Springer, Heidelberg (2011)

    Chapter  Google Scholar 

  15. Comprés Ureña, I.A., Riepen, M., Konow, M., Gerndt, M.: Invasive MPI on intel’s single-chip cloud computer. In: Proceedings of the 25th International Conference on Architecture of Computing Systems (2012)

    Google Scholar 

  16. Fuerlinger, K., Wright, N.J., Skinner, D.: Effective Performance Measurement at Petascale Using IPM. In: International Conference on Parallel and Distributed Systems, ICPADS (2010)

    Google Scholar 

  17. Held, J.: Single-chip Cloud Computer, an IA Tera-scale Research Processor. In: Guarracino, M.R., Vivien, F., Träff, J.L., Cannatoro, M., Danelutto, M., Hast, A., Perla, F., Knüpfer, A., Di Martino, B., Alexander, M. (eds.) Euro-Par-Workshop 2010. LNCS, vol. 6586, p. 85. Springer, Heidelberg (2011)

    Google Scholar 

  18. Mattson, T.G., Riepen, M., Lehnig, T., Brett, P., Haas, W., Kennedy, P., Howard, J., Vangal, S., Borkar, N., Ruhl, G., Dighe, S.: The 48-core SCC Processor: the Programmer’s View. In: Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis (2010)

    Google Scholar 

  19. Rotta, R.: On Efficient Message Passing on the Intel SCC. In: 3rd Many-core Applications Research Community (MARC) Symposium (2011)

    Google Scholar 

  20. Wong, F.C., Martin, R.P., Arpaci-Dusseau, R.H., Culler, D.E.: Architectural requirements and scalability of the nas parallel benchmarks. In: Proceedings of the Conference on Supercomputing (1999)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2012 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Comprés Ureña, I.A., Gerndt, M., Trinitis, C. (2012). Wait-Free Message Passing Protocol for Non-coherent Shared Memory Architectures. In: Träff, J.L., Benkner, S., Dongarra, J.J. (eds) Recent Advances in the Message Passing Interface. EuroMPI 2012. Lecture Notes in Computer Science, vol 7490. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-33518-1_19

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-33518-1_19

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-33517-4

  • Online ISBN: 978-3-642-33518-1

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics