Skip to main content

Providing Green Services in HPC Data Centers: A Methodology Based on Energy Estimation

  • Chapter
  • First Online:
Handbook on Data Centers

Abstract

A supercomputer is an infrastructure built from an interconnection of computers capable of performing tasks in parallel in order to achieve very high performance. They are used in order to run scientific applications in various fields like the prediction of severe weather phenomena and seismic waves. To meet new scientific challenges, the HPC community has set a new performance objective for the end of the decade: Exascale. To achieve such performance (1018 FLoat Operations Per Second), an exascale supercomputer will gather several millions of CPU cores running up to a billion trends and will consume several megawatts. The energy consumption issue at the exascale becomes even more worrying when we know that we already reach energy consumptions higher than 17 MW at the petascale while the DARPA set to 20 MW the threshold for exascale supercomputers. Hence, these systems that will be 30 times more performant than the current systems have to achieve an energy efficiency of 50 gigaFLOPS per watt while the current ones achieve between 2 and 3 gigaFLOPS per watt. As a consequence, reducing the energy consumption of high-performance computing infrastructures is a major challenge for the next years in order to be able to move to the exascale era.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 169.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Hardcover Book
USD 219.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    MPICH2: http://www.mcs.anl.gov/research/projects/mpich2/.

  2. 2.

    OpenMPI: http://www.open-mpi.org/.

  3. 3.

    Cloud Model 1: http://www.mmm.ucar.edu/people/bryan/cm1/.

  4. 4.

    NAS: http://www.nas.nasa.gov/publications/npb.html

References

  1. Aloisio, G. and Fiore, S. (2009). Towards Exascale Distributed Data Management. IJHPCA, 23(4):398–400.

    Google Scholar 

  2. Bergman, K., Borkar, S., Campbell, D., and others (2008). ExaScale Computing Study: Technology Challenges in Achieving Exascale Systems. In DARPA Information Processing Techniques Office, page pp. 278, Washington, DC.

    Google Scholar 

  3. Bouteiller, A., Bosilca, G., and Dongarra, J. (2010). Redesigning the message logging model for high performance. Concurrency and Computation: Practice and Experience, 22(16):2196–2211.

    Article  Google Scholar 

  4. Cappello, F., Caron, E., Daydé, M. J., Desprez, F., Jégou, Y., Primet, P. V.-B., Jeannot, E., Lanteri, S., Leduc, J., Melab, N., Mornet, G., Namyst, R., Quétier, B., and Richard., O. (2005). Grid'5000: A Large Scale, Reconfigurable, Controlable and Monitorable Grid Platform. In IEEE/ACM Grid 2005, Seattle, Washington, USA.

    Google Scholar 

  5. Cappello, F., Geist, A., Gropp, B., Kale, S., Kramer, B., and Snir, M. (2009). Toward exascale resilience. International Journal of High Performance Computing Applications, 23:374–388.

    Article  Google Scholar 

  6. Daly, J. T. (2006). A higher order estimate of the optimum checkpoint interval for restart dumps. Future Generation Comp. Syst., 22(3):303–312.

    Article  Google Scholar 

  7. Dias de Assuncao, M., Gelas, J.-P., Lefèvre, L., and Orgerie, A.-C. (2010a). The green grid5000: Instrumenting a grid with energy sensors. In 5th International Workshop on Distributed Cooperative Laboratories: Instrumenting the Grid (INGRID 2010), Poznan, Poland.

    Google Scholar 

  8. Dias de Assuncao, M., Orgerie, A.-C., and Lefèvre, L. (2010b). An analysis of power consumption logs from a monitored grid site. In IEEE/ACM International Conference on Green Computing and Communications (GreenCom-2010), pages 61–68, Hangzhou, China.

    Google Scholar 

  9. Diouri, M. E. M., Dolz, M. F., Glück, O., Lefèvre, L., Alonso, P., Catalán, S., Mayo, R., and Quintana-Ortí, E. S. (2013a). Solving some mysteries in power monitoring of servers: Take care of your wattmeters! In Energy Efficiency in Large Scale Distributed Systems (EE-LSDS), Vienna, Austria, April, 22–24 201–3.

    Google Scholar 

  10. Diouri, M. E. M., Glück, O., and Lefèvre, L. (2013b). Your Cluster is not Power Homogeneous: Take Care when Designing Green Schedulers! In 4 th IEEE International Green Computing Conference (IGCC), Arlington, VA USA.

    Google Scholar 

  11. Diouri, M. E. M., Glück, O., Lefèvre, L., and Cappello, F. (2013c). ECOFIT: A Framework to Estimate Energy Consumption of Fault Tolerance protocols during HPC executions. In 13th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGrid), Delft, Netherlands.

    Google Scholar 

  12. Diouri, M. E. M., Glück, O., Lefèvre, L., and Mignot, J.-C. (2013d). Energy Estimation for MPI Broadcasting Algorithms in Large Scale HPC Systems. In 20 th European MPI Users` Group Meeting on Recent Advances in Message Passing Interface (EuroMPI 2013), Madrid, Spain.

    Google Scholar 

  13. Diouri, M. E. M., Tsafack Chetsa, G. L., Glück, O., Lefèvre, L., Pierson, J.-M., Stolf, P., and Da Costa, G. (2013e). Energy efficiency in high-performance computing with and without knowledge of applications and services. International Journal of High Performance Computing Applications (IJHPCA), 27(3):232–243.

    Google Scholar 

  14. Etinski, M., Corbalan, J., Labarta, J., and Valero, M. (2010). Utilization driven power-aware parallel job scheduling. Computer Science - Research and Development, 25(3–4):207–216.

    Article  Google Scholar 

  15. Freeh, V. W., Lowenthal, D. K., Pan, F., Kappiah, N., Springer, R., Rountree, B., and Femal, M. E. (2007). Analyzing the energy-time trade-off in high-performance computing applications. IEEE Trans. Parallel Distrib. Syst., 18(6):835–848.

    Article  Google Scholar 

  16. Hermenier, F., Loriant, N., and Menaud, J.-M. (2006). Power Management in Grid Computing with Xen. In Frontiers of High Performance Computing and Networking - ISPA 2006 International Workshops, volume 4331 of Lecture Notes in Computer Science, pages 407–416, Sorrento, Italy.

    Google Scholar 

  17. Hlavacs, H., Da Costa, G., and Pierson, J.-M. (2009). Energy consumption of residential and professional switches. In IEEE CSE.

    Google Scholar 

  18. Hotta, Y., Sato, M., Kimura, H., Matsuoka, S., Boku, T., and Takahashi, D. (2006). Profile-based optimization of power performance by using dynamic voltage scaling on a pc cluster. In Proceedings of the 20th International in Parallel and Distributed Processing Symposium, IPDPS 2006.

    Google Scholar 

  19. Mahadevan, P., Sharma, P., Banerjee, S., and Ranganathan, P. (2009). A power benchmarking framework for network devices. In NETWORKING 2009 Conference, Aachen, Germany, May 11–15, 2009., pages 795–808.

    Google Scholar 

  20. Netzer, R. H. B. and Xu, J. (1995). Necessary and sufficient conditions for consistent global snapshots. IEEE Transactions on Parallel and Distributed Systems, 6(2):165–169.

    Article  Google Scholar 

  21. Orgerie, A.-C., Lefevre, L., and Gelas, J.-P. (2008). Save Watts in your Grid: Green Strategies for Energy-Aware Framework in Large Scale Distributed Systems. In ICPADS 2008: The 14th IEEE International Conference on Parallel and Distributed Systems, Melbourne, Australia.

    Google Scholar 

  22. Pinheiro, E., Bianchini, R., Carrera, E. V., and Heath, T. (2001). Load balancing and unbalancing for power and performance in cluster-based systems. In IN WORKSHOP ON COMPILERS AND OPERATING SYSTEMS FOR LOW POWER.

    Google Scholar 

  23. Rabenseifner, R., Hager, G., and Jost, G. (2009). Hybrid mpi/openmp parallel programming on clusters of multi-core smp nodes. In Parallel, Distributed and Network-based Processing, 2009 17th Euromicro International Conference on, pages 427 –436.

    Google Scholar 

  24. Rao, C., Toutenburg, H., Fieger, A., Heumann, C., Nittner, T., and Scheid, S. (1999). Linear models: Least squares and alternatives. Springer Series in Statistics.

    Google Scholar 

  25. Wadsworth, D. M. and Chen, Z. (2008). Performance of MPI broadcast algorithms. In IEEE IPDPS 2008, Miami, Florida USA, April 14–18, 2008, pages 1–7.

    Google Scholar 

  26. Young, J. W. (1974). A first order approximation to the optimum checkpoint interval. Commun. ACM, 17(9):530–531.

    Article  MATH  Google Scholar 

Download references

Acknowledgment

Experiments presented in this chapter were carried out using the Grid'5000 experimental testbed, being developed under the INRIA ALADDIN development action with support from CNRS, RENATER and several Universities as well as other funding bodies (see http://www.grid5000.fr).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Mohammed El Mehdi Diouri .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2015 Springer Science+Business Media New York

About this chapter

Cite this chapter

Diouri, M., Glück, O., Lefèvre, L., Mignot, JC. (2015). Providing Green Services in HPC Data Centers: A Methodology Based on Energy Estimation. In: Khan, S., Zomaya, A. (eds) Handbook on Data Centers. Springer, New York, NY. https://doi.org/10.1007/978-1-4939-2092-1_9

Download citation

  • DOI: https://doi.org/10.1007/978-1-4939-2092-1_9

  • Published:

  • Publisher Name: Springer, New York, NY

  • Print ISBN: 978-1-4939-2091-4

  • Online ISBN: 978-1-4939-2092-1

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics