Skip to main content

Abstract

As parallel computers and parallel programming environments are used widely, people try to seek a synthetical metric to evaluate how problem sizes and system sizes influence the performance of parallel computers and parallel algorithms. Scalability is a try at this aspect. Scalability has been widely used in practice to describe how system sizes and problem sizes influence the performance of parallel computers and algorithms. It is so very important that manufacturers of parallel computers all claim that their parallel computers are scalable, and give their appropriate scaling ranges. For example, Intel Paragon XP/S was designed to scale between 16 and 1024 processors. IBM SP2 scaled over a range of 256 processors. The Cray Y-MP series scaled over a range of 16 processors, and Cray T3D was designed to scale between 16 and 1024 processors. The CM-2 scaling range is 8K to 64K processing elements, and the CM-5 scaling range is 1024 to 16K computers [12].

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 129.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 169.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Gordon Bell, Ultracomputers, A Teraflop before its time, Communication of the ACM35, 1992.

    Google Scholar 

  2. J. Berntsen, Communication efficient matrix multiplication on hypercubes, Parallel Computing 12, 1989, 335–342.

    Article  MATH  Google Scholar 

  3. L.E. Cannon, A cellular computer to implement the Kalman Filter Algorithm, PH.D. Thesis, Montana State University, 1969.

    Google Scholar 

  4. E. Dekel, D. Nassimi, and S. Sahni, Parallel matrix and graph algorithms, SIAM Journal of Computing 10, 1981, 657–673.

    Article  MathSciNet  MATH  Google Scholar 

  5. S.R. Deshpande and R.M. Jenevein, Scalability of a binary tree on a hypercube, In: Proc. of 1986 International Conference on Parallel Computing, 1986, 661–668.

    Google Scholar 

  6. J.J. Dongarra et al., Scalability issues affecting the design of a dense linear algebra library, Journal of Parallel and Distributed Computing 22, 1994.

    Google Scholar 

  7. G.C. Fox, et al., Matrix algorithms on a hypercube I: Matrix multiplication, Parallel Computing 4, 1987, 17–31.

    Article  MATH  Google Scholar 

  8. A.Y. Grama, A. Gupta, and V. Kumar, Isoefficiency: Measuring the scalability of parallel algorithms and architectures, IEEE Parallel & Distributed Technology, August 1993, 12–21.

    Google Scholar 

  9. A. Gupta and V. Kumar, Scalability of parallel algorithms for matrix multiplication, In: Proc. of 1993 International Conference on Parallel Processing, 1993, 115–123.

    Google Scholar 

  10. A. Gupta and V. Kumar, The Scalability of FFT on parallel computers, IEEE Transactions on Parallel and Distributed Systems, Vol.4, No.8, August, 1993.

    Google Scholar 

  11. J. Gustafson, Reevaluating Amdahl’s law, Communication of the ACM, Vol.31, No.5, May 1988.

    Google Scholar 

  12. Kai Hwang, Advanced Computer Architecture: Parallelism, Scalability, Programmability, McGraw-Hill, Inc., 1993.

    Google Scholar 

  13. S.P. Kumar and S.M. Stern, A scalability analysis of the butterfly multiprocessor, In: Proc. of Super computing’s 88, 1988.

    Google Scholar 

  14. V. Kumar and V.N. Rao, Parallel depth-first search II. Analysis, International Journal of Parallel Programming 16, No.6, 1987.

    Google Scholar 

  15. V. Kumar and A. Gupta, Analyzing scalability of parallel algorithms and architectures, Journal of Parallel and distributed Computing 22, 1994, 379–391.

    Article  Google Scholar 

  16. V. Kumar, et al., Scalable load balancing techniques for parallel computers, Journal of Parallel and distributed Computing 22, 1994, 60–79.

    Article  Google Scholar 

  17. Y.E. Ma and D.G. Shea, Downward scalability of parallel processing, In: Proc. of Supercomputing’s 88, 1988.

    Google Scholar 

  18. D. Muller-Wichards and W. Ronsch, Scalability of algorithms: An analytic approach, Parallel Computing 21, 1995.

    Google Scholar 

  19. D. Nussbaum and A. Agarwal, Scalability of parallel machines, Communication of ACM 34, No.3, 1991, 57–61.

    Article  MathSciNet  Google Scholar 

  20. S. Sahni and V. Thanvantri, Performance metrics: Keeping the focus on runtime, IEEE Parallel & Distributed Technology, Spring, 1996.

    Google Scholar 

  21. J.P. Singh, et al., Scaling parallel programs for multiprocessors: Methodology and examples, Computer, July 1993.

    Google Scholar 

  22. A. Sivasubramaniam et al., An approach to scalability study of shared memory parallel systems, In: ACMSIGMETRICSs 94.

    Google Scholar 

  23. X.H. Sun and L. Ni, Scalable problems and memory-bounded speedup, Journal of Parallel and Distributed Computing 19, 1993, 27–37.

    Article  MATH  Google Scholar 

  24. Xian-He Sun and D.T. Rover, Scalability of parallel algorithm-machine combinations, IEEE Transactions on Parallel and Distributed Systems, Vol.5, No.6, June 1994, 599–613.

    Article  Google Scholar 

  25. Xian-He Sun, The relation of scalability and execution time, In: Proc. of IPPS’s 96), IEEE Computer Society Press,. 1996

    Google Scholar 

  26. Xingfu Wu and Wei Li, Scalability of parallel algorithm implemantation, In: Proc. of the 2th Intern. Symp. on Parallel Architectures, Algorithms, and NetworksflSPANs 96), IEEE Computer Society Press, June 1996, 559–56

    Google Scholar 

  27. Xingfu Wu, Scalable Parallel Computing Performance Models and Their Applications, Ph.D thesis, Beijing University of Aeronautics & Astronautics, 1996.

    Google Scholar 

  28. Xingfu Wu and Wei Li, Performance models for scalable cluster computing, Journal of Systems Architecture, Vol. 44, No. 3, Elsevier Science Publishers B.V. (North-Holland), Dec. 1997, 189–205.

    Article  Google Scholar 

  29. Xingfu Wu, An approach to scalability of parallel matrix multiplication algorithms, COCOON’97, Lecture Notes in Computer Science 1276, Springer-Verlag, August 1997, 492–501.

    Google Scholar 

  30. Xiaodong Zhang, et al., Latency metric: An experimental method for measuring and evaluating parallel program and architecture scalability, Journal of Parallel and Distributed Computing 22, 1994.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Rights and permissions

Reprints and permissions

Copyright information

© 1999 Springer Science+Business Media New York

About this chapter

Cite this chapter

Wu, X. (1999). Scalability. In: Performance Evaluation, Prediction and Visualization of Parallel Systems. The Kluwer International Series on Asian Studies in Computer and Information Science, vol 4. Springer, Boston, MA. https://doi.org/10.1007/978-1-4615-5147-8_3

Download citation

  • DOI: https://doi.org/10.1007/978-1-4615-5147-8_3

  • Publisher Name: Springer, Boston, MA

  • Print ISBN: 978-1-4613-7343-8

  • Online ISBN: 978-1-4615-5147-8

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics