Advertisement

On Automatic Parallelization of Irregular Reductions on Scalable Shared Memory Systems⋆

  • E. Gutiérrez
  • O. Plata
  • E. L. Zapata
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 1685)

Abstract

This paper presents a new parallelization method for reductions of arrays with subscripted subscripts on scalable shared-memory multiprocessors. The mapping of computations is based on the conflict-free write distribution of the reduction vector across the processors. The proposed method is general, scalable, and easy to implement on a compiler. A performance evaluation and comparison with other existing techniques is presented. From the experimental results, the proposed method is a clear alternative to the array expansion and privatized buffer methods, usual on state-of-the-art parallelizing compilers, like Polaris or SUIF.

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. [1]
    R. Asenjo, E. Gutierrez, Y. Lin, D. Padua, B. Pottengerg and E. Zapata, On the Automatic Parallelization of Sparse and Irregular Fortran Codes, TR1512, Univ. of Illinois at Urbana-Champaign, Ctr. for Supercomputing R&D., Dec. 1996.Google Scholar
  2. [2]
    W. Blume, R. Doallo, R. Eigemann, et. al., Parallel Programming with Polaris, IEEE Computer, 29(12):78–82, Dec. 1996.Google Scholar
  3. [3]
    I. Foster, R. Schreiber and P. Havlak, HPF-2, Scope of Activities and Motivating Applications, Tech. Rep. CRPC-TR94492, Rice Univ., Nov. 1994.Google Scholar
  4. [4]
    M.W. Hall, J.M. Anderson, S.P. Amarasinghe, et. al., Maximizing Multiprocessor Performance with the SUIF Compiler IEEE Computer, 29(12), Dec. 1996.Google Scholar
  5. [5]
    H. Han and C.-W. Tseng, Improving Compiler and Run-Time Support for Irregular Reductions, 11th Workshop on Languages and Compilers for Parallel Computing, Chapel Hill, NC, Aug. 1998.Google Scholar
  6. [6]
    High Performance Fortran Language Specification, Version 2.0. High Performance Fortran Forum, Oct. 1996Google Scholar
  7. [7]
    Y. Lin and D. Padua, On the Automatic Parallelization of Sparse and Irregular Fortran Programs, 4th Workshop on Languages, Compilers and Runtime Systems for Scalable Computers, Pittsburgh, PA, May 1998.Google Scholar
  8. [8]
    OpenMP: A Proposed Industry Standard API for Shared Memory Programming, OpenMP Architecture Review Board, http://www.openmp.org), 1997.
  9. [9]
    R. Ponnusamy, J. Saltz, A. Choudhary, S. Hwang and G. Fox, Runtime Support and Compilation Methods for User-Specified Data Distributions, IEEE Trans. on Parallel and Distributed Systems, 6(8):815–831, Jun. 1995.CrossRefGoogle Scholar
  10. [10]
    B. Pottenger and R. Eigenmann, Idiom Recognition in the Polaris Parallelizing Compiler, 9th ACM Int’l Conf. on Supercomputing, Barcelona, Spain, pp. 444–448, Jul. 1995.Google Scholar
  11. [11]
    Silicon Graphics, Inc. MIPSpro Automatic Parallelization. SGI, Inc. 1998Google Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 1999

Authors and Affiliations

  • E. Gutiérrez
    • 1
  • O. Plata
    • 1
  • E. L. Zapata
    • 1
  1. 1.Department of Computer ArchitectureUniversity of MálagaMálagaSpain

Personalised recommendations