Skip to main content
Log in

Efficient diagonalization of oversized matrices on a distributed-memory multiprocessor

  • Published:
Annals of Operations Research Aims and scope Submit manuscript

Abstract

On parallel architectures, Jacobi methods for computing the singular value decomposition (SVD) and the symmetric eigenvalue decomposition (EVD) have been established as one of the most popular algorithms due to their excellent parallelism. Most of the Jacobi algorithms for distributed-memory architectures have been developed under the assumption that matrices can be distributed over the processors by square blocks of an even order or column blocks with an even number of columns. Obviously, there is a limit on the number of processors while we need to deal with problems of various sizes. We propose algorithms to diagonalize oversized matrices on a given distributed-memory multiprocessor with good load balancing and minimal message passing. Performances of the proposed algorithms vary greatly, depending on the relation between the problem size and the number of available processors. We give theoretical performance analyses which suggest the faster algorithm for a given problem size on a given distributed-memory multiprocessor. Finally, we present a new implementation for the convergence test of the algorithms on a distributed-memory multiprocessor and the implementation results of the algorithms on the NCUBE/seven hypercube architecture.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. M. Berry and A. Sameh, Parallel algorithms for the singular value and dense symmetric eigenvalue problems, CSRD No. 761, University of Illinois (1988).

  2. R.P. Brent, F.T. Luk and C.F. Van Loan, Computation of the singular value decomposition using mesh-connected processors, J. VLSI Computer Systems 1(1985)242–270.

    Google Scholar 

  3. P.J. Eberlein, On using the Jacobi method on the hypercube, in:Proc. 2nd Conf. on Hypercube Multiprocessors, ed. M.T. Heath (1987) pp. 605–611.

  4. P.J. Eberlein, On one-sided Jacobi methods for parallel computation, SIAM J. Alg. Disc. Meth. 8(1987)790–796.

    Google Scholar 

  5. P.J. Eberlein and H. Park, Efficient implementation of Jacobi algorithms and Jacobi sets on distributed memory architectures, J. Parall. and Dist. Comput., special issue on algorithms for hypercube computers, to appear.

  6. P.J. Eberlein and H. Park, Eigensystem computation on hypercube architectures, in:Proc. 4th Conf. on Hypercubes, Concurrent Computers, and Applications, to appear.

  7. M.R. Hestenes, Inversion of matrices by biorthogonalization and related results, J. Soc. Indust. Appl. Math. 6(1958)51–90.

    Article  Google Scholar 

  8. F.T. Luk and H. Park, On parallel Jacobi orderings, SIAM J. Sci. Statist. Comput. 10 (1989) 18–26.

    Article  Google Scholar 

  9. F.T. Luk and H. Park, A proof of convergence for two parallel Jacobi SVD algorithms, IEEE Trans. on Computers 38(1989)806–811.

    Article  Google Scholar 

  10. D.P. O'Leary and G.W. Stewart, Data-flow algorithms for parallel matrix computations, Comm. ACM 28(1985)841–853.

    Article  Google Scholar 

  11. R. Schreiber, Solving eigenvalue and singular value problems on an undersized systolic array, SIAM J. Sci. Stat. Comput. 7(1986)441–451.

    Article  Google Scholar 

  12. D.S. Scott, M.T. Heath and R.C. Ward, Parallel block Jacobi eigenvalue algorithms using systolic arrays, Lin. Alg. and its Appl. 77(1986)345–355.

    Article  Google Scholar 

  13. C. Van Loan, The block Jacobi method for computing the singular value decomposition, Report TR 85-680, Department of Computer Science, Cornell University (1985).

Download references

Author information

Authors and Affiliations

Authors

Additional information

This work was supported by National Science Foundation grant CCR-8813493. This work was partly done during the author's visit to the Mathematical Science Section, Engineering Physics and Mathematics Division, Oak Ridge National Laboratory, while participating in the Special Year on Numerical Linear Algebra, 1988, sponsored by the UTK Departments of Computer Science and Mathematics, and the ORNL Algebra sponsored by the UTK Departments of Computer Science and Mathematics, and the ORNL Mathematical Sciences Section, Engineering Physics and Mathematics Division.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Park, H. Efficient diagonalization of oversized matrices on a distributed-memory multiprocessor. Ann Oper Res 22, 253–269 (1990). https://doi.org/10.1007/BF02023056

Download citation

  • Issue Date:

  • DOI: https://doi.org/10.1007/BF02023056

Keywords

Navigation