Supercomputing pp 229-254 | Cite as

Strategies for cache and local memory management by global program transformation

  • Dennis Gannon
  • William Jalby
  • Kyle Gallivan
Session 4A: Compilers And Restructuring Techniques I
Part of the Lecture Notes in Computer Science book series (LNCS, volume 297)


Local Memory Dependence Graph Data Dependence Iteration Space Distance Vector 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

7. References

  1. [Allen83]
    J.R. Allen, “Dependence Analysis for Subscripted Variables and Its Application to Program Transformations,” Ph.D. Thesis, Rice University, Houston, Texas, April 1983.Google Scholar
  2. [AlKe84]
    J. Allen, and K. Kennedy, “A Parallel Programming Environment,” technical report, Rice COMP TR84-3, July 1984.Google Scholar
  3. [ASKL79]
    W. Abu-Sufah, D. Kuck and D. Lawrie, “Automatic Program Transformations for Virtual Memory Computers,” Proc. of the 1979 Nat'l Computer Conf., June, 1979, 969–974.Google Scholar
  4. [ASKL81]
    W. Abu-Sufah, D. Kuck and D. Lawrie, “On the Performance Enhancement of Paging Systems Through Program Analysis and Program Transformation,” IEEE Trans. Comp. V. C30 no. 5, May, 1981, pages 341–356.Google Scholar
  5. [BuCy86]
    M. Burke, R. Cytron, “Interprocedural Dependence Analysis and Parallelization,” Proc. Sigplan 86 Symposium on Compiler Construction, 21(7):162–175, July 1986.Google Scholar
  6. [CoMK69]
    Coffman, E., McKeller, “The organization of matricies and matrix operations in the paged multiprogramming environment,” CACM V. 12 pp. 153–165. March 1969.Google Scholar
  7. [CGST85]
    W. Crowther, J. Goodhue, E. Starr, R Thomas, W. Milliken, T. Blackadar, "Performance Measurements on a 128-node Butterfly Parallel Processor," Proceedings of 1985 International Conference on Parallel Processing, pp. 531–540, 1985.Google Scholar
  8. [Cytron84]
    R. Cytron, “Compile-time Scheduling and Optimization for Asynchronous Machines,” Ph.D. Thesis, University of Illinois, Urbana-Champaign Aug., 1984 Report No. UIUCDCS-R-84-1177).Google Scholar
  9. [FeOt83]
    J. Ferante, K. Ottenstein, J. Warren, “The Program Dependence Graph and Its Uses in Optimization,” IBM Technical Report RC 10208, Aug. 1983.Google Scholar
  10. [GaJa87]
    Gannon, D., Jalby, W., “The Influence of Memory Heirarchy on Algorithm Organization: Programming FFTs on a Vector Multiprocessor,” to appear in “The Characteristics of Parallel Algorithms,” Gannon, Jamieson, Douglas, eds, MIT Press, 1987.Google Scholar
  11. [Kenn80]
    K. Kennedy, “Automatic translation of Fortran programs to vector form,” Rice Technical Report 476-029-4, Rice University, October 1980Google Scholar
  12. [KKLW80]
    D. Kuck, R. Kuhn, B. Leasure and M. Wolfe, “The Structure of an Advanced Vectorizer for Pipelined Processors,” IEEE Computer Society, proc. of the 4th Inter"l Computer Software and App. Conf., October, 1980, 709–715.Google Scholar
  13. [KKLPW81]
    D. J. Kuck, R. H. Kuhn, B. Leasure, D. H. Padua and M. Wolfe, “Dependence graphs and compiler optimizations,” Conference Record of Eighth Annual ACM Symposium on Principles of Programming Languages, Williamsburg, VA., January 1981.Google Scholar
  14. [KuWM84]
    D. Kuck, M. Wolfe, and J. McGraw, “A Debate: Retire FORTRAN?,” Physics Today, May, 1984, 67–75.Google Scholar
  15. [Padua79]
    D. Padua, “Multiprocessors: Discussion of Some Theoretical and Practical Problems,” Ph.D. Thesis, University of Illinois, Urbana-Champaign, Nov. 1979.Google Scholar
  16. [PaKu80]
    D. Padua and D. Kuck, “High-Speed Multiprocessors and Compilation Techniques,” IEEE Transactions on Computers, Vol. C-29, No. 9, September, 1980, 763–776.Google Scholar
  17. [PaWo86]
    D. Padua and M. Wolfe, “Advanced Compiler Optimizations for Supercomputers,” CACM, 29(12):1184–1201, Dec. 1986.Google Scholar
  18. [PhNo85]
    G. Phister, A. Norton, “Hot Spot Contention and Combining in Multistage Interconnection Networks,” Proceeding of the 1985 International Conference on Parallel Processing, 1985, 790–797.Google Scholar
  19. [Poly86]
    C. Polychronopoulos, “On Program Restructuring, Scheduling, and Communication for Parallel Processor Systems,” Ph.D. Thesis, University of Illinois Center for Supercomputer Research and Development. CSRD TR.595, Aug. 1986.Google Scholar
  20. [SaSa75]
    Saraj, Sahni, “Approximate Algorithms for the 0,1 Knapsack Problem,” JACM V. 22 no. 1 pp 115–124.Google Scholar
  21. [KWDG87]
    Wang, K.-Y., Gannon, D., “Applying AI Techniques to Program Optimization for Parallel Computers,” To appear in “AI Machines and Supercomputer Systems”, Hwang, DeGroot, eds. McGraw Hill, 1987.Google Scholar
  22. [Wolfe82]
    M. Wolfe, “Optimizing Supercompilers for Supercomputers,” Ph.D. Thesis, Dept. of Computer Science, University of Illinois, Urbana-Chanpaign, 1982, Report no. UIUCDCS-R-82-1105.Google Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 1988

Authors and Affiliations

  • Dennis Gannon
    • 1
  • William Jalby
    • 2
  • Kyle Gallivan
    • 3
  1. 1.Dept. of Computer ScienceIndiana University and CSRD Univ. of IllinoisUrbana-Champaign
  2. 2.INRIA and CSRD Univ. of IllinoisUrbana-Champaign
  3. 3.CSRD Univ. of IllinoisUrbana-Champaign

Personalised recommendations