Advertisement

New techniques for Cycle Shrinking

  • Yves Robert
  • Siang W. Song
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 605)

Abstract

Several loop transformations techniques have been designed to extract parallelism from nested loop structures. We first review two important approaches, known as Generalized Cycle Shrinking presented by Shang, O'Keefe and Fortes and the Index Shift Method introduced by Liu, Ho and Sheu. The main result of the paper is a new methodology that permits to combine cycle shrinking techniques with the index shift method. We present a new optimization method that produces the best scheduling vector, and we show that we can outperform previous results by an arbitrary speedup factor.

Keywords

Hull 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. [1]
    U. Banerjee, “An introduction to a formal theory of dependence analysis”, The Journal of Supercomputing 2, 1988, pp. 133–149.CrossRefGoogle Scholar
  2. [2]
    A. Darte and Y. Robert, Scheduling uniform loop nests, Technical Report LIP-IMAG 92-10, Ecole Normale Supérieure de Lyon, February 1992.Google Scholar
  3. [3]
    M.L. Dowling, “Optimal code parallelization using unimodular transformations”, Parallel Computing, 16, 1990, pp. 157–171.CrossRefGoogle Scholar
  4. [4]
    C.E. Leiserson and J.B. Saxe, “Optimizing synchronous systems”, Journal of VLSI Computer Systems, Vol. 1, April, 1983, pp. 41–67.Google Scholar
  5. [5]
    L.S. Liu, C.W. Ho, and J.P. Sheu, “On the parallelism of nested for-loops using index shift method”, Proceedings of International Conference on Parallel Processing, August 1990, pp. 11-119–11-123.Google Scholar
  6. [6]
    J.K. Peir and R. Cytron, “Minimum distance: a method for partitioning recurrences for multiprocessors”, IEEE Transactions on Computers, Vol. 38, No. 8, August 1989, pp. 1203–1211.CrossRefGoogle Scholar
  7. [7]
    C.D. Polychronopoulos, “Compiler optimization for enhancing parallelism and their impact on architecture design”, IEEE Transactions on Computers, Vol. 37, No. 8, August 1988, pp. 991–1004.CrossRefGoogle Scholar
  8. [8]
    C.D. Polychronopoulos, Parallel Programming and Compilers, Kluwer Academic Publishers, Boston, 1988.Google Scholar
  9. [9]
    Y. Robert and S.W. Song, Revisiting cycle shrinking, Technical Report LIP-IMAG 91–31, Ecole Normale Supérieure de Lyon, September 1991.Google Scholar
  10. [10]
    W. Shang and J.A.B. Fortes, “Time optimal linear schedules for algorithms with uniform dependences”, Proceedings of International Conference on Systolic Arrays, May 1988, pp. 393–402.Google Scholar
  11. [11]
    W. Shang and J.A.B. Fortes, “Time optimal linear schedules for algorithms with uniform dependencies”, IEEE Transactions on Computers, Vol. 40, No. 6, June 1991, pp. 723–742.CrossRefGoogle Scholar
  12. [12]
    W. Shang, M.T. O'Keefe, J.A.B. Fortes, “On loop transformations for generalized cycle shrinking”, Proceedings of International Conference on Parallel Processing, August 1991, pp. II-132–II-141.Google Scholar
  13. [13]
    M. Wolfe, Optimizing Supercompilers for Supercomputers, MIT Press, Cambridge MA, 1989.Google Scholar
  14. [14]
    M. Wolfe, “Data dependence and program restructuring”, The Journal of Supercomputing 4, 1990, pp. 321–344.CrossRefGoogle Scholar

Copyright information

© Springer-Verlag 1992

Authors and Affiliations

  • Yves Robert
    • 1
  • Siang W. Song
    • 2
  1. 1.Laboratoire de l'Informatique du Parallélisme - IMAG École Normale Supérieure de LyonLyon Cédex 07France
  2. 2.Institute of Mathematics and StatisticsUniversity of São PauloSão Paulo, SPBrazil

Personalised recommendations