Principles of speculative run-time parallelization Conference Paper uri icon

abstract

  • Springer-Verlag Berlin Heidelberg 1999. Current parallelizing compilers cannot identify a significant fraction of parallelizable loops because they have complex or statically insufficiently defined access patterns. We advocate a novel framework for the identification of parallel loops. It speculatively executes a loop as a doall and applies a fully parallel data dependence test to check for any unsatisfied data dependencies; if the test fails, then the loop is re-executed serially. We will present the principles of the design and implementation of a compiler that employs both run-time and static techniques to parallelize dynamic applications. Run-time optimizations always represent a tradeoff between a speculated potential benefit and a certain (sure) overhead that must be paid. We will introduce techniques that take advantage of classic compiler methods to reduce the cost of run-time optimization thus tilting the outcome of speculation in favor of significant performance gains. Experimental results from the PERFECT, SPEC and NCSA Benchmark suites show that these techniques yield speedups not obtainable by any other known method.

name of conference

  • Languages and Compilers for Parallel Computing, 11th International Workshop, LCPC'98, Chapel Hill, NC, USA, August 7-9, 1998, Proceedings

published proceedings

  • LANGUAGES AND COMPILERS FOR PARALLEL COMPUTING

author list (cited authors)

  • Patel, D., & Rauchwerger, L.

citation count

  • 4

complete list of authors

  • Patel, D||Rauchwerger, L

editor list (cited editors)

  • Chatterjee, S., Prins, J. F., Carter, L., Ferrante, J., Li, Z., Sehr, D. C., & Yew, P.

publication date

  • January 1999