The limited ability of compilers to nd the parallelism in programs is a signi cant barrier to the use of high performance computers. It forces programmers to resort to parallelizing their programs by hand, adding another level of complexity to the programming task. We show evidence that compilers can be improved, through static and run-time techniques, to the extent that a signi cant group of scienti c programs may be parallelized automatically. Symbolic dependence analysis and array privatization, plus run-time versions of those techniques are shown to be important to the success of this e ort. If we can succeed to parallelize programs automatically, the acceptance and use of large-scale parallel processors will be enhanced greatly
A certain balance between language features and compiler complexity is required to achieve reasonabl...
Parallel computers can provide impressive speedups, but unfortunately such speedups are difficult to...
Current parallelizing compilers cannot identify a significant fraction of parallelizable loops becau...
The goal of parallelizing, or restructuring, compilers is to detect and exploit parallelism in seque...
This paper presents an overview of automatic program parallelization techniques. It covers dependenc...
Modern computers will increasingly rely on parallelism to achieve high computation rates. Techniques...
Compiling for parallelism is a longstanding topic of compiler research. This book describes the fund...
This paper demonstrates that significant improvements to automatic parallelization technology requir...
Thesis (Ph. D.--University of Rochester. Dept. of Computer Science, 1991. Simultaneously published i...
A new technique for estimating and understanding the speed improvement that can result from executin...
Automatic parallelization techniques for finding loop-based parallelism fail to find efficient paral...
The notion of dependence captures the most important properties of a program for efficient execution...
Characteristics of full applications found in scientific computing industries today lead to challeng...
It is the goal of the Polaris project to develop a new parallelizing compiler that will overcome li...
To effectively translate real programs written in standard, sequential languages into parallel compu...
A certain balance between language features and compiler complexity is required to achieve reasonabl...
Parallel computers can provide impressive speedups, but unfortunately such speedups are difficult to...
Current parallelizing compilers cannot identify a significant fraction of parallelizable loops becau...
The goal of parallelizing, or restructuring, compilers is to detect and exploit parallelism in seque...
This paper presents an overview of automatic program parallelization techniques. It covers dependenc...
Modern computers will increasingly rely on parallelism to achieve high computation rates. Techniques...
Compiling for parallelism is a longstanding topic of compiler research. This book describes the fund...
This paper demonstrates that significant improvements to automatic parallelization technology requir...
Thesis (Ph. D.--University of Rochester. Dept. of Computer Science, 1991. Simultaneously published i...
A new technique for estimating and understanding the speed improvement that can result from executin...
Automatic parallelization techniques for finding loop-based parallelism fail to find efficient paral...
The notion of dependence captures the most important properties of a program for efficient execution...
Characteristics of full applications found in scientific computing industries today lead to challeng...
It is the goal of the Polaris project to develop a new parallelizing compiler that will overcome li...
To effectively translate real programs written in standard, sequential languages into parallel compu...
A certain balance between language features and compiler complexity is required to achieve reasonabl...
Parallel computers can provide impressive speedups, but unfortunately such speedups are difficult to...
Current parallelizing compilers cannot identify a significant fraction of parallelizable loops becau...