Current parallelizing compilers do a reasonable job of extracting parallelism from programs with regular, well behaved, statically analyzable access patterns. However, they cannot extract a significant fraction of the available parallelism if the program has a complex and/or statically insufficiently defined access pattern, e.g., simulation programs with irregular domains and/or dynamically changing interactions. Since such programs represent a large fraction of all applications, techniques are needed for extracting their inherent parallelism at run–time. In this paper we give a new run–time technique for finding an optimal parallel execution schedule for a partially parallel loop, i.e., a loop whose parallelization requires synchronization...
In this paper, we survey loop parallelization algorithms, analyzing the dependence representations t...
This paper presents a set of new run-time tests for speculative parallelization of loops that defy p...
In this paper, we present two new run-time algorithms for the parallelization of loops that have ind...
In this paper we give a new run--time technique for finding an optimal parallel execution schedule f...
The goal of parallelizing, or restructuring, compilers is to detect and exploit parallelism in seque...
Current parallelizing compilers cannot extract a significant fraction of the available parallelism i...
Current parallelizing compilers cannot identify a significant fraction of parallelizable loops becau...
When the inter-iteration dependency pattern of the iterations of a loop cannot be determined statica...
Current parallelizing compilers cannot identify a significant fraction of parallelizable loops becau...
This paper demonstrates that significant improvements to automatic parallelization technology requir...
[[abstract]]A run-time technique based on the inspector-executor scheme is proposed in this paper to...
Current parallelizing compilers cannot extract a significant fraction of the available parallelism i...
While automatic parallelization of loops usually relies on compile-time analysis of data dependences...
Parallelizing compilers promise to exploit the parallelism available in a given program, particularl...
Current parallelizing compilers cannot extract a significant fraction of the available parallelism ...
In this paper, we survey loop parallelization algorithms, analyzing the dependence representations t...
This paper presents a set of new run-time tests for speculative parallelization of loops that defy p...
In this paper, we present two new run-time algorithms for the parallelization of loops that have ind...
In this paper we give a new run--time technique for finding an optimal parallel execution schedule f...
The goal of parallelizing, or restructuring, compilers is to detect and exploit parallelism in seque...
Current parallelizing compilers cannot extract a significant fraction of the available parallelism i...
Current parallelizing compilers cannot identify a significant fraction of parallelizable loops becau...
When the inter-iteration dependency pattern of the iterations of a loop cannot be determined statica...
Current parallelizing compilers cannot identify a significant fraction of parallelizable loops becau...
This paper demonstrates that significant improvements to automatic parallelization technology requir...
[[abstract]]A run-time technique based on the inspector-executor scheme is proposed in this paper to...
Current parallelizing compilers cannot extract a significant fraction of the available parallelism i...
While automatic parallelization of loops usually relies on compile-time analysis of data dependences...
Parallelizing compilers promise to exploit the parallelism available in a given program, particularl...
Current parallelizing compilers cannot extract a significant fraction of the available parallelism ...
In this paper, we survey loop parallelization algorithms, analyzing the dependence representations t...
This paper presents a set of new run-time tests for speculative parallelization of loops that defy p...
In this paper, we present two new run-time algorithms for the parallelization of loops that have ind...