Dependencies between loop iterations cannot always be characterized during program compilation. Doacross loops typically make use of a-priori knowledge of inter-iteration dependencies to carry out required synchronizations. A type of doacross loop is proposed that allows the scheduling of iterations of a loop among processors without advance knowledge of inter-iteration dependencies. The method proposed for loop iterations requires that parallelizable preprocessing and postprocessing steps be carried out during program execution
To reach high performances on massively parallel machines, optimizing compilers have to treat very e...
[[abstract]]A run-time technique based on the inspector-executor scheme is proposed in this paper to...
We present two algorithms to minimize the amount of synchronization added when parallelizing a loop ...
Run time methods are studied to automatically parallelize and schedule iterations of a do loop in ce...
Loops are the main source of parallelism in scientific programs. Hence, several techniques were dev...
While automatic parallelization of loops usually relies on compile-time analysis of data dependences...
Parallelizing compilers promise to exploit the parallelism available in a given program, particularl...
Discussed are how loop level parallelism is detected in a nonprocedural dataflow program, and how a ...
Loops in scientific and engineering applications provide a rich source of parallelism. In order to o...
[[abstract]]The main function of parallelizing compilers is to analyze sequential programs, in parti...
Uncountable loops (such as while loops in C) and if-conditions are some of the most common construct...
This work leverages an original dependency analysis to parallelize loops regardless of their form i...
It is extremely difficult to parallelize DOACROSS loops with non-uniform loop-carried dependences. I...
To exploit parallelism in Fortran code, this dissertation consists of a study of the following three...
Synchronizations are necessary when there are dependences between concurrent processes. However, man...
To reach high performances on massively parallel machines, optimizing compilers have to treat very e...
[[abstract]]A run-time technique based on the inspector-executor scheme is proposed in this paper to...
We present two algorithms to minimize the amount of synchronization added when parallelizing a loop ...
Run time methods are studied to automatically parallelize and schedule iterations of a do loop in ce...
Loops are the main source of parallelism in scientific programs. Hence, several techniques were dev...
While automatic parallelization of loops usually relies on compile-time analysis of data dependences...
Parallelizing compilers promise to exploit the parallelism available in a given program, particularl...
Discussed are how loop level parallelism is detected in a nonprocedural dataflow program, and how a ...
Loops in scientific and engineering applications provide a rich source of parallelism. In order to o...
[[abstract]]The main function of parallelizing compilers is to analyze sequential programs, in parti...
Uncountable loops (such as while loops in C) and if-conditions are some of the most common construct...
This work leverages an original dependency analysis to parallelize loops regardless of their form i...
It is extremely difficult to parallelize DOACROSS loops with non-uniform loop-carried dependences. I...
To exploit parallelism in Fortran code, this dissertation consists of a study of the following three...
Synchronizations are necessary when there are dependences between concurrent processes. However, man...
To reach high performances on massively parallel machines, optimizing compilers have to treat very e...
[[abstract]]A run-time technique based on the inspector-executor scheme is proposed in this paper to...
We present two algorithms to minimize the amount of synchronization added when parallelizing a loop ...