We present new techniques for compilation of arbitrarily nested loops with affine dependences for distributed-memory parallel architectures. Our framework is implemented as a source-level transformer that uses the polyhedral model, and generates parallel code with communication expressed with the Message Passing Interface (MPI) library. Compared to all previous approaches, ours is a significant advance either (1) with respect to the generality of input code handled, or (2) efficiency of communication code, or both. We provide experimental results on a cluster of multicores demonstrating its effectiveness. In some cases, code we generate outperforms manually parallelized codes, and in another case is within 25% of it. To the best of our know...
International audienceThere may be a huge gap between the statements outlined by programmers in a pr...
Programming for parallel architectures that do not have a shared address space is extremely difficul...
International audienceMany automatic software parallelization systems have been proposed in the past...
2013 Spring.Includes bibliographical references.With the introduction of multi-core processors, moti...
Code generation and programming have become ever more challenging over the last decade due to the sh...
International audienceIn this paper, we present original techniques for the generation and the effic...
Minimizing communication overhead when mapping affine loop nests onto distributed memory parallel co...
The polyhedral model is known to be a powerful framework to reason about high level loop transformat...
International audienceAffine transformations have proven to be powerful for loop restructuring due t...
International audienceWe propose a framework based on an original generation and use of algorithmic ...
Abstract In this paper, an approach to the problem of exploiting parallelism within nested loops is ...
Affine transformations have proven to be powerful for loop restructuring due to their ability to mod...
Executing a program in parallel machines needs not only to find sufficient parallelism in a program,...
Supercompilers perform complex program transformations which often result in new loop bounds. This p...
Recent advances in polyhedral compilation technology have made it feasible to automatically transfor...
International audienceThere may be a huge gap between the statements outlined by programmers in a pr...
Programming for parallel architectures that do not have a shared address space is extremely difficul...
International audienceMany automatic software parallelization systems have been proposed in the past...
2013 Spring.Includes bibliographical references.With the introduction of multi-core processors, moti...
Code generation and programming have become ever more challenging over the last decade due to the sh...
International audienceIn this paper, we present original techniques for the generation and the effic...
Minimizing communication overhead when mapping affine loop nests onto distributed memory parallel co...
The polyhedral model is known to be a powerful framework to reason about high level loop transformat...
International audienceAffine transformations have proven to be powerful for loop restructuring due t...
International audienceWe propose a framework based on an original generation and use of algorithmic ...
Abstract In this paper, an approach to the problem of exploiting parallelism within nested loops is ...
Affine transformations have proven to be powerful for loop restructuring due to their ability to mod...
Executing a program in parallel machines needs not only to find sufficient parallelism in a program,...
Supercompilers perform complex program transformations which often result in new loop bounds. This p...
Recent advances in polyhedral compilation technology have made it feasible to automatically transfor...
International audienceThere may be a huge gap between the statements outlined by programmers in a pr...
Programming for parallel architectures that do not have a shared address space is extremely difficul...
International audienceMany automatic software parallelization systems have been proposed in the past...