This paper presents a compiling technique to generate parallel code with explicit local communications for a mesh-connected distributed memory, MIMD architecture. Our compiling technique works for the geometric paradigm of parallel computation, i.e. a data-parallel paradigm where array data structures are partitioned and assigned to a set of processing nodes, which, to perform their identical tasks, need to exchange some of the data allocated to them. This means that some data dependencies exist between computations mapped onto different nodes. Our technique solves them statically, by inserting explicit communications in the parallel code running on each node. We also demonstrate that the technique produces an optimal code as Ear as the exp...
An important research topic is parallelizing of compilers to generate local memory access sequences ...
We discuss some techniques for preserving locality of reference in index spaces when mapped to memor...
This article focuses on principles for the design of efficient parallel algorithms for distributed m...
This paper presents a compiling technique to generate parallel code with explicit local communicatio...
This paper presents a compiling technique to generate parallel code with explicit local communicatio...
The motivation of this thesis was to develop strategies that would enable unstructured mesh based co...
International audienceIn this paper we concentrate on embedded parallel architectures with heterogen...
International audienceIn this paper, we present original techniques for the generation and the effic...
Scientific and engineering applications often involve structured meshes. These meshes may be nested ...
Producción CientíficaCurrent multicomputers are typically built as interconnected clusters of shared...
Programming for parallel architectures that do not have a shared address space is extremely difficul...
Distributed memory multiprocessors are increasingly being used to provide high performance for advan...
Data-parallel languages, such as H scIGH P scERFORMANCE F scORTRAN or F scORTRAN D, provide a machin...
We present new techniques for compilation of arbitrarily nested loops with affine dependences for di...
Parallel computing hardware is affordable and accessible, yet parallel programming is not as widespr...
An important research topic is parallelizing of compilers to generate local memory access sequences ...
We discuss some techniques for preserving locality of reference in index spaces when mapped to memor...
This article focuses on principles for the design of efficient parallel algorithms for distributed m...
This paper presents a compiling technique to generate parallel code with explicit local communicatio...
This paper presents a compiling technique to generate parallel code with explicit local communicatio...
The motivation of this thesis was to develop strategies that would enable unstructured mesh based co...
International audienceIn this paper we concentrate on embedded parallel architectures with heterogen...
International audienceIn this paper, we present original techniques for the generation and the effic...
Scientific and engineering applications often involve structured meshes. These meshes may be nested ...
Producción CientíficaCurrent multicomputers are typically built as interconnected clusters of shared...
Programming for parallel architectures that do not have a shared address space is extremely difficul...
Distributed memory multiprocessors are increasingly being used to provide high performance for advan...
Data-parallel languages, such as H scIGH P scERFORMANCE F scORTRAN or F scORTRAN D, provide a machin...
We present new techniques for compilation of arbitrarily nested loops with affine dependences for di...
Parallel computing hardware is affordable and accessible, yet parallel programming is not as widespr...
An important research topic is parallelizing of compilers to generate local memory access sequences ...
We discuss some techniques for preserving locality of reference in index spaces when mapped to memor...
This article focuses on principles for the design of efficient parallel algorithms for distributed m...