International audienceIn this paper, we present original techniques for the generation and the efficient execution of communication code for parallel loop nests, in the framework of the compilation of HPFf-like languages on distributed memory parallel computers. The problem is studied through its two components: on one hand, the generation of a fast description of communication sets by the compiler and, on the other hand, the implementation of efficient transfers at run-time. Both take into account the characteristics of the distributed array management, notably the memory contiguity
Distributed memory multiprocessors are increasingly being used to provide high performance for advan...
Communication overhead in multiprocessor systems, as exemplified by cache coherency traffic and glob...
Distributed-memory multicomputers, such as the Intel iPSC/860, the Intel Paragon, the IBM SP-1 /SP-2...
International audienceIn this paper, we present original techniques for the generation and the effic...
International audienceHigh Performance Fortran and other similar languages have been designed as a m...
This work was also published as a Rice University thesis/dissertation: http://hdl.handle.net/1911/19...
International audienceHigh Performance Fortran and other similar languages have been designed as a m...
Compilation of parallel loops is one of the most important parts in parallel compilation and optimiz...
[[abstract]]In distributed memory multicomputers, local memory accesses are much faster than those i...
Data-parallel languages allow programmers to use the familiar machine-independent programming style ...
[[abstract]]An increasing number of programming languages, such as Fortran 90, HPF, and APL, provide...
International audienceThis paper presents an environment for programming distributed memory computer...
Data distribution functions are introduced. They are matced with scheduling functions. The processor...
The difficulty of programming distributed memory parallel architectures is an impediment to the expl...
Distributed-memory message-passing machines deliver scalable perfor-mance but are difficult to progr...
Distributed memory multiprocessors are increasingly being used to provide high performance for advan...
Communication overhead in multiprocessor systems, as exemplified by cache coherency traffic and glob...
Distributed-memory multicomputers, such as the Intel iPSC/860, the Intel Paragon, the IBM SP-1 /SP-2...
International audienceIn this paper, we present original techniques for the generation and the effic...
International audienceHigh Performance Fortran and other similar languages have been designed as a m...
This work was also published as a Rice University thesis/dissertation: http://hdl.handle.net/1911/19...
International audienceHigh Performance Fortran and other similar languages have been designed as a m...
Compilation of parallel loops is one of the most important parts in parallel compilation and optimiz...
[[abstract]]In distributed memory multicomputers, local memory accesses are much faster than those i...
Data-parallel languages allow programmers to use the familiar machine-independent programming style ...
[[abstract]]An increasing number of programming languages, such as Fortran 90, HPF, and APL, provide...
International audienceThis paper presents an environment for programming distributed memory computer...
Data distribution functions are introduced. They are matced with scheduling functions. The processor...
The difficulty of programming distributed memory parallel architectures is an impediment to the expl...
Distributed-memory message-passing machines deliver scalable perfor-mance but are difficult to progr...
Distributed memory multiprocessors are increasingly being used to provide high performance for advan...
Communication overhead in multiprocessor systems, as exemplified by cache coherency traffic and glob...
Distributed-memory multicomputers, such as the Intel iPSC/860, the Intel Paragon, the IBM SP-1 /SP-2...