Reducing communication overhead is crucial for improving the performance of programs on distributed-memory machines. Compilers for data-parallel languages must perform communication optimizations in order to minimize this overhead. In this paper, we show how to combine dependence analysis, traditionally used to optimize regular communication, and a data-flow analysis method originally developed to improve placement of irregular communication. Our approach allows us to perform more extensive optimizations --- message vectorization, elimination of redundant messages, and overlapping communication with computation. We also present preliminary experimental results that demonstrate the benefits of the proposed method. 1 Introduction Distributed...
Most of the current compiler projects for distributed memory architectures leave the critical and ti...
Irregular applications pose challenges in optimizing communication, due to the difficulty of analyzi...
The topic of intermediate languages for optimizing and parallelizing compilers has received much at...
Data-parallel languages allow programmers to use the familiar machine-independent programming style ...
This work was also published as a Rice University thesis/dissertation: http://hdl.handle.net/1911/19...
Reducing communication overhead is extremely important in distributed-memory message-passing archite...
In distributed-memory message passing architectures reducing communication cost is extremely importa...
Reducing communication overhead is extremely important in distributed-memory message-passing archite...
This paper describes a number of optimizations that can be used to support the efficient execution o...
The performance of a High Performance Parallel or Distributed Computation depends heavily on minimiz...
The increasing attention toward distributed shared memory systems attests to the fact that programme...
The computational speed of individual processors in distributed memory computers is increasing faste...
Communication coalescing is a static optimization that can reduce both communication frequency and r...
Distributed-memory multicomputers, such as the Intel iPSC/860, the Intel Paragon, the IBM SP-1 /SP-2...
Data parallel languages like High Performance Fortran (HPF) are emerging as the architecture indepen...
Most of the current compiler projects for distributed memory architectures leave the critical and ti...
Irregular applications pose challenges in optimizing communication, due to the difficulty of analyzi...
The topic of intermediate languages for optimizing and parallelizing compilers has received much at...
Data-parallel languages allow programmers to use the familiar machine-independent programming style ...
This work was also published as a Rice University thesis/dissertation: http://hdl.handle.net/1911/19...
Reducing communication overhead is extremely important in distributed-memory message-passing archite...
In distributed-memory message passing architectures reducing communication cost is extremely importa...
Reducing communication overhead is extremely important in distributed-memory message-passing archite...
This paper describes a number of optimizations that can be used to support the efficient execution o...
The performance of a High Performance Parallel or Distributed Computation depends heavily on minimiz...
The increasing attention toward distributed shared memory systems attests to the fact that programme...
The computational speed of individual processors in distributed memory computers is increasing faste...
Communication coalescing is a static optimization that can reduce both communication frequency and r...
Distributed-memory multicomputers, such as the Intel iPSC/860, the Intel Paragon, the IBM SP-1 /SP-2...
Data parallel languages like High Performance Fortran (HPF) are emerging as the architecture indepen...
Most of the current compiler projects for distributed memory architectures leave the critical and ti...
Irregular applications pose challenges in optimizing communication, due to the difficulty of analyzi...
The topic of intermediate languages for optimizing and parallelizing compilers has received much at...