We discuss the design and high-performance implementation of collective communications operations on distributed-memory computer architectures. Using a combination of known techniques (many of which were first proposed in the 1980s and early 1990s) along with careful exploitation of communication modes supported by MPI, we have developed implementations that have improved performance in most situations compared to those currently supported by public domain implementations of MPI such as MPICH. Performance results from a large Intel Xeon/Pentium 4 (R) processor cluster are included.
Parallel computing on clusters of workstations and personal computers has very high potential, since...
The performance of MPI implementation operations still presents critical issues for high performance...
Parallel computing on clusters of workstations and personal computers has very high potential, sinc...
We have implemented eight of the MPI collective routines using MPI point-to-point communication rou...
Collective communication is an important subset of Message Passing Interface. Improving the perform...
Collective communication is an important subset of Message Passing Interface. Improving the perform...
Collective communication allows efficient communication and synchronization among a collection of pr...
Abstract Many parallel applications from scientific computing use collective MPI communication oper-...
Massively parallel computers (MPC) are characterized by the distribution of memory among an ensemble...
This paper describes a novel methodology for implementing a common set of collective communication o...
Previous studies of application usage show that the per-formance of collective communications are cr...
We evaluate the architectural support of collective communication operations on the IBM SP2, Cray T3...
In order for collective communication routines to achieve high performance on different platforms, t...
This work presents and evaluates algorithms for MPI collective communication operations on high perf...
: The significance of collective communication operations for scalable parallel systems has been wel...
Parallel computing on clusters of workstations and personal computers has very high potential, since...
The performance of MPI implementation operations still presents critical issues for high performance...
Parallel computing on clusters of workstations and personal computers has very high potential, sinc...
We have implemented eight of the MPI collective routines using MPI point-to-point communication rou...
Collective communication is an important subset of Message Passing Interface. Improving the perform...
Collective communication is an important subset of Message Passing Interface. Improving the perform...
Collective communication allows efficient communication and synchronization among a collection of pr...
Abstract Many parallel applications from scientific computing use collective MPI communication oper-...
Massively parallel computers (MPC) are characterized by the distribution of memory among an ensemble...
This paper describes a novel methodology for implementing a common set of collective communication o...
Previous studies of application usage show that the per-formance of collective communications are cr...
We evaluate the architectural support of collective communication operations on the IBM SP2, Cray T3...
In order for collective communication routines to achieve high performance on different platforms, t...
This work presents and evaluates algorithms for MPI collective communication operations on high perf...
: The significance of collective communication operations for scalable parallel systems has been wel...
Parallel computing on clusters of workstations and personal computers has very high potential, since...
The performance of MPI implementation operations still presents critical issues for high performance...
Parallel computing on clusters of workstations and personal computers has very high potential, sinc...