Previous studies of application usage show that the per-formance of collective communications are critical for high-performance computing and are often overlooked when compared to the point-to-point performance. In this paper, we analyze and attempt to improve intra-cluster col-lective communication in the context of the widely deployed MPI programming paradigm by extending accepted mod-els of point-to-point communication, such as Hockney, LogP/LogGP, and PLogP. The predictions from the mod-els were compared to the experimentally gathered data and our findings were used to optimize the implementa-tion of collective operations in the FT-MPI library. 1
We discuss the design and high-performance implementation of collective communications operations on...
Further performance improvements of parallel simulation applications will not be reached by simply s...
This work presents and evaluates algorithms for MPI collective communication operations on high perf...
Many parallel applications from scientific computing use MPI collective communication operations to ...
Collective communications occupy 20-90% of total execution times in many MPI applications. In this p...
Collective communications occupy 20-90% of total execution times in many MPI applications. In this p...
In order for collective communication routines to achieve high performance on different platforms, t...
The performance of collective communication operations is one of the deciding factors in the overa...
The performance of collective communication operations is one of the deciding factors in the overa...
Abstract Many parallel applications from scientific computing use collective MPI communication oper-...
Many parallel applications from scientific computing use MPI collective communication operations to ...
The performance of MPI implementation operations still presents critical issues for high performance...
Collective communication is an important subset of Message Passing Interface. Improving the perform...
Collective communication is an important subset of Message Passing Interface. Improving the perform...
We have implemented eight of the MPI collective routines using MPI point-to-point communication rou...
We discuss the design and high-performance implementation of collective communications operations on...
Further performance improvements of parallel simulation applications will not be reached by simply s...
This work presents and evaluates algorithms for MPI collective communication operations on high perf...
Many parallel applications from scientific computing use MPI collective communication operations to ...
Collective communications occupy 20-90% of total execution times in many MPI applications. In this p...
Collective communications occupy 20-90% of total execution times in many MPI applications. In this p...
In order for collective communication routines to achieve high performance on different platforms, t...
The performance of collective communication operations is one of the deciding factors in the overa...
The performance of collective communication operations is one of the deciding factors in the overa...
Abstract Many parallel applications from scientific computing use collective MPI communication oper-...
Many parallel applications from scientific computing use MPI collective communication operations to ...
The performance of MPI implementation operations still presents critical issues for high performance...
Collective communication is an important subset of Message Passing Interface. Improving the perform...
Collective communication is an important subset of Message Passing Interface. Improving the perform...
We have implemented eight of the MPI collective routines using MPI point-to-point communication rou...
We discuss the design and high-performance implementation of collective communications operations on...
Further performance improvements of parallel simulation applications will not be reached by simply s...
This work presents and evaluates algorithms for MPI collective communication operations on high perf...