In the MPI parallel programming model, communication remains the bottleneck that prevents applications to achieve greater performance and scalability. Due to this problem it is important to know the behavior of this communication in each application. For this reason, we propose the use of community detection algorithms to identify from the communication matrix the clusters of ranks that maximize intracluster communication and minimize intercluster communication. The aim of this project is providing another tool to identify how you can improve the performance of a MPI application
The performance of MPI implementation operations still presents critical issues for high performance...
The emergence of meta computers and computational grids makes it feasible to run parallel programs o...
We present a technique for performance analysis that helps users understand the communication behavi...
A parallelization based on MPI and OpenMP of an algorithm that evaluates and counts all the possible...
Collective communications occupy 20-90% of total execution times in many MPI applications. In this p...
Abstract Many parallel applications from scientific computing use collective MPI communication oper-...
In order for collective communication routines to achieve high performance on different platforms, t...
Previous studies of application usage show that the per-formance of collective communications are cr...
Many parallel applications from scientific computing use MPI collective communication operations to ...
This work presents and evaluates algorithms for MPI collective communication operations on high perf...
The performance of collective communication operations is one of the deciding factors in the overa...
Analyzing parallel programs has become increasingly difficult due to the immense amount of informati...
Many parallel applications from scientific computing use MPI collective communication operations to ...
There are various community detection algorithms which that have been developed. Among them, Louvain...
Further performance improvements of parallel simulation applications will not be reached by simply s...
The performance of MPI implementation operations still presents critical issues for high performance...
The emergence of meta computers and computational grids makes it feasible to run parallel programs o...
We present a technique for performance analysis that helps users understand the communication behavi...
A parallelization based on MPI and OpenMP of an algorithm that evaluates and counts all the possible...
Collective communications occupy 20-90% of total execution times in many MPI applications. In this p...
Abstract Many parallel applications from scientific computing use collective MPI communication oper-...
In order for collective communication routines to achieve high performance on different platforms, t...
Previous studies of application usage show that the per-formance of collective communications are cr...
Many parallel applications from scientific computing use MPI collective communication operations to ...
This work presents and evaluates algorithms for MPI collective communication operations on high perf...
The performance of collective communication operations is one of the deciding factors in the overa...
Analyzing parallel programs has become increasingly difficult due to the immense amount of informati...
Many parallel applications from scientific computing use MPI collective communication operations to ...
There are various community detection algorithms which that have been developed. Among them, Louvain...
Further performance improvements of parallel simulation applications will not be reached by simply s...
The performance of MPI implementation operations still presents critical issues for high performance...
The emergence of meta computers and computational grids makes it feasible to run parallel programs o...
We present a technique for performance analysis that helps users understand the communication behavi...