The current MPI model defines a one-to-one relationship between MPI processes and MPI ranks. This model captures many use cases effectively, such as one MPI process per core and one MPI process per node. However, this semantic has limited interoperability be-tween MPI and other programming models that use threads within a node. In this paper, we describe an extension to MPI that introduces communication endpoints as a means to relax the one-to-one rela-tionship between processes and threads. Endpoints enable a greater degree interoperability between MPI and other programming mod-els, and we illustrate their potential for additional performance and computation management benefits through the decoupling of ranks from processes
Abstract—With the increasing prominence of many-core archi-tectures and decreasing per-core resource...
The MPI standard provides tool builders with an efficient profiling interface, PMPI. Although many t...
This paper addresses performance portability of MPI code on multiprogrammed shared memory machines. ...
Hybrid MPI+threads programming is gaining prominence as an alternative to the traditional "MPI every...
Abstract—Modern high-speed interconnection networks are designed with capabilities to support commun...
We propose extensions to the Message-Passing Interface (MPI) Standard that provide for dynamic proce...
We propose extensions to the Message Passing In-terface (MPI) that generalize the MPI communicator c...
This paper describes current activities of the MPI-2 Forum. The MPI - 2 Forum is a group of parallel...
Message Passing Interface (MPI), as an effort to unify message passing systems to achieve portabilit...
The need for intuitive parallel programming designs has grown with the rise of modern many-core proc...
Hybrid MPI+Threads programming has emerged as an alternative model to the “MPI everywhere ” model to...
The new MPI 4.0 standard includes a new chapter about partitioned point-to-point communication opera...
Many-core architectures, such as the Intel Xeon Phi, provide dozens of cores and hundreds of hardwar...
This paper describes MICE, a prototype implementation of MPI on the Converse interoperable parallel ...
Supercomputing applications rely on strong scaling to achieve faster results on a larger number of p...
Abstract—With the increasing prominence of many-core archi-tectures and decreasing per-core resource...
The MPI standard provides tool builders with an efficient profiling interface, PMPI. Although many t...
This paper addresses performance portability of MPI code on multiprogrammed shared memory machines. ...
Hybrid MPI+threads programming is gaining prominence as an alternative to the traditional "MPI every...
Abstract—Modern high-speed interconnection networks are designed with capabilities to support commun...
We propose extensions to the Message-Passing Interface (MPI) Standard that provide for dynamic proce...
We propose extensions to the Message Passing In-terface (MPI) that generalize the MPI communicator c...
This paper describes current activities of the MPI-2 Forum. The MPI - 2 Forum is a group of parallel...
Message Passing Interface (MPI), as an effort to unify message passing systems to achieve portabilit...
The need for intuitive parallel programming designs has grown with the rise of modern many-core proc...
Hybrid MPI+Threads programming has emerged as an alternative model to the “MPI everywhere ” model to...
The new MPI 4.0 standard includes a new chapter about partitioned point-to-point communication opera...
Many-core architectures, such as the Intel Xeon Phi, provide dozens of cores and hundreds of hardwar...
This paper describes MICE, a prototype implementation of MPI on the Converse interoperable parallel ...
Supercomputing applications rely on strong scaling to achieve faster results on a larger number of p...
Abstract—With the increasing prominence of many-core archi-tectures and decreasing per-core resource...
The MPI standard provides tool builders with an efficient profiling interface, PMPI. Although many t...
This paper addresses performance portability of MPI code on multiprogrammed shared memory machines. ...