Current trends in computing and system architecture point towards a need for accelerators such as GPUs to have inherent communication capabilities. We review previous and current software libraries that provide pseudo-communication abilities through direct message passing. We show how these libraries are beneficial to the HPC community, but are not forward-thinking enough. We give motivation as to why MPI should be extended to support these accelerators, and provide a road map of achievable milestones to complete such an extension, some of which require advances in hardware and device drivers
A number of efforts have been undertaken to integrate GPU functionality into an HPC environment, wit...
Recently MPI implementations have been extended to support accelerator devices, Intel Many Integrate...
Modern HPC platforms are using multiple CPU, GPUs and high-performance interconnects per node. Unfor...
Current trends in computing and system architecture point towards a need for accelerators such as GP...
Abstract—Data movement in high-performance computing systems accelerated by graphics processing unit...
Abstract—Current implementations of MPI are unaware of accelerator memory (i.e., GPU device memory) ...
In High Performance Computing (HPC), minimizing communication overhead is one of the most important ...
Abstract—We present and analyze two new communication libraries, cudaMPI and glMPI, that provide an ...
The Message-Passing Interface (MPI) is a widely-used standard library for programming parallel appli...
Abstract—Accelerator awareness has become a pressing issue in data movement models, such as MPI, bec...
MPI is the new de facto standard for multicomputer and cluster message passing. This short paper des...
The Message Passing Interface (MPI) has been extremely successful as a portable way to program high...
Modern HPC platforms are using multiple CPU, GPUs and high-performance interconnects per node. Unfor...
The Message Passing Interface (MPI) has been extremely successful as a portable way to program high-...
International audienceOverlapping communications with computation is an efficient way to amortize th...
A number of efforts have been undertaken to integrate GPU functionality into an HPC environment, wit...
Recently MPI implementations have been extended to support accelerator devices, Intel Many Integrate...
Modern HPC platforms are using multiple CPU, GPUs and high-performance interconnects per node. Unfor...
Current trends in computing and system architecture point towards a need for accelerators such as GP...
Abstract—Data movement in high-performance computing systems accelerated by graphics processing unit...
Abstract—Current implementations of MPI are unaware of accelerator memory (i.e., GPU device memory) ...
In High Performance Computing (HPC), minimizing communication overhead is one of the most important ...
Abstract—We present and analyze two new communication libraries, cudaMPI and glMPI, that provide an ...
The Message-Passing Interface (MPI) is a widely-used standard library for programming parallel appli...
Abstract—Accelerator awareness has become a pressing issue in data movement models, such as MPI, bec...
MPI is the new de facto standard for multicomputer and cluster message passing. This short paper des...
The Message Passing Interface (MPI) has been extremely successful as a portable way to program high...
Modern HPC platforms are using multiple CPU, GPUs and high-performance interconnects per node. Unfor...
The Message Passing Interface (MPI) has been extremely successful as a portable way to program high-...
International audienceOverlapping communications with computation is an efficient way to amortize th...
A number of efforts have been undertaken to integrate GPU functionality into an HPC environment, wit...
Recently MPI implementations have been extended to support accelerator devices, Intel Many Integrate...
Modern HPC platforms are using multiple CPU, GPUs and high-performance interconnects per node. Unfor...