This paper explores the challenges in implementing a message passing interface usable on systems with data-parallel processors. As a case study, we design and implement the ''DCGN'' API on NVIDIA GPUs that is similar to MPI and allows full access to the underlying architecture. We introduce the notion of data-parallel thread-groups as a way to map resources to MPI ranks. We use a method that also allows the data-parallel processors to run autonomously from user-written CPU code. In order to facilitate communication, we use a sleep-based polling system to store and retrieve messages. Unlike previous systems, our method provides both performance and flexibility. By running a test suite of applications with different communication requirements...
Current trends in computing and system architecture point towards a need for accelerators such as GP...
Data-parallel languages such as High Performance Fortran (HPF) present a simple execution model in w...
The Message Passing Interface (MPI) has been extremely successful as a portable way to program high-...
This paper explores the challenges in implementing a message passing interface usable on systems wit...
Abstract—We present and analyze two new communication libraries, cudaMPI and glMPI, that provide an ...
Communication hardware and software have a significant impact on the performance of clusters and sup...
Parallel computing on clusters of workstations and personal computers has very high potential, since...
The Message-Passing Interface (MPI) is a widely-used standard library for programming parallel appli...
Parallel computing on clusters of workstations and personal computers has very high potential, sinc...
Message Passing Interface[2] is the de facto standard for multicomputer and cluster message passing;...
The Message Passing Interface (MPI) is a standard in parallel computing, and can also be used as a h...
Shared-memory and message-passing are two op- posite models to develop parallel computations. The sh...
Novel reconfigurable computing platforms enable effi-cient realizations of complex signal processing...
In this paper, we describe the design and implementation of a datamovement and control substrate (DM...
Abstract—Data movement in high-performance computing systems accelerated by graphics processing unit...
Current trends in computing and system architecture point towards a need for accelerators such as GP...
Data-parallel languages such as High Performance Fortran (HPF) present a simple execution model in w...
The Message Passing Interface (MPI) has been extremely successful as a portable way to program high-...
This paper explores the challenges in implementing a message passing interface usable on systems wit...
Abstract—We present and analyze two new communication libraries, cudaMPI and glMPI, that provide an ...
Communication hardware and software have a significant impact on the performance of clusters and sup...
Parallel computing on clusters of workstations and personal computers has very high potential, since...
The Message-Passing Interface (MPI) is a widely-used standard library for programming parallel appli...
Parallel computing on clusters of workstations and personal computers has very high potential, sinc...
Message Passing Interface[2] is the de facto standard for multicomputer and cluster message passing;...
The Message Passing Interface (MPI) is a standard in parallel computing, and can also be used as a h...
Shared-memory and message-passing are two op- posite models to develop parallel computations. The sh...
Novel reconfigurable computing platforms enable effi-cient realizations of complex signal processing...
In this paper, we describe the design and implementation of a datamovement and control substrate (DM...
Abstract—Data movement in high-performance computing systems accelerated by graphics processing unit...
Current trends in computing and system architecture point towards a need for accelerators such as GP...
Data-parallel languages such as High Performance Fortran (HPF) present a simple execution model in w...
The Message Passing Interface (MPI) has been extremely successful as a portable way to program high-...