Modern multi-core clusters are increasingly using GPUs to achieve higher performance and power efficiency. In such clusters, efficient communication among processes with data residing in GPU memory is of paramount importance to the performance of MPI applications. This paper investigates the efficient design of intranode MPI Allreduce operation in GPU clusters. We propose two design alternatives that ex-ploit in-GPU reduction and fast intranode communication capabilities of modern GPUs. Our GPU shared-buffer aware design and GPU-aware Binomial reduce-broadcast algorith-mic approach provide significant speedup over MVAPICH2 by up to 22 and 16 times, respectively
GPUs are frequently used to accelerate data-parallel workloads across a wide variety of application ...
Abstract. We present a new, simple algorithmic idea for exploiting the potential for bidirectional c...
We present GPMR, our MapReduce library that leverages the power of GPU clusters for large-scale comp...
Abstract—Current implementations of MPI are unaware of accelerator memory (i.e., GPU device memory) ...
Abstract—Accelerator awareness has become a pressing issue in data movement models, such as MPI, bec...
International audienceHeterogeneous supercomputers are now considered the most valuable solution to ...
Abstract—We present and analyze two new communication libraries, cudaMPI and glMPI, that provide an ...
After the introduction of CUDA by Nvidia, the GPUs became devices capable of accelerating any genera...
Abstract — Modern processors have multiple cores on a chip to overcome power consumption and heat di...
GPUs gain high popularity in High Performance Computing, due to their massive parallelism and high p...
Graphic Processing Units (GPUs) are widely used in high performance computing, due to their high com...
GPUs are widely used in high performance computing, due to their high computational power and high p...
Hybrid nodes containing GPUs are rapidly becoming the norm in parallel machines. We have conducted s...
Today, GPUs and other parallel accelerators are widely used in high performance computing, due to th...
Clusters of several thousand nodes interconnected with InfiniBand, an emerging high-performance inte...
GPUs are frequently used to accelerate data-parallel workloads across a wide variety of application ...
Abstract. We present a new, simple algorithmic idea for exploiting the potential for bidirectional c...
We present GPMR, our MapReduce library that leverages the power of GPU clusters for large-scale comp...
Abstract—Current implementations of MPI are unaware of accelerator memory (i.e., GPU device memory) ...
Abstract—Accelerator awareness has become a pressing issue in data movement models, such as MPI, bec...
International audienceHeterogeneous supercomputers are now considered the most valuable solution to ...
Abstract—We present and analyze two new communication libraries, cudaMPI and glMPI, that provide an ...
After the introduction of CUDA by Nvidia, the GPUs became devices capable of accelerating any genera...
Abstract — Modern processors have multiple cores on a chip to overcome power consumption and heat di...
GPUs gain high popularity in High Performance Computing, due to their massive parallelism and high p...
Graphic Processing Units (GPUs) are widely used in high performance computing, due to their high com...
GPUs are widely used in high performance computing, due to their high computational power and high p...
Hybrid nodes containing GPUs are rapidly becoming the norm in parallel machines. We have conducted s...
Today, GPUs and other parallel accelerators are widely used in high performance computing, due to th...
Clusters of several thousand nodes interconnected with InfiniBand, an emerging high-performance inte...
GPUs are frequently used to accelerate data-parallel workloads across a wide variety of application ...
Abstract. We present a new, simple algorithmic idea for exploiting the potential for bidirectional c...
We present GPMR, our MapReduce library that leverages the power of GPU clusters for large-scale comp...