Collective operations are among the most important communication operations in shared- and distributed-memory parallel applications. In this paper, we analyze the tradeoffs between energy, mem-ory, and runtime of different algorithms that implement such operations. We show that existing algo-rithms have varying behavior and that no known algorithm is optimal in all three regards. We als
Developing efficient programs for distributed systems is difficult because computations must be effi...
We discuss the design and high-performance implementation of collective communications operations on...
Accelerated computing has become pervasive for increasing the computational power and energy efficie...
In this paper we present several algorithms for performing all-to-many personalized communication on...
Collective communication allows efficient communication and synchronization among a collection of pr...
Abstract Many parallel applications from scientific computing use collective MPI communication oper-...
This paper describes a number of optimizations that can be used to support the efficient execution o...
Collective operations are common features of parallel programming models that are frequently used in...
: The significance of collective communication operations for scalable parallel systems has been wel...
Energy consumption by computer systems has emerged as an important concern, both at the level of ind...
This talk discusses optimized collective algorithms and the benefits of leveraging independent hardw...
Many parallel applications from scientific computing use MPI collective communication operations to ...
Collective communications occupy 20-90% of total execution times in many MPI applications. In this p...
Abstract. In this paper we present the design, implementation and evaluation of a runtime system bas...
PVM and other distributed computing systems have enabled the use of networks of workstations for par...
Developing efficient programs for distributed systems is difficult because computations must be effi...
We discuss the design and high-performance implementation of collective communications operations on...
Accelerated computing has become pervasive for increasing the computational power and energy efficie...
In this paper we present several algorithms for performing all-to-many personalized communication on...
Collective communication allows efficient communication and synchronization among a collection of pr...
Abstract Many parallel applications from scientific computing use collective MPI communication oper-...
This paper describes a number of optimizations that can be used to support the efficient execution o...
Collective operations are common features of parallel programming models that are frequently used in...
: The significance of collective communication operations for scalable parallel systems has been wel...
Energy consumption by computer systems has emerged as an important concern, both at the level of ind...
This talk discusses optimized collective algorithms and the benefits of leveraging independent hardw...
Many parallel applications from scientific computing use MPI collective communication operations to ...
Collective communications occupy 20-90% of total execution times in many MPI applications. In this p...
Abstract. In this paper we present the design, implementation and evaluation of a runtime system bas...
PVM and other distributed computing systems have enabled the use of networks of workstations for par...
Developing efficient programs for distributed systems is difficult because computations must be effi...
We discuss the design and high-performance implementation of collective communications operations on...
Accelerated computing has become pervasive for increasing the computational power and energy efficie...