Governments, universities, and companies expend vast resources building the top supercomputers. The processors and interconnect networks become faster, while the number of nodes grows exponentially. Problems of scale emerge, not least of which is collective performance. This thesis identifies and proposes solutions for two major scalability problems. Our first contribution is a novel algorithm for process-partitioning and remapping for exascale systems that has far better time and space scaling than known algorithms. Our evaluations predict an improvement of up to 60x for large exascale systems and arbitrary reduction in the large temporary buffer space required for generating new communicators. Our second contribution consist...
Graph Convolutional Networks (GCNs) are extensively utilized for deep learning on graphs. The large ...
Proceedings of: First International Workshop on Sustainable Ultrascale Computing Systems (NESUS 2014...
This is a post-peer-review, pre-copyedit version of an article published in [insert journal title]. ...
Governments, universities, and companies expend vast resources building the top supercomputers. The...
Supercomputers continue to expand both in size and complexity as we reach the beginning of the exasc...
The current trends in high performance computing show that large machines with tens of thousands of ...
Many modern services need to routinely perform tasks on a large scale. This prompts us to consider t...
Artículo de publicación ISIIn this paper we study distributed algorithms on massive graphs where li...
This work presents and evaluates algorithms for MPI collective communication operations on high perf...
Collective communications occupy 20-90% of total execution times in many MPI applications. In this p...
pre-printThe placement of tasks in a parallel application on specific nodes of a supercomputer can s...
Sparse matrix operations dominate the cost of many scientific applications. In parallel, the perform...
The divergence of application behavior from optimal network usage leads to performance bottlenecks i...
We report on a project to develop a unified approach for building a library of collective communicat...
Technology trends suggest that future machines will relyon parallelism to meet increasing performanc...
Graph Convolutional Networks (GCNs) are extensively utilized for deep learning on graphs. The large ...
Proceedings of: First International Workshop on Sustainable Ultrascale Computing Systems (NESUS 2014...
This is a post-peer-review, pre-copyedit version of an article published in [insert journal title]. ...
Governments, universities, and companies expend vast resources building the top supercomputers. The...
Supercomputers continue to expand both in size and complexity as we reach the beginning of the exasc...
The current trends in high performance computing show that large machines with tens of thousands of ...
Many modern services need to routinely perform tasks on a large scale. This prompts us to consider t...
Artículo de publicación ISIIn this paper we study distributed algorithms on massive graphs where li...
This work presents and evaluates algorithms for MPI collective communication operations on high perf...
Collective communications occupy 20-90% of total execution times in many MPI applications. In this p...
pre-printThe placement of tasks in a parallel application on specific nodes of a supercomputer can s...
Sparse matrix operations dominate the cost of many scientific applications. In parallel, the perform...
The divergence of application behavior from optimal network usage leads to performance bottlenecks i...
We report on a project to develop a unified approach for building a library of collective communicat...
Technology trends suggest that future machines will relyon parallelism to meet increasing performanc...
Graph Convolutional Networks (GCNs) are extensively utilized for deep learning on graphs. The large ...
Proceedings of: First International Workshop on Sustainable Ultrascale Computing Systems (NESUS 2014...
This is a post-peer-review, pre-copyedit version of an article published in [insert journal title]. ...