Run-time compilation techniques have been shown effective for automating the parallelization of loops with unstructured indirect data accessing patterns. However, it is still an open problem to efficiently parallelize sparse matrix factorizations commonly used in iterative numerical problems. The difficulty is that a factorization process contains irregularlyinterleaved communication and computation with varying granularities and it is hard to obtain scalable performance on distributed memory machines. In this paper, we present an inspector/executor approach for parallelizing such applications by embodying automatic graph scheduling techniques to optimize interleaved communication and computation. We describe a run-time system called RAPID ...
Factorizing sparse matrices using direct multifrontal methods generates directed tree-shaped task g...
This paper investigates the execution of tree-shaped task graphs using multiple processors. Each edg...
Task graphs are used for scheduling tasks on parallel processors when the tasks have dependencies. I...
International audienceIn this paper, we propose a generic method of automatic parallelization for sp...
Problems in the class of unstructured sparse matrix computations are characterized by highly irregul...
Automatic scheduling for directed acyclic graphs (DAG) and its applications for coarsegrained irregu...
this article we investigate the trade-off between time and space efficiency in scheduling and execut...
Many irregular scientific computing problems can be modeled by directed acyclic task graphs (DAGs). ...
This paper presents a compiler and runtime framework for parallelizing sparse matrix computations th...
Automatic scheduling in parallel/distributed systems for coarse grained irregular problems such as s...
Scientific workloads are often described as directed acyclic task graphs. In this paper, we focus o...
textWe present a methodology for exploiting shared-memory parallelism within matrix computations by ...
Algorithms are often parallelized based on data dependence analysis manually or by means of parallel...
This dissertation advances the state of the art for scalable high-performance graph analytics and da...
Runtime specialization optimizes programs based on partial information available only at run time. I...
Factorizing sparse matrices using direct multifrontal methods generates directed tree-shaped task g...
This paper investigates the execution of tree-shaped task graphs using multiple processors. Each edg...
Task graphs are used for scheduling tasks on parallel processors when the tasks have dependencies. I...
International audienceIn this paper, we propose a generic method of automatic parallelization for sp...
Problems in the class of unstructured sparse matrix computations are characterized by highly irregul...
Automatic scheduling for directed acyclic graphs (DAG) and its applications for coarsegrained irregu...
this article we investigate the trade-off between time and space efficiency in scheduling and execut...
Many irregular scientific computing problems can be modeled by directed acyclic task graphs (DAGs). ...
This paper presents a compiler and runtime framework for parallelizing sparse matrix computations th...
Automatic scheduling in parallel/distributed systems for coarse grained irregular problems such as s...
Scientific workloads are often described as directed acyclic task graphs. In this paper, we focus o...
textWe present a methodology for exploiting shared-memory parallelism within matrix computations by ...
Algorithms are often parallelized based on data dependence analysis manually or by means of parallel...
This dissertation advances the state of the art for scalable high-performance graph analytics and da...
Runtime specialization optimizes programs based on partial information available only at run time. I...
Factorizing sparse matrices using direct multifrontal methods generates directed tree-shaped task g...
This paper investigates the execution of tree-shaped task graphs using multiple processors. Each edg...
Task graphs are used for scheduling tasks on parallel processors when the tasks have dependencies. I...