Abstract—The data-driven task parallelism execution model can support parallel programming models that are well suited for large-scale distributed-memory parallel computing, for example, simulations and analysis pipelines running on clusters and clouds. We describe a novel compiler intermediate representation and optimizations for this execution model, including adaptions of standard techniques alongside novel techniques. These techniques are applied to Swift/T, a high-level scripting language for flexible data flow composition of functions, which may be serial or use lower-level parallel programming models such as MPI and OpenMP. This paper presents preliminary results, indicating that our compiler optimizations reduce communication overhe...
This paper describes methods to adapt existing optimizing compilers for sequential languages to prod...
Achieving high performance in task-parallel runtime systems, especially with high degrees of paralle...
Despite the performance potential of parallel systems, several factors have hindered their widesprea...
Abstract—Swift/T is a high-level language for writing concise, deterministic scripts that compose se...
Abstract—Many scientific applications are conceptually built up from independent component tasks as ...
It has become common knowledge that parallel programming is needed for scientific applications, part...
International audienceThis paper presents a technique for representing the high level semantics of p...
In this paper, we propose and evaluate practical, automatic techniques that exploit compiler analysi...
Distributed Memory Multicomputers (DMMs) such as the IBM SP-2, the Intel Paragon and the Thinking Ma...
For a wide variety of applications, both task and data parallelism must be exploited to achieve the ...
In this paper we analyze the effect of compiler optimizations on fine grain parallelism in scalar pr...
Achieving high performance in task-parallel runtime systems, especially with high degrees of paralle...
Over the past few decades, scientific research has grown to rely increasingly on simulation and othe...
As the demand increases for high performance and power efficiency in modern computer runtime systems...
Modern parallel programming models perform their best under the particular patterns they are tuned t...
This paper describes methods to adapt existing optimizing compilers for sequential languages to prod...
Achieving high performance in task-parallel runtime systems, especially with high degrees of paralle...
Despite the performance potential of parallel systems, several factors have hindered their widesprea...
Abstract—Swift/T is a high-level language for writing concise, deterministic scripts that compose se...
Abstract—Many scientific applications are conceptually built up from independent component tasks as ...
It has become common knowledge that parallel programming is needed for scientific applications, part...
International audienceThis paper presents a technique for representing the high level semantics of p...
In this paper, we propose and evaluate practical, automatic techniques that exploit compiler analysi...
Distributed Memory Multicomputers (DMMs) such as the IBM SP-2, the Intel Paragon and the Thinking Ma...
For a wide variety of applications, both task and data parallelism must be exploited to achieve the ...
In this paper we analyze the effect of compiler optimizations on fine grain parallelism in scalar pr...
Achieving high performance in task-parallel runtime systems, especially with high degrees of paralle...
Over the past few decades, scientific research has grown to rely increasingly on simulation and othe...
As the demand increases for high performance and power efficiency in modern computer runtime systems...
Modern parallel programming models perform their best under the particular patterns they are tuned t...
This paper describes methods to adapt existing optimizing compilers for sequential languages to prod...
Achieving high performance in task-parallel runtime systems, especially with high degrees of paralle...
Despite the performance potential of parallel systems, several factors have hindered their widesprea...