This artifact accompanies the paper "Parallel Block-Delayed Sequences" at PPoPP'22. The paper presents a library-only technique for fusing collection-oriented operations (map, zip, filter, flatten, scan, etc.) which improves both run-time and space usage by avoiding unnecessary intermediate allocations. This artifact provides source code for libraries in both C++ and Parallel ML, implementing block-delayed sequences. It also includes scripts to run experiments and reproduce the results in the paper, supporting our claims that block-delayed sequences provide significant improvements in both time and space
This paper considers the implementation of a two-point diagonally implicit block method for solving ...
Parallelizing compiler technology has improved in re-cent years. One area in which compilers have ma...
Multicore architec- tures are increasingly common, but real architectures have unpredictable timing ...
This artifact contains the benchmarks for the loop fission algorithm presented in paper "Distributin...
This dissertation presents two new developments in the area of computer program preparation for para...
The PC-set method and the parallel technique are two methods for performing compiled unit-delay simu...
Block-processing is a powerful and popular technique for increasing computation speed by simultaneou...
This artifact is intended to demonstrate the workflow followed by our simulation approach for at-sca...
Accesses to shared resources in multi-core systems raise predictability issues. The delay in accessi...
This artifact is concerned with Section 5 (Evaluation) of the paper "Accelerating Nested Data Parall...
International audienceAccesses to shared resources in multi-core systems raise predictability issues...
Parallel computing hardware is ubiquitous, ranging from cell-phones with multiple cores to super-com...
The PC-set method and the parallel technique are two methods for generating compiled unit-delay simu...
Abstract—Dask enables parallel and out-of-core computation. We couple blocked algorithms with dynami...
The PC-set method and the parallel technique are two methods for generating compiled unit-delay simu...
This paper considers the implementation of a two-point diagonally implicit block method for solving ...
Parallelizing compiler technology has improved in re-cent years. One area in which compilers have ma...
Multicore architec- tures are increasingly common, but real architectures have unpredictable timing ...
This artifact contains the benchmarks for the loop fission algorithm presented in paper "Distributin...
This dissertation presents two new developments in the area of computer program preparation for para...
The PC-set method and the parallel technique are two methods for performing compiled unit-delay simu...
Block-processing is a powerful and popular technique for increasing computation speed by simultaneou...
This artifact is intended to demonstrate the workflow followed by our simulation approach for at-sca...
Accesses to shared resources in multi-core systems raise predictability issues. The delay in accessi...
This artifact is concerned with Section 5 (Evaluation) of the paper "Accelerating Nested Data Parall...
International audienceAccesses to shared resources in multi-core systems raise predictability issues...
Parallel computing hardware is ubiquitous, ranging from cell-phones with multiple cores to super-com...
The PC-set method and the parallel technique are two methods for generating compiled unit-delay simu...
Abstract—Dask enables parallel and out-of-core computation. We couple blocked algorithms with dynami...
The PC-set method and the parallel technique are two methods for generating compiled unit-delay simu...
This paper considers the implementation of a two-point diagonally implicit block method for solving ...
Parallelizing compiler technology has improved in re-cent years. One area in which compilers have ma...
Multicore architec- tures are increasingly common, but real architectures have unpredictable timing ...