On current multiprocessor architectures one must carefully distribute data in memory in order to achieve high performance. Process partitioning is the operation of rewriting an algorithm as a collection of tasks, each operating primarily on its own portion of the data, to carry out the computation in parallel. A semi-automatic approach to process partitioning is considered in which the compiler, guided by advice from the user, automatically transforms programs into such an interacting task system. This approach is illustrated with a picture processing example written in BLAZE, which is transformed into a task system maximizing locality of memory reference
Many real-time computations such as process control and robotic applications may be naturally distri...
This paper outlines two methods which we believe will play an important role in any distributed memo...
We describe an approach to parallel compilation that seeks to harness the vast amount of fine-grain ...
The general problem studied is that of segmenting or partitioning programs for distribution across a...
Distributed-memory parallel computers are increasingly being used to provide high levels of performa...
grantor: University of TorontoScalable shared memory multiprocessors are becoming increasi...
εm is a high-level programming system which puts parallelism within the reach of scientists who are ...
There is a trend towards using accelerators to increase performance and energy efficiency of general...
Outlined here are two methods which we believe will play an important role in any distributed memory...
This work was also published as a Rice University thesis/dissertation: http://hdl.handle.net/1911/16...
The goal of this dissertation is to give programmers the ability to achieve high performance by focu...
For distributed memory multicomputers such as the Intel Paragon, the IBM SP-2, the NCUBE/2, and the ...
In order to utilize parallel computers, four approaches, broadly speaking, to the provision of paral...
Coordinated Science Laboratory was formerly known as Control Systems LaboratoryOffice of Naval Resea...
The main contribution of the effort in the last two years is the introduction of the MOPPS system. A...
Many real-time computations such as process control and robotic applications may be naturally distri...
This paper outlines two methods which we believe will play an important role in any distributed memo...
We describe an approach to parallel compilation that seeks to harness the vast amount of fine-grain ...
The general problem studied is that of segmenting or partitioning programs for distribution across a...
Distributed-memory parallel computers are increasingly being used to provide high levels of performa...
grantor: University of TorontoScalable shared memory multiprocessors are becoming increasi...
εm is a high-level programming system which puts parallelism within the reach of scientists who are ...
There is a trend towards using accelerators to increase performance and energy efficiency of general...
Outlined here are two methods which we believe will play an important role in any distributed memory...
This work was also published as a Rice University thesis/dissertation: http://hdl.handle.net/1911/16...
The goal of this dissertation is to give programmers the ability to achieve high performance by focu...
For distributed memory multicomputers such as the Intel Paragon, the IBM SP-2, the NCUBE/2, and the ...
In order to utilize parallel computers, four approaches, broadly speaking, to the provision of paral...
Coordinated Science Laboratory was formerly known as Control Systems LaboratoryOffice of Naval Resea...
The main contribution of the effort in the last two years is the introduction of the MOPPS system. A...
Many real-time computations such as process control and robotic applications may be naturally distri...
This paper outlines two methods which we believe will play an important role in any distributed memo...
We describe an approach to parallel compilation that seeks to harness the vast amount of fine-grain ...