In machines like the Intel iPSC/2 and the BBN Butterfly, local memory operations are much faster than inter-processor communication. When writing programs for these machines, programmers must worry about exploiting spatial locality of reference. This is tedious and reduces the level of abstraction at the which the programmer works. We are implementing a parallelizing compiler that will shoulder much of that burden. Given a sequential, shared memory program and a specification of how data structures are to be mapped across the processors, our compiler will perform process decomposition to exploit locality of reference. In this paper, we discuss some experiments in parallelizing SIMPLE, a large scientific benchmark from Los Alamos, fo...
Distributed Memory Multicomputers (DMMs) such as the IBM SP-2, the Intel Paragon and the Thinking Ma...
This work was also published as a Rice University thesis/dissertation: http://hdl.handle.net/1911/16...
Before it can achieve wide acceptance,parallel computation must be made significantly easier to prog...
In the context of sequential computers, it is common practice to exploit temporal locality of refer...
Over the past two decades tremendous progress has been made in both the design of parallel architect...
INTRODUCTION The SPMD (Single-Program Multiple-Data Stream) model has been widely adopted as the ba...
This work was also published as a Rice University thesis/dissertation: http://hdl.handle.net/1911/16...
Reduction recognition and optimization are crucial techniques in parallelizing compilers. They are u...
Distributed-memory message-passing machines deliver scalable perfor-mance but are difficult to progr...
Data-parallel languages, such as H scIGH P scERFORMANCE F scORTRAN or F scORTRAN D, provide a machin...
Parallel computing is regarded by most computer scientists as the most likely approach for significa...
We describe an approach to parallel compilation that seeks to harness the vast amount of fine-grain ...
Power consumption and fabrication limitations are increasingly playing significant roles in the desi...
The goal of this dissertation is to give programmers the ability to achieve high performance by focu...
This paper introduces the goals of the Portable, Scalable, Architecture Independent (PSI) Compiler P...
Distributed Memory Multicomputers (DMMs) such as the IBM SP-2, the Intel Paragon and the Thinking Ma...
This work was also published as a Rice University thesis/dissertation: http://hdl.handle.net/1911/16...
Before it can achieve wide acceptance,parallel computation must be made significantly easier to prog...
In the context of sequential computers, it is common practice to exploit temporal locality of refer...
Over the past two decades tremendous progress has been made in both the design of parallel architect...
INTRODUCTION The SPMD (Single-Program Multiple-Data Stream) model has been widely adopted as the ba...
This work was also published as a Rice University thesis/dissertation: http://hdl.handle.net/1911/16...
Reduction recognition and optimization are crucial techniques in parallelizing compilers. They are u...
Distributed-memory message-passing machines deliver scalable perfor-mance but are difficult to progr...
Data-parallel languages, such as H scIGH P scERFORMANCE F scORTRAN or F scORTRAN D, provide a machin...
Parallel computing is regarded by most computer scientists as the most likely approach for significa...
We describe an approach to parallel compilation that seeks to harness the vast amount of fine-grain ...
Power consumption and fabrication limitations are increasingly playing significant roles in the desi...
The goal of this dissertation is to give programmers the ability to achieve high performance by focu...
This paper introduces the goals of the Portable, Scalable, Architecture Independent (PSI) Compiler P...
Distributed Memory Multicomputers (DMMs) such as the IBM SP-2, the Intel Paragon and the Thinking Ma...
This work was also published as a Rice University thesis/dissertation: http://hdl.handle.net/1911/16...
Before it can achieve wide acceptance,parallel computation must be made significantly easier to prog...