As VLSI chip sizes and densities increase, it becomes possible to put many processing elements on a single chip and connect them together with a low latency communication network. In this paper we propose a software system, SUDS (Software Un-Do System), that leverages these resources using speculation to exploit parallelism in integer programs with many data dependences. We demonstrate that in order to achieve parallel speedups a speculation system must deliver memory request latencies lower than about 30 cycles. We give a cost breakdown for our current working implementation of SUDS that has a memory request latency that is nearly able to meet this goal. In addition, we identify the three primitive runtime operations that are necessary to ...
Thesis (Ph. D.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer...
A software distributed shared memory (DSM) system allows shared memory parallel programs to execute ...
Effectively utilizing available parallelism is becoming harder and harder as systems evolve to many-...
As VLSI chip sizes and densities increase, it becomes possible to put many processing elements on a ...
Memory dependence prediction allows out-of-order is-sue processors to achieve high degrees of instru...
One of the main performance bottlenecks of processors today is the discrepancy between processor and...
The traditional single-core processors are being replaced by chip multiprocessors (CMPs) where sever...
We present a software approach to design a thread-level data dependence speculation system targeting...
Memory dependence prediction allows out-of-order issue processors to achieve high degrees of instruc...
In computer systems, latency tolerance is the use of concurrency to achieve high performance in spit...
In this thesis, I describe the design and implementation of a software data cache for speculative lo...
As the existing techniques that empower the modern high-performance processors are being refined and...
PhD ThesisCurrent microprocessors improve performance by exploiting instruction-level parallelism (I...
Current microprocessors improve performance by exploiting instruction-level parallelism (ILP). ILP h...
grantor: University of TorontoTo fully exploit the potential of single-chip multiprocessor...
Thesis (Ph. D.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer...
A software distributed shared memory (DSM) system allows shared memory parallel programs to execute ...
Effectively utilizing available parallelism is becoming harder and harder as systems evolve to many-...
As VLSI chip sizes and densities increase, it becomes possible to put many processing elements on a ...
Memory dependence prediction allows out-of-order is-sue processors to achieve high degrees of instru...
One of the main performance bottlenecks of processors today is the discrepancy between processor and...
The traditional single-core processors are being replaced by chip multiprocessors (CMPs) where sever...
We present a software approach to design a thread-level data dependence speculation system targeting...
Memory dependence prediction allows out-of-order issue processors to achieve high degrees of instruc...
In computer systems, latency tolerance is the use of concurrency to achieve high performance in spit...
In this thesis, I describe the design and implementation of a software data cache for speculative lo...
As the existing techniques that empower the modern high-performance processors are being refined and...
PhD ThesisCurrent microprocessors improve performance by exploiting instruction-level parallelism (I...
Current microprocessors improve performance by exploiting instruction-level parallelism (ILP). ILP h...
grantor: University of TorontoTo fully exploit the potential of single-chip multiprocessor...
Thesis (Ph. D.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer...
A software distributed shared memory (DSM) system allows shared memory parallel programs to execute ...
Effectively utilizing available parallelism is becoming harder and harder as systems evolve to many-...