Powerful memory models, including hierarchies with block transfer or with pipeline of accesses have been proposed in theory and are partially realized in commercial systems, to reduce the average memory overhead per operation. However, even for such powerful models, there are simple direct flow programs, with no branches and no indirect addressing, that require non-constant overhead, resulting in superlinear execution time. Indeed, we characterize a wide, natural class of machines, including nearly all previously proposed models, and develop a technique which yields superlinear time lower bounds on any machine of this class, for suitable direct-flow computations. We propose the Address Dependence Model (ADM) for machines with pipelined memo...
In this work, a model of computation for shared memory parallelism is presented. To address fundamen...
To design effective large-scale multiprocessors, designers need to understand the characteristics of...
Memory access time is a key factor limiting the performance of large-scale, shared-memory multiproce...
Powerful memory models, including hierarchies with block transfer or with pipeline of accesses have ...
The capability of the Random Access Machine (RAM) to execute any instruction in constant time is not...
In a recent paper (SPAA'01), we have established that the Pipelined Hierarchical Random Access Machi...
This paper formulates and investigates the question of whether a given algorithm can be coded in a w...
Any computational model which relies on a physical system is likely to be subject to the fact that i...
AbstractWe study the time relationships between several models of computation (variants of counter m...
We consider extensible processor designs in which the number of gates and the distance that a signal...
Processors have become faster at a much quicker rate than memory access time, creating wide gap betw...
The trend in high-performance microprocessor design is toward increasing computational power on the ...
The memories of real life computers usually have a hierarchical structure with levels like registers...
This work was also published as a Rice University thesis/dissertation: http://hdl.handle.net/1911/16...
Processor performance is directly impacted by the latency of the memory system. As processor core cy...
In this work, a model of computation for shared memory parallelism is presented. To address fundamen...
To design effective large-scale multiprocessors, designers need to understand the characteristics of...
Memory access time is a key factor limiting the performance of large-scale, shared-memory multiproce...
Powerful memory models, including hierarchies with block transfer or with pipeline of accesses have ...
The capability of the Random Access Machine (RAM) to execute any instruction in constant time is not...
In a recent paper (SPAA'01), we have established that the Pipelined Hierarchical Random Access Machi...
This paper formulates and investigates the question of whether a given algorithm can be coded in a w...
Any computational model which relies on a physical system is likely to be subject to the fact that i...
AbstractWe study the time relationships between several models of computation (variants of counter m...
We consider extensible processor designs in which the number of gates and the distance that a signal...
Processors have become faster at a much quicker rate than memory access time, creating wide gap betw...
The trend in high-performance microprocessor design is toward increasing computational power on the ...
The memories of real life computers usually have a hierarchical structure with levels like registers...
This work was also published as a Rice University thesis/dissertation: http://hdl.handle.net/1911/16...
Processor performance is directly impacted by the latency of the memory system. As processor core cy...
In this work, a model of computation for shared memory parallelism is presented. To address fundamen...
To design effective large-scale multiprocessors, designers need to understand the characteristics of...
Memory access time is a key factor limiting the performance of large-scale, shared-memory multiproce...