As technological process shrinks and clock rate increases, instruction caches can no longer be accessed in one cycle. Alternatives are implementing smaller caches (with higher miss rate) or large caches with a pipelined access (with higher branch misprediction penalty). In both cases, the performance obtained is far from the obtained by an ideal large cache with one-cycle access. In this paper we present cache line guided prestaging (CLGP), a novel mechanism that overcomes the limitations of current instruction cache implementations. CLGP employs prefetching to charge future cache lines into a set of fast prestage buffers. These buffers are managed efficiently by the CLGP algorithm, trying to fetch from them as much as possible. Therefore, ...
Cache performance analysis is becoming increasingly important in microprocessor design. This work ex...
Despite large caches, main-memory access latencies still cause significant performance losses in man...
The large number of cache misses of current applications coupled with the increasing cache miss late...
As technological process shrinks and clock rate increases, instruction caches can no longer be acces...
Instruction cache miss latency is becoming an increasingly important performance bottleneck, especia...
The full text of this article is not available on SOAR. WSU users can access the article via IEEE Xp...
Instruction cache misses can severely limit the performance of both superscalar processors and high ...
Ever increasing memory latencies and deeper pipelines push memory farther from the processor. Prefet...
In this dissertation, we provide hardware solutions to increase the efficiency of the cache hierarch...
The exponentially increasing gap between processors and off-chip memory, as measured in processor cy...
As the degree of instruction-level parallelism in superscalar architectures increases, the gap betwe...
A common mechanism to perform hardware-based prefetching for regular accesses to arrays and chained...
Instruction prefetching is an important aspect of contemporary high performance computer architectur...
It is well known that memory latency is a major deterrent to achieving the maximum possible performa...
Abstract. Given the increasing gap between processors and memory, prefetching data into cache become...
Cache performance analysis is becoming increasingly important in microprocessor design. This work ex...
Despite large caches, main-memory access latencies still cause significant performance losses in man...
The large number of cache misses of current applications coupled with the increasing cache miss late...
As technological process shrinks and clock rate increases, instruction caches can no longer be acces...
Instruction cache miss latency is becoming an increasingly important performance bottleneck, especia...
The full text of this article is not available on SOAR. WSU users can access the article via IEEE Xp...
Instruction cache misses can severely limit the performance of both superscalar processors and high ...
Ever increasing memory latencies and deeper pipelines push memory farther from the processor. Prefet...
In this dissertation, we provide hardware solutions to increase the efficiency of the cache hierarch...
The exponentially increasing gap between processors and off-chip memory, as measured in processor cy...
As the degree of instruction-level parallelism in superscalar architectures increases, the gap betwe...
A common mechanism to perform hardware-based prefetching for regular accesses to arrays and chained...
Instruction prefetching is an important aspect of contemporary high performance computer architectur...
It is well known that memory latency is a major deterrent to achieving the maximum possible performa...
Abstract. Given the increasing gap between processors and memory, prefetching data into cache become...
Cache performance analysis is becoming increasingly important in microprocessor design. This work ex...
Despite large caches, main-memory access latencies still cause significant performance losses in man...
The large number of cache misses of current applications coupled with the increasing cache miss late...