Modern high-performance cache implementations use subarrays to reduce the capacitive load on the bit-lines and achieve faster access time [6]. To overlap bitline precharging time with address decoding and wordline assertion, caches typically precharge all subarrays simultaneously prior to a cache access. Though only a small number of subarrays are accessed on a cache access, precharging all subarrays leads high energy consumption in modern and future high-performance deep-sub micron caches, because current CMOS scaling trends significantly increase the leakage from bitlines as process generation evolves [2, 3]. To precharge only the required subarrays (and reduce precharging activity and energy), delayed pre-charging technique was proposed,...
Abstract—Cache compression improves the performance of a multi-core system by being able to store mo...
Memory access latency is the primary performance bottle-neck in modern computer systems. Prefetching...
[[abstract]]Focusing on the way-predicting cache with sub-block placement, we propose a new cache sc...
The performance of superscalar processors is more sensitive to the memory system delay than their si...
In this dissertation, we provide hardware solutions to increase the efficiency of the cache hierarch...
The full text of this article is not available on SOAR. WSU users can access the article via IEEE Xp...
IEEE Computer Society Annual Symposium on VLSI : April 7-9, 2008 : Montpellier, FranceThe share of l...
The line size/performance trade-offs in off-chip second-level caches in light of energy-efficiency a...
As the degree of instruction-level parallelism in superscalar architectures increases, the gap betwe...
A well known performance bottleneck in computer architecture is the so-called memory wall. This term...
4th International Workshop on Dependable Embedded Systems : October 9, 2007 : Beijing, ChinaLeakage ...
Memory access latency is the primary performance bottle-neck in modern computer systems. Prefetching...
Low-latency data access is essential for performance. To achieve this, processors use fast first-lev...
Conventional cache prefetching approaches can be either hardware-based, generally by using a one-blo...
Deep-submicron CMOS designs maintain high transistor switching speeds by scaling down the supply vol...
Abstract—Cache compression improves the performance of a multi-core system by being able to store mo...
Memory access latency is the primary performance bottle-neck in modern computer systems. Prefetching...
[[abstract]]Focusing on the way-predicting cache with sub-block placement, we propose a new cache sc...
The performance of superscalar processors is more sensitive to the memory system delay than their si...
In this dissertation, we provide hardware solutions to increase the efficiency of the cache hierarch...
The full text of this article is not available on SOAR. WSU users can access the article via IEEE Xp...
IEEE Computer Society Annual Symposium on VLSI : April 7-9, 2008 : Montpellier, FranceThe share of l...
The line size/performance trade-offs in off-chip second-level caches in light of energy-efficiency a...
As the degree of instruction-level parallelism in superscalar architectures increases, the gap betwe...
A well known performance bottleneck in computer architecture is the so-called memory wall. This term...
4th International Workshop on Dependable Embedded Systems : October 9, 2007 : Beijing, ChinaLeakage ...
Memory access latency is the primary performance bottle-neck in modern computer systems. Prefetching...
Low-latency data access is essential for performance. To achieve this, processors use fast first-lev...
Conventional cache prefetching approaches can be either hardware-based, generally by using a one-blo...
Deep-submicron CMOS designs maintain high transistor switching speeds by scaling down the supply vol...
Abstract—Cache compression improves the performance of a multi-core system by being able to store mo...
Memory access latency is the primary performance bottle-neck in modern computer systems. Prefetching...
[[abstract]]Focusing on the way-predicting cache with sub-block placement, we propose a new cache sc...