DRAM caches have been shown to be an effective way to utilize the bandwidth and capacity of 3D stacked DRAM. Although they can capture the spatial and temporal data locality of applications, their access latency is still substantially higher than conventional on-chip SRAM caches. Moreover, their tag access latency and storage overheads are excessive. Storing tags for a large DRAM cache in SRAM is impractical as it would occupy a significant fraction of the processor chip. Storing them in the DRAM itself incurs high access overheads. Attempting to cache the DRAM tags on the processor adds a constant delay to the access time. In this paper, we introduce FusionCache, a DRAM cache that offers more efficient tag accesses by fusing DRAM cache tag...
This paper presents a technique for minimizing chip-area cost of implementing an on-chip cache memor...
As DRAM access latencies approach a thousand instructionexecution times and on-chip caches grow to m...
We propose to overcome the memory capacity limitation of GPUs with a Heterogeneous Memory Stack (HMS...
DRAM caches have shown excellent potential in capturing the spatial and temporal data locality of ap...
Abstract—Recent research advocates large die-stacked DRAM caches in manycore servers to break the me...
Recent research advocates large die-stacked DRAM caches in manycore servers to break the memory late...
Abstract—Die-stacked DRAM caches represent an emerging technology that offers a new level of cache b...
IEEE Due to the large storage capacity, high bandwidth and low latency, 3D DRAM is proposed to be th...
Hybrid main memories composed of DRAM as a cache to scalable non-volatile memories such as phase-cha...
The bandwidth of traditional DRAM is pin limited and so does not scale wellwith the increasing deman...
In this paper, we present Bi-Modal Cache - a flexible stacked DRAM cache organization which simultan...
© 2017 Association for Computing Machinery. Placing the DRAM in the same package as a processor enab...
We characterize the cache behavior of an in-memory tag table and demonstrate that an optimized imple...
Abstract—This paper analyzes the trade-offs in architecting stacked DRAM either as part of main memo...
Die-stacking is a new technology that allows multiple integrated circuits to be stacked on top of ea...
This paper presents a technique for minimizing chip-area cost of implementing an on-chip cache memor...
As DRAM access latencies approach a thousand instructionexecution times and on-chip caches grow to m...
We propose to overcome the memory capacity limitation of GPUs with a Heterogeneous Memory Stack (HMS...
DRAM caches have shown excellent potential in capturing the spatial and temporal data locality of ap...
Abstract—Recent research advocates large die-stacked DRAM caches in manycore servers to break the me...
Recent research advocates large die-stacked DRAM caches in manycore servers to break the memory late...
Abstract—Die-stacked DRAM caches represent an emerging technology that offers a new level of cache b...
IEEE Due to the large storage capacity, high bandwidth and low latency, 3D DRAM is proposed to be th...
Hybrid main memories composed of DRAM as a cache to scalable non-volatile memories such as phase-cha...
The bandwidth of traditional DRAM is pin limited and so does not scale wellwith the increasing deman...
In this paper, we present Bi-Modal Cache - a flexible stacked DRAM cache organization which simultan...
© 2017 Association for Computing Machinery. Placing the DRAM in the same package as a processor enab...
We characterize the cache behavior of an in-memory tag table and demonstrate that an optimized imple...
Abstract—This paper analyzes the trade-offs in architecting stacked DRAM either as part of main memo...
Die-stacking is a new technology that allows multiple integrated circuits to be stacked on top of ea...
This paper presents a technique for minimizing chip-area cost of implementing an on-chip cache memor...
As DRAM access latencies approach a thousand instructionexecution times and on-chip caches grow to m...
We propose to overcome the memory capacity limitation of GPUs with a Heterogeneous Memory Stack (HMS...