Die-stacking technology allows conventional DRAM to be integrated with processors. While numerous opportunities to make use of such stacked DRAM exist, one promising way is to use it as a large cache. Although previous studies show that DRAM caches can deliver per-formance benefits, there remain inefficiencies as well as significant hardware costs for auxiliary structures. This paper presents two innovations that exploit the bursty nature of memory requests to streamline the DRAM cache. The first is a low-cost Hit-Miss Pre-dictor (HMP) that virtually eliminates the hardware overhead of the previously proposed multi-megabyte MissMap structure. The second is a Self-Balancing Dispatch (SBD) mechanism that dynam-ically sends some requests to th...
IEEE Due to the large storage capacity, high bandwidth and low latency, 3D DRAM is proposed to be th...
Abstract—This paper analyzes the trade-offs in architecting stacked DRAM either as part of main memo...
PosterDRAM vendors have traditionally optimized for low cost and high performance, often making desi...
the tight integration of significant quantities of DRAM with high-performance computation logic. How...
textContemporary DRAM systems have maintained impressive scaling by managing a careful balance betwe...
Die-stacking is a new technology that allows multiple integrated circuits to be stacked on top of ea...
Die-stacked DRAM has been proposed for use as a large, high-bandwidth, last-level cache with hundred...
In this paper, we present Bi-Modal Cache - a flexible stacked DRAM cache organization which simultan...
Abstract—Recent research advocates large die-stacked DRAM caches in manycore servers to break the me...
The effective bandwidth of the FPGA external memory, usually DRAM, is extremely sensitive to the acc...
Die stacking memory technology can enable gigascale DRAM caches that can operate at 4x-8x higher ban...
This paper presents an operating system managed die-stacked DRAM called i-MIRROR that mirrors high l...
Multiple-channel die-stacked DRAMs have been used for maximizing the performance and minimizing the ...
Recent research advocates large die-stacked DRAM caches in manycore servers to break the memory late...
Die-stacked DRAM can provide large amounts of in-package, high-bandwidth cache storage. For server a...
IEEE Due to the large storage capacity, high bandwidth and low latency, 3D DRAM is proposed to be th...
Abstract—This paper analyzes the trade-offs in architecting stacked DRAM either as part of main memo...
PosterDRAM vendors have traditionally optimized for low cost and high performance, often making desi...
the tight integration of significant quantities of DRAM with high-performance computation logic. How...
textContemporary DRAM systems have maintained impressive scaling by managing a careful balance betwe...
Die-stacking is a new technology that allows multiple integrated circuits to be stacked on top of ea...
Die-stacked DRAM has been proposed for use as a large, high-bandwidth, last-level cache with hundred...
In this paper, we present Bi-Modal Cache - a flexible stacked DRAM cache organization which simultan...
Abstract—Recent research advocates large die-stacked DRAM caches in manycore servers to break the me...
The effective bandwidth of the FPGA external memory, usually DRAM, is extremely sensitive to the acc...
Die stacking memory technology can enable gigascale DRAM caches that can operate at 4x-8x higher ban...
This paper presents an operating system managed die-stacked DRAM called i-MIRROR that mirrors high l...
Multiple-channel die-stacked DRAMs have been used for maximizing the performance and minimizing the ...
Recent research advocates large die-stacked DRAM caches in manycore servers to break the memory late...
Die-stacked DRAM can provide large amounts of in-package, high-bandwidth cache storage. For server a...
IEEE Due to the large storage capacity, high bandwidth and low latency, 3D DRAM is proposed to be th...
Abstract—This paper analyzes the trade-offs in architecting stacked DRAM either as part of main memo...
PosterDRAM vendors have traditionally optimized for low cost and high performance, often making desi...