Memory hierarchies play an important role in microarchitectural design to bridge the performance gap between modern microprocessors and main memory. However, memory hierarchies are inefficient due to storing waste. This dissertation quantifies two types of waste, dead blocks and data redundancy. This dissertation studies waste in diverse memory hierarchies and proposes techniques to reduce waste to improve performance with limited overhead. This dissertation observes that waste of dead blocks in an inclusive last level cache consists of two kinds of blocks: blocks that are highly accessed in core caches and blocks that have low temporal locality in both core caches and the last-level cache. Blindly replacing all dead blocks in an inclusive...
textOne of the major limiters to computer system performance has been the access to main memory, wh...
Program redundancy analysis and optimization have been an important component in optimizing compiler...
The increasing speed gap between microprocessors and off-chip DRAM makes last-level caches (LLCs) a ...
Memory hierarchies play an important role in microarchitectural design to bridge the performance gap...
Programming languages with automatic memory management are continuing to grow in popularity due to e...
This dissertation addresses two sets of challenges facing processor design as the industry enters th...
As a throughput-oriented device, Graphics Processing Unit(GPU) has already integrated with cache, wh...
This thesis states that dynamic profiling of the memory reference stream can improve energy and per...
Reference counting is a garbage-collection technique that maintains a per-object count of the number...
textModern microprocessors devote a large portion of their chip area to caches in order to bridge t...
Cache working-set adaptation is key as embedded systems move to multiprocessor and Simultaneous Mult...
Memory subsystem with larger capacity and deeper hierarchy has been designed to achieve the maximum ...
Minimizing power, increasing performance, and delivering effective memory bandwidth are today's prim...
Architects have adopted the shared memory model that implicitly manages cache coherence and cache ca...
DRAM caches are important for enabling effective heterogeneous memory systems that can transparently...
textOne of the major limiters to computer system performance has been the access to main memory, wh...
Program redundancy analysis and optimization have been an important component in optimizing compiler...
The increasing speed gap between microprocessors and off-chip DRAM makes last-level caches (LLCs) a ...
Memory hierarchies play an important role in microarchitectural design to bridge the performance gap...
Programming languages with automatic memory management are continuing to grow in popularity due to e...
This dissertation addresses two sets of challenges facing processor design as the industry enters th...
As a throughput-oriented device, Graphics Processing Unit(GPU) has already integrated with cache, wh...
This thesis states that dynamic profiling of the memory reference stream can improve energy and per...
Reference counting is a garbage-collection technique that maintains a per-object count of the number...
textModern microprocessors devote a large portion of their chip area to caches in order to bridge t...
Cache working-set adaptation is key as embedded systems move to multiprocessor and Simultaneous Mult...
Memory subsystem with larger capacity and deeper hierarchy has been designed to achieve the maximum ...
Minimizing power, increasing performance, and delivering effective memory bandwidth are today's prim...
Architects have adopted the shared memory model that implicitly manages cache coherence and cache ca...
DRAM caches are important for enabling effective heterogeneous memory systems that can transparently...
textOne of the major limiters to computer system performance has been the access to main memory, wh...
Program redundancy analysis and optimization have been an important component in optimizing compiler...
The increasing speed gap between microprocessors and off-chip DRAM makes last-level caches (LLCs) a ...