As cache hierarchies become deeper and the number of cores on a chip increases, managing caches becomes more impor-tant for performance and energy. However, current hard-ware cache management policies do not always adapt opti-mally to the applications behavior: e.g., caches may be pol-luted by data structures whose locality cannot be captured by the caches, and producer-consumer communication in-curs multiple round trips of coherence messages per cache line transferred. We propose load and store instructions that carry hints regarding into which cache(s) the accessed data should be placed. Our instructions allow software to convey locality information to the hardware, while incur-ring minimal hardware cost and not affecting correctness. Our...
This paper introduces the abstract concept of value-aware caches, which exploit value locality rathe...
Cache memory is one of the most important components of a computer system. The cache allows quickly...
As the gap between memory and processor speeds continues to widen, cache efficiency is an increasing...
Thesis: Ph. D., Massachusetts Institute of Technology, Department of Electrical Engineering and Comp...
Memory subsystem with larger capacity and deeper hierarchy has been designed to achieve the maximum ...
Memory subsystem with larger capacity and deeper hierarchy has been designed to achieve the maximum ...
Journal ArticleIn future multi-cores, large amounts of delay and power will be spent accessing data...
Memory (cache, DRAM, and disk) is in charge of providing data and instructions to a computer\u27s pr...
Memory (cache, DRAM, and disk) is in charge of providing data and instructions to a computer\u27s pr...
Set-associative caches are traditionally managed using hardwarebased lookup and replacement schemes ...
Contention for shared cache resources has been recognized as a major bottleneck for multicores—espec...
Contention for shared cache resources has been recognized as a major bottleneck for multicores—espec...
This paper describes Constrained Associative-Mapping-of-Tracking-Entries (C-AMTE), a scalable mechan...
Contention for shared cache resources has been recognized as a major bottleneck for multicores—espec...
The performance gap between processor and memory continues to remain a major performance bottleneck ...
This paper introduces the abstract concept of value-aware caches, which exploit value locality rathe...
Cache memory is one of the most important components of a computer system. The cache allows quickly...
As the gap between memory and processor speeds continues to widen, cache efficiency is an increasing...
Thesis: Ph. D., Massachusetts Institute of Technology, Department of Electrical Engineering and Comp...
Memory subsystem with larger capacity and deeper hierarchy has been designed to achieve the maximum ...
Memory subsystem with larger capacity and deeper hierarchy has been designed to achieve the maximum ...
Journal ArticleIn future multi-cores, large amounts of delay and power will be spent accessing data...
Memory (cache, DRAM, and disk) is in charge of providing data and instructions to a computer\u27s pr...
Memory (cache, DRAM, and disk) is in charge of providing data and instructions to a computer\u27s pr...
Set-associative caches are traditionally managed using hardwarebased lookup and replacement schemes ...
Contention for shared cache resources has been recognized as a major bottleneck for multicores—espec...
Contention for shared cache resources has been recognized as a major bottleneck for multicores—espec...
This paper describes Constrained Associative-Mapping-of-Tracking-Entries (C-AMTE), a scalable mechan...
Contention for shared cache resources has been recognized as a major bottleneck for multicores—espec...
The performance gap between processor and memory continues to remain a major performance bottleneck ...
This paper introduces the abstract concept of value-aware caches, which exploit value locality rathe...
Cache memory is one of the most important components of a computer system. The cache allows quickly...
As the gap between memory and processor speeds continues to widen, cache efficiency is an increasing...