In this paper we examine the application of offline algorithms for determining the optimal sequence of loads and superloads (a load of multiple consecutive cache lines) for direct-mapped caches. We evaluate potential gains in terms of miss rate and bandwidth and find that in many cases optimal superloading can noticeably reduce the miss rate without appreciably increasing bandwidth. Then we examine how this performance potential might be realized. We examine the effectiveness of a dynamic online algorithm and of static analysis (profiling) for superloading and compare these to next-line prefetching. Experimental results show improvements comparable to those of the optimal algorithm in terms of miss rates
Directly mapped caches are an attractive option for processor designers as they combine fast lookup ...
The speed of processors increases much faster than the memory access time. This makes memory accesse...
Fast Internet content delivery relies on two layers of caches on the request path. Firstly, content ...
Reconsider the competitiveness ofon-line strategies using k servers versus the optimal off-line stra...
Projections of computer technology forecast proces-sors with peak performance of 1,000 MIPS in the r...
The replacement policies known as MIN and OPT are optimal for a two-level memory hierarchy. The comp...
Abstract—On multicore processors, applications are run shar-ing the cache. This paper presents onlin...
Suppose that a program makes a sequence of m accesses (references) to data blocks, the cache can hol...
This thesis studies the use of software methods to improve memory performance in a heterogeneous cac...
Abstract. Given the increasing gap between processors and memory, prefetching data into cache become...
Directly mapped caches are an attractive option for processor designers as they combine fast lookup ...
Blocking is a well-known optimization technique for improving the effectiveness of memory hierarchie...
A novel model for adaptive cache size control (MACSC) at runtime is proposed in this paper. It autom...
On multicore processors, applications are run sharing the cache. This paper presents online optimiza...
Caching is fundamental to performance in distributed information retrieval systems such as the World...
Directly mapped caches are an attractive option for processor designers as they combine fast lookup ...
The speed of processors increases much faster than the memory access time. This makes memory accesse...
Fast Internet content delivery relies on two layers of caches on the request path. Firstly, content ...
Reconsider the competitiveness ofon-line strategies using k servers versus the optimal off-line stra...
Projections of computer technology forecast proces-sors with peak performance of 1,000 MIPS in the r...
The replacement policies known as MIN and OPT are optimal for a two-level memory hierarchy. The comp...
Abstract—On multicore processors, applications are run shar-ing the cache. This paper presents onlin...
Suppose that a program makes a sequence of m accesses (references) to data blocks, the cache can hol...
This thesis studies the use of software methods to improve memory performance in a heterogeneous cac...
Abstract. Given the increasing gap between processors and memory, prefetching data into cache become...
Directly mapped caches are an attractive option for processor designers as they combine fast lookup ...
Blocking is a well-known optimization technique for improving the effectiveness of memory hierarchie...
A novel model for adaptive cache size control (MACSC) at runtime is proposed in this paper. It autom...
On multicore processors, applications are run sharing the cache. This paper presents online optimiza...
Caching is fundamental to performance in distributed information retrieval systems such as the World...
Directly mapped caches are an attractive option for processor designers as they combine fast lookup ...
The speed of processors increases much faster than the memory access time. This makes memory accesse...
Fast Internet content delivery relies on two layers of caches on the request path. Firstly, content ...