In this thesis, we propose two optimization techniques to reduce power consumption in L1 caches (data, texture and constant), shared memory and L2 cache. The first optimization technique targets static power. Evaluation of GPGPU applications shows that once a cache block is accessed by a thread, it takes several hundreds of clock cycles until the same block is accessed again. The long inter-access cycle can be used to put cache cells into drowsy mode and reduce static power. While drowsy cells reduce static power, they increase access time as voltage of a cache cell in drowsy mode should be raised before the block can be accessed. To mitigate performance impact of drowsy cells, we propose a novel technique called coarse grained drowsy mode....
Hardware designers are constantly looking for ways to squeeze waste out of architectures to achieve ...
This report evaluates two distinct methods of improving the performance of GPU memory systems. Over ...
This thesis presents a novel dynamically reconfigurable active L1 instruction and data cache model, ...
As a throughput-oriented device, Graphics Processing Unit(GPU) has already integrated with cache, wh...
abstract: General-purpose processors propel the advances and innovations that are the subject of hum...
The computation power from graphics processing units (GPUs) has become prevalent in many fields of c...
The usage of Graphics Processing Units (GPUs) as an application accelerator has become increasingly ...
Current GPU computing models support a mixture of coherent and incoherent classes of memory operatio...
Power-performance efficiency has become a central focus that is challenging in heterogeneous process...
The massive parallelism provided by general-purpose GPUs (GPGPUs) possessing numerous compute thread...
Energy consumption and speed of execution have long been recognized as conflicting requirements for ...
This thesis proposes a buffered dual access mode cache to reduce power consumption in multicore cach...
abstract: With the massive multithreading execution feature, graphics processing units (GPUs) have b...
Power consumption in computing today has lead the industry towards energy efficient computing. As tr...
Green computing techniques aim to reduce the power foot print of modern embedded devices with partic...
Hardware designers are constantly looking for ways to squeeze waste out of architectures to achieve ...
This report evaluates two distinct methods of improving the performance of GPU memory systems. Over ...
This thesis presents a novel dynamically reconfigurable active L1 instruction and data cache model, ...
As a throughput-oriented device, Graphics Processing Unit(GPU) has already integrated with cache, wh...
abstract: General-purpose processors propel the advances and innovations that are the subject of hum...
The computation power from graphics processing units (GPUs) has become prevalent in many fields of c...
The usage of Graphics Processing Units (GPUs) as an application accelerator has become increasingly ...
Current GPU computing models support a mixture of coherent and incoherent classes of memory operatio...
Power-performance efficiency has become a central focus that is challenging in heterogeneous process...
The massive parallelism provided by general-purpose GPUs (GPGPUs) possessing numerous compute thread...
Energy consumption and speed of execution have long been recognized as conflicting requirements for ...
This thesis proposes a buffered dual access mode cache to reduce power consumption in multicore cach...
abstract: With the massive multithreading execution feature, graphics processing units (GPUs) have b...
Power consumption in computing today has lead the industry towards energy efficient computing. As tr...
Green computing techniques aim to reduce the power foot print of modern embedded devices with partic...
Hardware designers are constantly looking for ways to squeeze waste out of architectures to achieve ...
This report evaluates two distinct methods of improving the performance of GPU memory systems. Over ...
This thesis presents a novel dynamically reconfigurable active L1 instruction and data cache model, ...