GPUs are widely used to accelerate the training of machine learning workloads. As the machine learning models become increasingly larger, they require a longer time to train, which in turn leads to higher GPU energy consumption. This paper presents GPOEO, an online GPU energy optimization framework for machine learning training workloads. GPOEO dynamically determines the optimal energy configuration by employing a set of novel techniques for online measurement, multi-objective prediction modeling, and search optimization. To characterize the target workload behavior, GPOEO utilizes GPU performance counters. To reduce the performance counter profiling overhead, it uses an analytical model to detect the change of training iteration and only r...
Graphics Processing Units (GPUs) have revolutionized the computing landscape over the past decade. H...
Training deep learning (DL) models is a highly compute-intensive task since it involves operating on...
Energy and power are the main design constraints for modern high-performance computing systems. Inde...
In recent years, the focus of computing has moved away from performance-centric serial computation t...
Machine learning approaches have been widely adopted in recent years due to their capability of lear...
Largescale machine learning frameworks can accelerate training of a neural network by per forming ...
Training deep neural networks (DNNs) is becoming increasingly more resource- and energy-intensive ev...
We devise a performance model for GPU training of Deep Learning Recommendation Models (DLRM), whose ...
Power and energy is the first-class design constraint for multi-core processors and is a limiting fa...
We devise a performance model for GPU training of Deep Learning Recommendation Models (DLRM), whose ...
To improve the power consumption of parallel applications at the runtime, modern processors provide ...
Thanks to parallel processing, it is possible not only to reduce code runtime but also energy consum...
Energy optimization is an increasingly important aspect of today's high-performance computing applic...
A plethora of applications are using machine learning, the operations of which are becoming more com...
Energy optimization is an increasingly important aspect of today’s high-performance computing applic...
Graphics Processing Units (GPUs) have revolutionized the computing landscape over the past decade. H...
Training deep learning (DL) models is a highly compute-intensive task since it involves operating on...
Energy and power are the main design constraints for modern high-performance computing systems. Inde...
In recent years, the focus of computing has moved away from performance-centric serial computation t...
Machine learning approaches have been widely adopted in recent years due to their capability of lear...
Largescale machine learning frameworks can accelerate training of a neural network by per forming ...
Training deep neural networks (DNNs) is becoming increasingly more resource- and energy-intensive ev...
We devise a performance model for GPU training of Deep Learning Recommendation Models (DLRM), whose ...
Power and energy is the first-class design constraint for multi-core processors and is a limiting fa...
We devise a performance model for GPU training of Deep Learning Recommendation Models (DLRM), whose ...
To improve the power consumption of parallel applications at the runtime, modern processors provide ...
Thanks to parallel processing, it is possible not only to reduce code runtime but also energy consum...
Energy optimization is an increasingly important aspect of today's high-performance computing applic...
A plethora of applications are using machine learning, the operations of which are becoming more com...
Energy optimization is an increasingly important aspect of today’s high-performance computing applic...
Graphics Processing Units (GPUs) have revolutionized the computing landscape over the past decade. H...
Training deep learning (DL) models is a highly compute-intensive task since it involves operating on...
Energy and power are the main design constraints for modern high-performance computing systems. Inde...