Recent years have seen the adoption of Machine Learning (ML) techniques to predict the performance of large-scale applications, mostly at a coarse level. In contrast, we propose to use ML techniques for performance prediction at a much finer granularity, namely at the Basic Block (BB) level, which are single entry, single exit code blocks that are used for analysis by the compilers to break down a large code into manageable pieces. We extrapolate the basic block execution counts of GPU applications and use them for predicting the performance for large input sizes from the counts of smaller input sizes. We train a Poisson Neural Network (PNN) model using random input values as well as the lowest input values of the application to learn the r...
The parallel and distributed platforms of High Performance Computing available today have became mor...
General-purpose computing systems have benefited from technology scaling for several decades but are...
Heterogeneous processing using GPUs is here to stay and today spans mobile devices, laptops, and ...
CPUs and dedicated accelerators (namely GPUs and FPGAs) continue to grow increasingly large and comp...
Training machine learning (ML) algorithms is a computationally intensive process, which is frequentl...
The resurgence of machine learning since the late 1990s has been enabled by significant advances in ...
We devise a performance model for GPU training of Deep Learning Recommendation Models (DLRM), whose ...
Training machine learning (ML) algorithms is a computationally intensive process, which is frequentl...
Abstract—To exploit the abundant computational power of the world’s fastest supercomputers, an even ...
Scientific applications often require massive amounts of compute time and power. With the constantly...
Accurately modeling and predicting performance for large-scale applications becomes increasingly dif...
Consistently growing architectural complexity and machine scales make creating accurate performance ...
Big Data has been a catalyst force for the Machine Learning (ML) area, forcing us to rethink existin...
Today, machine learning (ML) workloads are nearly ubiquitous. Over the past decade, much effort has ...
The significant growth in computational power of mod-ern Graphics Processing Units(GPUs) coupled wit...
The parallel and distributed platforms of High Performance Computing available today have became mor...
General-purpose computing systems have benefited from technology scaling for several decades but are...
Heterogeneous processing using GPUs is here to stay and today spans mobile devices, laptops, and ...
CPUs and dedicated accelerators (namely GPUs and FPGAs) continue to grow increasingly large and comp...
Training machine learning (ML) algorithms is a computationally intensive process, which is frequentl...
The resurgence of machine learning since the late 1990s has been enabled by significant advances in ...
We devise a performance model for GPU training of Deep Learning Recommendation Models (DLRM), whose ...
Training machine learning (ML) algorithms is a computationally intensive process, which is frequentl...
Abstract—To exploit the abundant computational power of the world’s fastest supercomputers, an even ...
Scientific applications often require massive amounts of compute time and power. With the constantly...
Accurately modeling and predicting performance for large-scale applications becomes increasingly dif...
Consistently growing architectural complexity and machine scales make creating accurate performance ...
Big Data has been a catalyst force for the Machine Learning (ML) area, forcing us to rethink existin...
Today, machine learning (ML) workloads are nearly ubiquitous. Over the past decade, much effort has ...
The significant growth in computational power of mod-ern Graphics Processing Units(GPUs) coupled wit...
The parallel and distributed platforms of High Performance Computing available today have became mor...
General-purpose computing systems have benefited from technology scaling for several decades but are...
Heterogeneous processing using GPUs is here to stay and today spans mobile devices, laptops, and ...