Deep Neural Networks (DNN) are well understood to be one of the largest consumers of HPC resources and efficiently running their training and inference phases on modern heterogeneous architectures (and accelerators) poses an important challenge for the compilation community. Currently, DNNs are actively being studied by the automatic parallelization and polyhedral compilation communities for the same purpose. In this (initial) paper, we study the kernels of four varieties of DNN layers with the goal of applying automatic parallelization techniques for latest architectures. We show the affine (Polyhedral) nature of these kernels thereby showing that they are amenable to well known polyhedral compilation techniques. For benchmarking purposes,...
Improving the e ciency of neural networks has great potential impact due to their wide range of pos...
International audienceAutomatic parallel code generation from high-level abstractions such as those ...
Deep Neural Networks (DNNs) are extremely computationally demanding, which presents a large barrier ...
Deep Neural Networks (DNNs) have revolutionized many aspects of our lives. The use of DNNs is becomi...
At the heart of deep learning training and inferencing are computationally intensive primitives such...
Polyhedral compilation has been successful in analyzing, optimizing, automatically parallelizing a�...
In the past decade, deep learning became the prevalent methodology for predictive modeling thanks to...
International audienceHigh-level program optimizations, such as loop transformations, are critical f...
2013 Spring.Includes bibliographical references.With the introduction of multi-core processors, moti...
DNNs have been finding a growing number of applications including image classification, speech recog...
The polyhedral model for loop parallelization has proved to be an effective tool for ad-vanced optim...
Many modern (mobile) systems involve memory intensive computations. External memory accesses are cos...
On modern architectures, a missed optimization can translate into performance degradations reaching ...
International audienceThere may be a huge gap between the statements outlined by programmers in a pr...
International audienceThe polyhedral model is a powerful framework for automatic optimization and pa...
Improving the e ciency of neural networks has great potential impact due to their wide range of pos...
International audienceAutomatic parallel code generation from high-level abstractions such as those ...
Deep Neural Networks (DNNs) are extremely computationally demanding, which presents a large barrier ...
Deep Neural Networks (DNNs) have revolutionized many aspects of our lives. The use of DNNs is becomi...
At the heart of deep learning training and inferencing are computationally intensive primitives such...
Polyhedral compilation has been successful in analyzing, optimizing, automatically parallelizing a�...
In the past decade, deep learning became the prevalent methodology for predictive modeling thanks to...
International audienceHigh-level program optimizations, such as loop transformations, are critical f...
2013 Spring.Includes bibliographical references.With the introduction of multi-core processors, moti...
DNNs have been finding a growing number of applications including image classification, speech recog...
The polyhedral model for loop parallelization has proved to be an effective tool for ad-vanced optim...
Many modern (mobile) systems involve memory intensive computations. External memory accesses are cos...
On modern architectures, a missed optimization can translate into performance degradations reaching ...
International audienceThere may be a huge gap between the statements outlined by programmers in a pr...
International audienceThe polyhedral model is a powerful framework for automatic optimization and pa...
Improving the e ciency of neural networks has great potential impact due to their wide range of pos...
International audienceAutomatic parallel code generation from high-level abstractions such as those ...
Deep Neural Networks (DNNs) are extremely computationally demanding, which presents a large barrier ...