The memory space taken to host and process large tensor graphs is a limiting factor for embedded ConvNets. Even though many data-driven compression pipelines have proven their efficacy, this work shows there is still room for optimization at the intersection with compute-oriented optimizations. We demonstrate that tensor pruning via weight sparsification can cooperate with a model-agnostic tiling strategy, leading ConvNets towards a new feasible region of the solution space. The collected results show for the first time fast versions of MobileNets deployed at full scale on an ARM M7 core with 512KB of RAM and 2MB of FLASH memory
International audienceMany domains of scientific simulation (chemistry, condensed matter physics, da...
International audienceMany domains of scientific simulation (chemistry, condensed matter physics, da...
Thesis: Ph. D., Massachusetts Institute of Technology, Department of Electrical Engineering and Comp...
Popular Machine Learning (ML) and High Performance Computing (HPC) workloads contribute to a signifi...
140 pagesTensor algebra lives at the heart of big data applications. Where classical machine learnin...
The inherent sparsity present in convolutional neural networks (CNNs) offers a valuable opportunity ...
Deep neural networks (DNNs) have become the primary methods to solve machine learning and artificial...
The emergence of deep learning has launched many works in deep learning accelerators. To fully reali...
Tensor algorithms are a rapidly growing field of research with applications in many scientific domai...
In this thesis, we develop high performance algorithms for certain computations involving dense tens...
Tensor decomposition (TD) is an important method for extracting latent information from high-dimensi...
The emergence of deep learning has launched many works in deep learning accelerators. To fully reali...
In this thesis, we develop high performance algorithms for certain computations involving dense tens...
In this thesis, we develop high performance algorithms for certain computations involving dense tens...
University of Minnesota Ph.D. dissertation. April 2019. Major: Computer Science. Advisor: George Ka...
International audienceMany domains of scientific simulation (chemistry, condensed matter physics, da...
International audienceMany domains of scientific simulation (chemistry, condensed matter physics, da...
Thesis: Ph. D., Massachusetts Institute of Technology, Department of Electrical Engineering and Comp...
Popular Machine Learning (ML) and High Performance Computing (HPC) workloads contribute to a signifi...
140 pagesTensor algebra lives at the heart of big data applications. Where classical machine learnin...
The inherent sparsity present in convolutional neural networks (CNNs) offers a valuable opportunity ...
Deep neural networks (DNNs) have become the primary methods to solve machine learning and artificial...
The emergence of deep learning has launched many works in deep learning accelerators. To fully reali...
Tensor algorithms are a rapidly growing field of research with applications in many scientific domai...
In this thesis, we develop high performance algorithms for certain computations involving dense tens...
Tensor decomposition (TD) is an important method for extracting latent information from high-dimensi...
The emergence of deep learning has launched many works in deep learning accelerators. To fully reali...
In this thesis, we develop high performance algorithms for certain computations involving dense tens...
In this thesis, we develop high performance algorithms for certain computations involving dense tens...
University of Minnesota Ph.D. dissertation. April 2019. Major: Computer Science. Advisor: George Ka...
International audienceMany domains of scientific simulation (chemistry, condensed matter physics, da...
International audienceMany domains of scientific simulation (chemistry, condensed matter physics, da...
Thesis: Ph. D., Massachusetts Institute of Technology, Department of Electrical Engineering and Comp...