Deploying deep learning models on various devices has become an important topic. The wave of hardware specialization brings a diverse set of acceleration primitives for multi-dimensional tensor computations. These new acceleration primitives, along with the emerging machine learning models, bring tremendous engineering challenges. In this paper, we present TensorIR, a compiler abstraction for optimizing programs with these tensor computation primitives. TensorIR generalizes the loop nest representation used in existing machine learning compilers to bring tensor computation as the first-class citizen. Finally, we build an end-to-end framework on top of our abstraction to automatically optimize deep learning models for given tensor computatio...
Boosting the runtime performance of deep neural networks (DNNs) is critical due to their wide adopti...
A tensor network is a type of decomposition used to express and approximate large arrays of data. A ...
Complex tensor contraction expressions arise in accurate electronic structure models in quantum chem...
Multi-level intermediate representations (MLIR) show great promise for reducing the cost of building...
High-performance tensor programs are crucial to guarantee efficient execution of deep neural network...
Deep Learning (DL) has created a growing demand for simpler ways to develop complex models and effic...
Tensors are higher-dimensional analogs of matrices, and represent a key data abstraction for many ap...
Tensors are higher-order extensions of matrices. While matrix methods form the cornerstone of tradit...
Computational intensive applications such as pattern recognition, and natural language processing, a...
140 pagesTensor algebra lives at the heart of big data applications. Where classical machine learnin...
International audienceThe Deep learning processor (DLP), especially ASIC-based accelerators, have be...
Optimizing the implementation of tensor computations is essential to exploiting the full capacity of...
Introduction In the last two decade, tensor computations developed from a small and little known su...
Deep learning is a very computational intensive task. Traditionally GPUs have been used to speed-up ...
Thesis (Ph.D.)--University of Washington, 2022As the scaling and performance demands for deep learni...
Boosting the runtime performance of deep neural networks (DNNs) is critical due to their wide adopti...
A tensor network is a type of decomposition used to express and approximate large arrays of data. A ...
Complex tensor contraction expressions arise in accurate electronic structure models in quantum chem...
Multi-level intermediate representations (MLIR) show great promise for reducing the cost of building...
High-performance tensor programs are crucial to guarantee efficient execution of deep neural network...
Deep Learning (DL) has created a growing demand for simpler ways to develop complex models and effic...
Tensors are higher-dimensional analogs of matrices, and represent a key data abstraction for many ap...
Tensors are higher-order extensions of matrices. While matrix methods form the cornerstone of tradit...
Computational intensive applications such as pattern recognition, and natural language processing, a...
140 pagesTensor algebra lives at the heart of big data applications. Where classical machine learnin...
International audienceThe Deep learning processor (DLP), especially ASIC-based accelerators, have be...
Optimizing the implementation of tensor computations is essential to exploiting the full capacity of...
Introduction In the last two decade, tensor computations developed from a small and little known su...
Deep learning is a very computational intensive task. Traditionally GPUs have been used to speed-up ...
Thesis (Ph.D.)--University of Washington, 2022As the scaling and performance demands for deep learni...
Boosting the runtime performance of deep neural networks (DNNs) is critical due to their wide adopti...
A tensor network is a type of decomposition used to express and approximate large arrays of data. A ...
Complex tensor contraction expressions arise in accurate electronic structure models in quantum chem...