Accelerators designed for deep neural network (DNN) inference with extremely low operand widths, down to 1-bit, have become popular due to their ability to significantly reduce energy consumption during inference. This paper introduces a compiler-programmable flexible System-on-Chip (SoC) with mixed-precision support. This SoC is based on a Transport-Triggered Architecture (TTA) that facilitates efficient implementation of DNN workloads. By shifting the complexity of data movement from the hardware scheduler to the exposed-datapath compiler, DNN workloads can be implemented in an energy efficient yet flexible way. The architecture is fully supported by a compiler and can be programmed using C/C++/OpenCL. The SoC is implemented using 22nm FD...
The continued success of Deep Neural Networks (DNNs) in classification tasks has sparked a trend of ...
Proceedings of a meeting held 19-23 March 2018, Dresden, GermanyInternational audienceArtificial int...
Real-time inference of deep convolutional neural networks (CNNs) on embedded systems and SoCs would ...
Recently, accelerators for extremely quantized deep neural network (DNN) inference with operand widt...
Recently, accelerators for extremely quantized deep neural network (DNN) inference with operand widt...
Deep neural networks (DNN) are achieving state-of-the-art performance in many artificial intelligenc...
The computational requirements of artificial intelligence workloads are growing exponentially. In ad...
Deep Neural Networks (DNNs) have become a promising solution to inject AI in our daily lives from se...
The size of neural networks in deep learning techniques is increasing and varies significantly accor...
This paper introduces an energy-efficient design method for Deep Neural Network (DNN) accelerator. A...
The computation efficiency and flexibility of the accelerator hinder deep neural network (DNN) imple...
Deep Neural Networks (DNNs) are widely used in various application domains and achieve remarkable re...
Deep neural networks (DNNs) have shown extraordinary performance in recent years for various applica...
Today, hardware accelerators are widely accepted as a cost-effective solution for emerging applicati...
Modern deep neural network (DNN) applications demand a remarkable processing throughput usually unme...
The continued success of Deep Neural Networks (DNNs) in classification tasks has sparked a trend of ...
Proceedings of a meeting held 19-23 March 2018, Dresden, GermanyInternational audienceArtificial int...
Real-time inference of deep convolutional neural networks (CNNs) on embedded systems and SoCs would ...
Recently, accelerators for extremely quantized deep neural network (DNN) inference with operand widt...
Recently, accelerators for extremely quantized deep neural network (DNN) inference with operand widt...
Deep neural networks (DNN) are achieving state-of-the-art performance in many artificial intelligenc...
The computational requirements of artificial intelligence workloads are growing exponentially. In ad...
Deep Neural Networks (DNNs) have become a promising solution to inject AI in our daily lives from se...
The size of neural networks in deep learning techniques is increasing and varies significantly accor...
This paper introduces an energy-efficient design method for Deep Neural Network (DNN) accelerator. A...
The computation efficiency and flexibility of the accelerator hinder deep neural network (DNN) imple...
Deep Neural Networks (DNNs) are widely used in various application domains and achieve remarkable re...
Deep neural networks (DNNs) have shown extraordinary performance in recent years for various applica...
Today, hardware accelerators are widely accepted as a cost-effective solution for emerging applicati...
Modern deep neural network (DNN) applications demand a remarkable processing throughput usually unme...
The continued success of Deep Neural Networks (DNNs) in classification tasks has sparked a trend of ...
Proceedings of a meeting held 19-23 March 2018, Dresden, GermanyInternational audienceArtificial int...
Real-time inference of deep convolutional neural networks (CNNs) on embedded systems and SoCs would ...