The heavy burdens of computation and off-chip traffic impede deploying the large scale convolution neural network on embedded platforms. As CNN is attributed to the strong endurance to computation errors, employing block floating point (BFP) arithmetics in CNN accelerators could save the hardware cost and data traffics efficiently, while maintaining the classification accuracy. In this paper, we verify the effects of word width definitions in BFP to the CNN performance without retraining. Several typical CNN models, including VGG16, ResNet-18, ResNet-50 and GoogLeNet, were tested in this paper. Experiments revealed that 8-bit mantissa, including sign bit, in BFP representation merely induced less than 0.3% accuracy loss. In addition, we inv...
The resurgence of machine learning in various applications and it's inherent compute-intensive natur...
Convolutional Neural Networks (CNNs) are currently adopted to solve an ever greater number of proble...
Owing to the growth of the size of convolutional neural networks (CNNs), quantization and loop tilin...
This article analyzes the effects of approximate multiplication when performing inferences on deep c...
International audienceThe numerical format used for representing weights and activations plays a key...
The entangled guardbands in terms of timing specification and energy budget ensure a system against ...
The programmability of FPGA suits the constantly changing convolutional neural network (CNN). Howeve...
DNNs have been finding a growing number of applications including image classification, speech recog...
Today, computer vision (CV) problems are solved with unprecedented accuracy using convolutional neur...
Artificial Neural Networks (NNs) can effectively be used to solve many classification and regression...
The datapath bit-width of hardware accelerators for convolutional neural network (CNN) inference is ...
International audience—Deep Neural Networks are becoming the de-facto standard models for image unde...
Real-time inference of deep convolutional neural networks (CNNs) on embedded systems and SoCs would ...
Real-time inference of deep convolutional neural networks (CNNs) on embedded systems and SoCs would ...
Low-precision neural network models are crucial for reducing the memory footprint and computational...
The resurgence of machine learning in various applications and it's inherent compute-intensive natur...
Convolutional Neural Networks (CNNs) are currently adopted to solve an ever greater number of proble...
Owing to the growth of the size of convolutional neural networks (CNNs), quantization and loop tilin...
This article analyzes the effects of approximate multiplication when performing inferences on deep c...
International audienceThe numerical format used for representing weights and activations plays a key...
The entangled guardbands in terms of timing specification and energy budget ensure a system against ...
The programmability of FPGA suits the constantly changing convolutional neural network (CNN). Howeve...
DNNs have been finding a growing number of applications including image classification, speech recog...
Today, computer vision (CV) problems are solved with unprecedented accuracy using convolutional neur...
Artificial Neural Networks (NNs) can effectively be used to solve many classification and regression...
The datapath bit-width of hardware accelerators for convolutional neural network (CNN) inference is ...
International audience—Deep Neural Networks are becoming the de-facto standard models for image unde...
Real-time inference of deep convolutional neural networks (CNNs) on embedded systems and SoCs would ...
Real-time inference of deep convolutional neural networks (CNNs) on embedded systems and SoCs would ...
Low-precision neural network models are crucial for reducing the memory footprint and computational...
The resurgence of machine learning in various applications and it's inherent compute-intensive natur...
Convolutional Neural Networks (CNNs) are currently adopted to solve an ever greater number of proble...
Owing to the growth of the size of convolutional neural networks (CNNs), quantization and loop tilin...