International audienceThe most compute-intensive stage of deep neural network (DNN) training is matrix multiplication where the multiply-accumulate (MAC) operator is key. To reduce training costs, we consider using low-precision arithmetic for MAC operations. While low-precision training has been investigated in prior work, the focus has been on reducing the number of bits in weights or activations without compromising accuracy. In contrast, the focus in this paper is on implementation details beyond weight or activation width that affect area and accuracy. In particular, we investigate the impact of fixed-versus floating-point representations, multiplier rounding, and floatingpoint exceptional value support. Results suggest that (1) lowpre...
Understanding the bit-width precision is critical in compact representation of a Deep Neural Network...
International audienceDeep Neural Networks (DNN) represent a performance-hungry application. Floatin...
DNNs have been finding a growing number of applications including image classification, speech recog...
International audienceThe most compute-intensive stage of deep neural network (DNN) training is matr...
Due to limited size, cost and power, embedded devices do not offer the same computational throughput...
Posit™ arithmetic is a recent alternative format to the IEEE 754 standard for floating-point numbers...
Low-precision formats have recently driven major breakthroughs in neural network (NN) training and i...
Approximate computing has emerged as a promising approach to energy-efficient design of digital syst...
Low-precision formats have recently driven major breakthroughs in neural network (NN) training and i...
Mixed-precision (MP) arithmetic combining both single- and half-precision operands has been successf...
Analog mixed-signal (AMS) devices promise faster, more energy-efficient deep neural network (DNN) in...
This article analyzes the effects of approximate multiplication when performing inferences on deep c...
Hardware accelerators for Deep Neural Networks (DNNs) that use reduced precision parameters are more...
Training large-scale deep neural networks (DNNs) currently requires a significant amount of energy, ...
International audienceResource requirements for hardware acceleration of neural networks inference i...
Understanding the bit-width precision is critical in compact representation of a Deep Neural Network...
International audienceDeep Neural Networks (DNN) represent a performance-hungry application. Floatin...
DNNs have been finding a growing number of applications including image classification, speech recog...
International audienceThe most compute-intensive stage of deep neural network (DNN) training is matr...
Due to limited size, cost and power, embedded devices do not offer the same computational throughput...
Posit™ arithmetic is a recent alternative format to the IEEE 754 standard for floating-point numbers...
Low-precision formats have recently driven major breakthroughs in neural network (NN) training and i...
Approximate computing has emerged as a promising approach to energy-efficient design of digital syst...
Low-precision formats have recently driven major breakthroughs in neural network (NN) training and i...
Mixed-precision (MP) arithmetic combining both single- and half-precision operands has been successf...
Analog mixed-signal (AMS) devices promise faster, more energy-efficient deep neural network (DNN) in...
This article analyzes the effects of approximate multiplication when performing inferences on deep c...
Hardware accelerators for Deep Neural Networks (DNNs) that use reduced precision parameters are more...
Training large-scale deep neural networks (DNNs) currently requires a significant amount of energy, ...
International audienceResource requirements for hardware acceleration of neural networks inference i...
Understanding the bit-width precision is critical in compact representation of a Deep Neural Network...
International audienceDeep Neural Networks (DNN) represent a performance-hungry application. Floatin...
DNNs have been finding a growing number of applications including image classification, speech recog...