Analog mixed-signal (AMS) devices promise faster, more energy-efficient deep neural network (DNN) inference than their digital counterparts. However, recent studies show that DNNs on AMS devices with fixed-point numbers can incur an accuracy penalty because of precision loss. To mitigate this penalty, we present a novel AMS-compatible adaptive block floating-point (ABFP) number representation. We also introduce amplification (or gain) as a method for increasing the accuracy of the number representation without increasing the bit precision of the output. We evaluate the effectiveness of ABFP on the DNNs in the MLPerf datacenter inference benchmark -- realizing less than $1\%$ loss in accuracy compared to FLOAT32. We also propose a novel meth...
Deep learning training involves a large number of operations, which are dominated by high dimensiona...
The acceleration of deep-learning kernels in hardware relies on matrix multiplications that are exec...
In this paper, low end Digital Signal Processors (DSPs) are applied to accelerate integer neural net...
Approximate computing has emerged as a promising approach to energy-efficient design of digital syst...
Due to limited size, cost and power, embedded devices do not offer the same computational throughput...
Abstract — Deep machine learning (DML) holds the potential to revolutionize machine learning by auto...
The last decade has witnessed the breakthrough of deep neural networks (DNNs) in many fields. With t...
There is an urgent need for compact, fast, and power-efficient hardware implementations of state-of-...
Current state-of-the-art employs approximate multipliers to address the highly increased power deman...
The heavy burdens of computation and off-chip traffic impede deploying the large scale convolution n...
International audienceThe most compute-intensive stage of deep neural network (DNN) training is matr...
A neural network is quantized for the mitigation of nonlinear and components distortions in a 16-QAM...
Recently, analog compute-in-memory (CIM) architectures based on emerging analog non-volatile memory ...
Deep neural networks (DNNs) have achieved unprecedented capabilities in tasks such as analysis and r...
International audienceDeep Neural Networks (DNN) represent a performance-hungry application. Floatin...
Deep learning training involves a large number of operations, which are dominated by high dimensiona...
The acceleration of deep-learning kernels in hardware relies on matrix multiplications that are exec...
In this paper, low end Digital Signal Processors (DSPs) are applied to accelerate integer neural net...
Approximate computing has emerged as a promising approach to energy-efficient design of digital syst...
Due to limited size, cost and power, embedded devices do not offer the same computational throughput...
Abstract — Deep machine learning (DML) holds the potential to revolutionize machine learning by auto...
The last decade has witnessed the breakthrough of deep neural networks (DNNs) in many fields. With t...
There is an urgent need for compact, fast, and power-efficient hardware implementations of state-of-...
Current state-of-the-art employs approximate multipliers to address the highly increased power deman...
The heavy burdens of computation and off-chip traffic impede deploying the large scale convolution n...
International audienceThe most compute-intensive stage of deep neural network (DNN) training is matr...
A neural network is quantized for the mitigation of nonlinear and components distortions in a 16-QAM...
Recently, analog compute-in-memory (CIM) architectures based on emerging analog non-volatile memory ...
Deep neural networks (DNNs) have achieved unprecedented capabilities in tasks such as analysis and r...
International audienceDeep Neural Networks (DNN) represent a performance-hungry application. Floatin...
Deep learning training involves a large number of operations, which are dominated by high dimensiona...
The acceleration of deep-learning kernels in hardware relies on matrix multiplications that are exec...
In this paper, low end Digital Signal Processors (DSPs) are applied to accelerate integer neural net...