International audienceThe numerical format used for representing weights and activations plays a key role in the computational efficiency and robustness of CNNs. Recently, a 16-bit floating point format called Brain-Float 16 (bf16) has been proposed and implemented in hardware accelerators. However, the robustness of accelerators implemented with this format has not yet been studied. In this paper, we perform a comparison of the robustness of state-of-the art CNNs implemented with 8-bit integer, Brain-Float 16 and 32bit floating point formats. We also introduce an error detection and masking technique, called opportunistic parity (OP), which can detect and mask errors in the weights with zero storage overhead. With this technique, the robus...
FP8 is a natural progression for accelerating deep learning training inference beyond the 16-bit for...
The datapath bit-width of hardware accelerators for convolutional neural network (CNN) inference is ...
Convolutional Neural Networks (CNNs) are broadly used in safety-critical applications such as autono...
International audienceThe numerical format used for representing weights and activations plays a key...
The heavy burdens of computation and off-chip traffic impede deploying the large scale convolution n...
We propose a lightweight scheme where the formation of a data block is changed in such a way that it...
Quantization of neural networks has been one of the most popular techniques to compress models for e...
International audienceResource-constrained CNN implementations are subject to various reliability th...
Several hardware companies are proposing native Brain Float 16-bit (BF16) support for neural network...
In the neural network context, used in a variety of applications, binarised networks, which describe...
Today, computer vision (CV) problems are solved with unprecedented accuracy using convolutional neur...
International audienceIn this article, we propose a technique for improving the efficiency of convol...
This paper presents a convolutional neural network (CNN) accelerator that can skip zero weights and ...
As the utilization of CNN increases, many studies on lightweight, such as pruning, quantization, and...
This paper describes a hybrid weight-control strategy for the VLSI realization of programmable CNNs,...
FP8 is a natural progression for accelerating deep learning training inference beyond the 16-bit for...
The datapath bit-width of hardware accelerators for convolutional neural network (CNN) inference is ...
Convolutional Neural Networks (CNNs) are broadly used in safety-critical applications such as autono...
International audienceThe numerical format used for representing weights and activations plays a key...
The heavy burdens of computation and off-chip traffic impede deploying the large scale convolution n...
We propose a lightweight scheme where the formation of a data block is changed in such a way that it...
Quantization of neural networks has been one of the most popular techniques to compress models for e...
International audienceResource-constrained CNN implementations are subject to various reliability th...
Several hardware companies are proposing native Brain Float 16-bit (BF16) support for neural network...
In the neural network context, used in a variety of applications, binarised networks, which describe...
Today, computer vision (CV) problems are solved with unprecedented accuracy using convolutional neur...
International audienceIn this article, we propose a technique for improving the efficiency of convol...
This paper presents a convolutional neural network (CNN) accelerator that can skip zero weights and ...
As the utilization of CNN increases, many studies on lightweight, such as pruning, quantization, and...
This paper describes a hybrid weight-control strategy for the VLSI realization of programmable CNNs,...
FP8 is a natural progression for accelerating deep learning training inference beyond the 16-bit for...
The datapath bit-width of hardware accelerators for convolutional neural network (CNN) inference is ...
Convolutional Neural Networks (CNNs) are broadly used in safety-critical applications such as autono...