International audienceResource requirements for hardware acceleration of neural networks inference is notoriously high, both in terms of computation and storage. One way to mitigate this issue is to quantize parameters and activations. This is usually done by scaling and centering the distributions of weights and activations, on a kernel per kernel basis, so that a low-precision binary integer representation can be used. This work studies low-precision logarithmic number system (LNS) as an efficient alternative. Firstly, LNS has more dynamic than fixed-point for the same number of bits. Thus, when quantizing MNIST and CIFAR reference networks without retraining, the smallest format size achieving top-1 accuracy comparable to floating-point ...
Low-precision neural network models are crucial for reducing the memory footprint and computational...
A low cost, high-speed architecture for the computation of the binary logarithm is proposed. It is b...
International audienceThe most compute-intensive stage of deep neural network (DNN) training is matr...
International audienceResource requirements for hardware acceleration of neural networks inference i...
Convolutional Neural Networks (CNN) have become a popular solution for computer vision problems. How...
International audienceEconomical hardware often uses a FiXed-point Number System (FXNS), whose const...
International audienceEconomical hardware often uses a FiXed-point Number System (FXNS), whose const...
International audienceThe ever-growing cost of both training and inference for state-of-the-art neur...
For energy efficiency, many low-bit quantization methods for deep neural networks (DNNs) have been p...
Low-precision neural networks represent both weights and activations with few bits, drastically redu...
Energy-efficient computing and ultra-low-power operation are requirements for many application areas...
The increase in sophistication of neural network models in recent years has exponentially expanded m...
Today, computer vision (CV) problems are solved with unprecedented accuracy using convolutional neur...
Research has shown that deep neural networks contain significant redundancy, and thus that high clas...
We present the implementation of binary and ternary neural networks in the hls4ml library, designed ...
Low-precision neural network models are crucial for reducing the memory footprint and computational...
A low cost, high-speed architecture for the computation of the binary logarithm is proposed. It is b...
International audienceThe most compute-intensive stage of deep neural network (DNN) training is matr...
International audienceResource requirements for hardware acceleration of neural networks inference i...
Convolutional Neural Networks (CNN) have become a popular solution for computer vision problems. How...
International audienceEconomical hardware often uses a FiXed-point Number System (FXNS), whose const...
International audienceEconomical hardware often uses a FiXed-point Number System (FXNS), whose const...
International audienceThe ever-growing cost of both training and inference for state-of-the-art neur...
For energy efficiency, many low-bit quantization methods for deep neural networks (DNNs) have been p...
Low-precision neural networks represent both weights and activations with few bits, drastically redu...
Energy-efficient computing and ultra-low-power operation are requirements for many application areas...
The increase in sophistication of neural network models in recent years has exponentially expanded m...
Today, computer vision (CV) problems are solved with unprecedented accuracy using convolutional neur...
Research has shown that deep neural networks contain significant redundancy, and thus that high clas...
We present the implementation of binary and ternary neural networks in the hls4ml library, designed ...
Low-precision neural network models are crucial for reducing the memory footprint and computational...
A low cost, high-speed architecture for the computation of the binary logarithm is proposed. It is b...
International audienceThe most compute-intensive stage of deep neural network (DNN) training is matr...