Low-precision formats have recently driven major breakthroughs in neural network (NN) training and inference by reducing the memory footprint of the NN models and improving the energy efficiency of the underlying hardware architectures. Narrow integer data types have been vastly investigated for NN inference and have successfully been pushed to the extreme of ternary and binary representations. In contrast, most training-oriented platforms use at least 16-bit floating-point (FP) formats. Lower-precision data types such as 8-bit FP formats and mixed-precision techniques have only recently been explored in hardware implementations. We present MiniFloat-NN, a RISC-V instruction set architecture extension for low-precision NN training, providin...
sis presents FPRaker, a processing element for composing training accelerators. Training manipulates...
Applying machine learning to various applications has gained significant momentum in recent years. H...
Approximate computing has emerged as a promising approach to energy-efficient design of digital syst...
Low-precision formats have recently driven major breakthroughs in neural network (NN) training and i...
Due to limited size, cost and power, embedded devices do not offer the same computational throughput...
International audienceThe most compute-intensive stage of deep neural network (DNN) training is matr...
The high computational complexity, memory footprints, and energy requirements of machine learning mo...
Heavily quantized fixed-point arithmetic is becoming a common approach to deploy Convolutional Neura...
Mixed-precision (MP) arithmetic combining both single- and half-precision operands has been successf...
Nowadays, two groundbreaking factors are emerging in neural networks. Firstly, there is the RISC-V o...
Low bit-width Quantized Neural Networks (QNNs) enable deployment of complex machine learning models ...
Several hardware companies are proposing native Brain Float 16-bit (BF16) support for neural network...
The proliferation of embedded Neural Processing Units (NPUs) is enabling the adoption of Tiny Machin...
The unprecedented growth in DNN model complexity, size and the amount of training data have led to a...
FP8 is a natural progression for accelerating deep learning training inference beyond the 16-bit for...
sis presents FPRaker, a processing element for composing training accelerators. Training manipulates...
Applying machine learning to various applications has gained significant momentum in recent years. H...
Approximate computing has emerged as a promising approach to energy-efficient design of digital syst...
Low-precision formats have recently driven major breakthroughs in neural network (NN) training and i...
Due to limited size, cost and power, embedded devices do not offer the same computational throughput...
International audienceThe most compute-intensive stage of deep neural network (DNN) training is matr...
The high computational complexity, memory footprints, and energy requirements of machine learning mo...
Heavily quantized fixed-point arithmetic is becoming a common approach to deploy Convolutional Neura...
Mixed-precision (MP) arithmetic combining both single- and half-precision operands has been successf...
Nowadays, two groundbreaking factors are emerging in neural networks. Firstly, there is the RISC-V o...
Low bit-width Quantized Neural Networks (QNNs) enable deployment of complex machine learning models ...
Several hardware companies are proposing native Brain Float 16-bit (BF16) support for neural network...
The proliferation of embedded Neural Processing Units (NPUs) is enabling the adoption of Tiny Machin...
The unprecedented growth in DNN model complexity, size and the amount of training data have led to a...
FP8 is a natural progression for accelerating deep learning training inference beyond the 16-bit for...
sis presents FPRaker, a processing element for composing training accelerators. Training manipulates...
Applying machine learning to various applications has gained significant momentum in recent years. H...
Approximate computing has emerged as a promising approach to energy-efficient design of digital syst...