Several hardware companies are proposing native Brain Float 16-bit (BF16) support for neural network training. The usage of Mixed Precision (MP) arithmetic with floating-point 32-bit (FP32) and 16-bit half-precision aims at improving memory and floating-point operations throughput, allowing faster training of bigger models. This paper proposes a binary analysis tool enabling the emulation of lower precision numerical formats in Neural Network implementation without the need for hardware support. This tool is used to analyze BF16 usage in the training phase of a 3D Generative Adversarial Network (3DGAN) simulating High Energy Physics detectors. The binary tool allows us to confirm that BF16 can provide results with similar accuracy as the fu...
Detailed simulation is one of the most expensive tasks, in terms of time and computing resources for...
Approximate computing has emerged as a promising approach to energy-efficient design of digital syst...
Low-precision formats have recently driven major breakthroughs in neural network (NN) training and i...
Several hardware companies are proposing native Brain Float 16-bit (BF16) support for neural network...
Mixed-precision (MP) arithmetic combining both single- and half-precision operands has been successf...
Due to limited size, cost and power, embedded devices do not offer the same computational throughput...
We present the implementation of binary and ternary neural networks in the hls4ml library, designed ...
International audienceGraphics Processing Units (GPUs) offer the possibility to execute floating-poi...
An important aspect of modern automation is machine learning. Specifically, neural networks are used...
The subject of this thesis is neural network acceleration with the goal of reducing the number of fl...
Modern computational tasks are often required to not only guarantee predefined accuracy, but get the...
Applying machine learning to various applications has gained significant momentum in recent years. H...
This paper proposes a novel approximate bfloat16 multiplier with on-the-fly adjustable accuracy for ...
This electronic version was submitted by the student author. The certified thesis is available in th...
Machine learning has risen to prominence in recent years thanks to advancements in computer technolo...
Detailed simulation is one of the most expensive tasks, in terms of time and computing resources for...
Approximate computing has emerged as a promising approach to energy-efficient design of digital syst...
Low-precision formats have recently driven major breakthroughs in neural network (NN) training and i...
Several hardware companies are proposing native Brain Float 16-bit (BF16) support for neural network...
Mixed-precision (MP) arithmetic combining both single- and half-precision operands has been successf...
Due to limited size, cost and power, embedded devices do not offer the same computational throughput...
We present the implementation of binary and ternary neural networks in the hls4ml library, designed ...
International audienceGraphics Processing Units (GPUs) offer the possibility to execute floating-poi...
An important aspect of modern automation is machine learning. Specifically, neural networks are used...
The subject of this thesis is neural network acceleration with the goal of reducing the number of fl...
Modern computational tasks are often required to not only guarantee predefined accuracy, but get the...
Applying machine learning to various applications has gained significant momentum in recent years. H...
This paper proposes a novel approximate bfloat16 multiplier with on-the-fly adjustable accuracy for ...
This electronic version was submitted by the student author. The certified thesis is available in th...
Machine learning has risen to prominence in recent years thanks to advancements in computer technolo...
Detailed simulation is one of the most expensive tasks, in terms of time and computing resources for...
Approximate computing has emerged as a promising approach to energy-efficient design of digital syst...
Low-precision formats have recently driven major breakthroughs in neural network (NN) training and i...