Recent advances in deep learning have been driven by ever-increasing model sizes, with networks growing to millions or even billions of parameters. Such enormous models call for fast and energy-efficient hardware accelerators. We study the potential of Analog AI accelerators based on Non-Volatile Memory, in particular Phase Change Memory (PCM), for software-equivalent accurate inference of natural language processing applications. We demonstrate a path to software-equivalent accuracy for the GLUE benchmark on BERT (Bidirectional Encoder Representations from Transformers), by combining noise-aware training to combat inherent PCM drift and noise sources, together with reduced-precision digital attention-block computation down to INT6
The machine learning revolution under way brought us neural networks that outperform humans at a var...
Deep neural networks (DNNs) have achieved unprecedented capabilities in tasks such as analysis and r...
The objective of this research is to accelerate deep neural networks (DNNs) with emerging non-volati...
Non-volatile analog memory devices such as phase-change memory (PCM) enable designing dedicated conn...
Matrix-Vector Multiplications (MVMs) represent a heavy workload for both training and inference in D...
In-memory computing using resistive memory devices is a promising non-von Neumann approach for makin...
Neuromorphic computing embraces the “device history” offered by many analog non-volatile memory (NVM...
There is an urgent need for compact, fast, and power-efficient hardware implementations of state-of-...
Brain-inspired computation promises a paradigm shift in information processing, both in terms of its...
The search for energy efficient circuital implementations of neural networks has led to the explorat...
Memory devices, such as the phase change memory (PCM), have recently shown significant breakthroughs...
Deep neural networks (DNNs) have revolutionized the field of artificial intelligence and have achiev...
Recently, analog compute-in-memory (CIM) architectures based on emerging analog non-volatile memory ...
The last decade has witnessed the breakthrough of deep neural networks (DNNs) in many fields. With t...
The neural computation field had finally delivered on its promises in 2013 when the University of To...
The machine learning revolution under way brought us neural networks that outperform humans at a var...
Deep neural networks (DNNs) have achieved unprecedented capabilities in tasks such as analysis and r...
The objective of this research is to accelerate deep neural networks (DNNs) with emerging non-volati...
Non-volatile analog memory devices such as phase-change memory (PCM) enable designing dedicated conn...
Matrix-Vector Multiplications (MVMs) represent a heavy workload for both training and inference in D...
In-memory computing using resistive memory devices is a promising non-von Neumann approach for makin...
Neuromorphic computing embraces the “device history” offered by many analog non-volatile memory (NVM...
There is an urgent need for compact, fast, and power-efficient hardware implementations of state-of-...
Brain-inspired computation promises a paradigm shift in information processing, both in terms of its...
The search for energy efficient circuital implementations of neural networks has led to the explorat...
Memory devices, such as the phase change memory (PCM), have recently shown significant breakthroughs...
Deep neural networks (DNNs) have revolutionized the field of artificial intelligence and have achiev...
Recently, analog compute-in-memory (CIM) architectures based on emerging analog non-volatile memory ...
The last decade has witnessed the breakthrough of deep neural networks (DNNs) in many fields. With t...
The neural computation field had finally delivered on its promises in 2013 when the University of To...
The machine learning revolution under way brought us neural networks that outperform humans at a var...
Deep neural networks (DNNs) have achieved unprecedented capabilities in tasks such as analysis and r...
The objective of this research is to accelerate deep neural networks (DNNs) with emerging non-volati...