Existing approaches that partition a convolutional neural network (CNN) onto multiple accelerators and pipeline the training computations avoid or limit the use of stale weights, and they either underutilize accelerators or increase memory footprint. We explore the impact of stale weights on the statistical efficiency and performance in a pipelined backpropagation scheme that maximizes accelerator utilization and keeps memory overhead modest. We use LeNet-5, AlexNet, VGG and ResNet, and show that when pipelining is limited to early layers in a network, training with stale weights converges and results in models with comparable inference accuracies to those resulting from non-pipelined training. However, when pipelining is deeper in the netw...
Training of convolutional neural networks (CNNs) on embedded platforms to support on-device learning...
While deep neural networks (DNNs) have shown to be successful in several domains like computer visio...
This paper presents a convolutional neural network (CNN) accelerator that can skip zero weights and ...
Existing approaches that partition a convolutional neural network (CNN) onto multiple accelerators a...
The growth in size and complexity of convolutional neural networks (CNNs) is forcing the partitionin...
Deep neural network models are commonly used in various real-life applications due to their high pre...
To break the three lockings during backpropagation (BP) process for neural network training, multipl...
DNNs have been finding a growing number of applications including image classification, speech recog...
This work is focused on the pruning of some convolutional neural networks (CNNs) and improving their...
Deep Neural Networks (DNNs) have begun to permeate all corners of electronic society due to their hi...
Hardware accelerators for neural network inference can exploit common data properties for performanc...
Motivated by the goal of enabling energy-efficient and/or lower-cost hardware implementations of dee...
We provide an efficient implementation of the backpropagation algorithm, specialized to the case whe...
Convolutional neural networks (CNNs) outperform traditional machine learning algorithms across a wid...
Convolutional Neural Networks (CNNs) are brain-inspired computational models designed to recognize p...
Training of convolutional neural networks (CNNs) on embedded platforms to support on-device learning...
While deep neural networks (DNNs) have shown to be successful in several domains like computer visio...
This paper presents a convolutional neural network (CNN) accelerator that can skip zero weights and ...
Existing approaches that partition a convolutional neural network (CNN) onto multiple accelerators a...
The growth in size and complexity of convolutional neural networks (CNNs) is forcing the partitionin...
Deep neural network models are commonly used in various real-life applications due to their high pre...
To break the three lockings during backpropagation (BP) process for neural network training, multipl...
DNNs have been finding a growing number of applications including image classification, speech recog...
This work is focused on the pruning of some convolutional neural networks (CNNs) and improving their...
Deep Neural Networks (DNNs) have begun to permeate all corners of electronic society due to their hi...
Hardware accelerators for neural network inference can exploit common data properties for performanc...
Motivated by the goal of enabling energy-efficient and/or lower-cost hardware implementations of dee...
We provide an efficient implementation of the backpropagation algorithm, specialized to the case whe...
Convolutional neural networks (CNNs) outperform traditional machine learning algorithms across a wid...
Convolutional Neural Networks (CNNs) are brain-inspired computational models designed to recognize p...
Training of convolutional neural networks (CNNs) on embedded platforms to support on-device learning...
While deep neural networks (DNNs) have shown to be successful in several domains like computer visio...
This paper presents a convolutional neural network (CNN) accelerator that can skip zero weights and ...