Efficient inference of Deep Neural Networks (DNNs) is essential to making AI ubiquitous. Two important algorithmic techniques have shown promise for enabling efficient inference - sparsity and binarization. These techniques translate into weight sparsity and weight repetition at the hardware-software level allowing the deployment of DNNs with critically low power and latency requirements. We propose a new method called signed-binary networks to improve further efficiency (by exploiting both weight sparsity and weight repetition) while maintaining similar accuracy. Our method achieves comparable accuracy on ImageNet and CIFAR10 datasets with binary and can lead to $>69\%$ sparsity. We observe real speedup when deploying these models on gener...
Network binarization (i.e., binary neural networks, BNNs) can efficiently compress deep neural netwo...
Deep neural networks have achieved impressive results in computer vision and machine learning. Unfor...
Binarization of feature representation is critical for Binarized Neural Networks (BNNs). Currently, ...
Binary neural networks (BNNs) have attracted broad research interest due to their efficient storage ...
DNNs have been finding a growing number of applications including image classification, speech recog...
In trained deep neural networks, unstructured pruning can reduce redundant weights to lower storage ...
Thesis (Ph.D.)--University of Washington, 2020The recent renaissance of deep neural networks has lea...
As deep neural networks grow larger, they suffer from a huge number of weights, and thus reducing th...
How to train a binary neural network (BinaryNet) with both high compression rate and high accuracy o...
This paper presents a convolutional neural network (CNN) accelerator that can skip zero weights and ...
Hardware accelerators for neural network inference can exploit common data properties for performanc...
Deep neural networks (DNNs) are increasing their presence in a wide range of applications, and their...
The growing energy and performance costs of deep learning have driven the community to reduce the si...
The ever-growing computational demands of increasingly complex machine learning models frequently ne...
Deep Neural Networks (DNNs) have become ubiquitous, achieving state-of-the-art results across a wide...
Network binarization (i.e., binary neural networks, BNNs) can efficiently compress deep neural netwo...
Deep neural networks have achieved impressive results in computer vision and machine learning. Unfor...
Binarization of feature representation is critical for Binarized Neural Networks (BNNs). Currently, ...
Binary neural networks (BNNs) have attracted broad research interest due to their efficient storage ...
DNNs have been finding a growing number of applications including image classification, speech recog...
In trained deep neural networks, unstructured pruning can reduce redundant weights to lower storage ...
Thesis (Ph.D.)--University of Washington, 2020The recent renaissance of deep neural networks has lea...
As deep neural networks grow larger, they suffer from a huge number of weights, and thus reducing th...
How to train a binary neural network (BinaryNet) with both high compression rate and high accuracy o...
This paper presents a convolutional neural network (CNN) accelerator that can skip zero weights and ...
Hardware accelerators for neural network inference can exploit common data properties for performanc...
Deep neural networks (DNNs) are increasing their presence in a wide range of applications, and their...
The growing energy and performance costs of deep learning have driven the community to reduce the si...
The ever-growing computational demands of increasingly complex machine learning models frequently ne...
Deep Neural Networks (DNNs) have become ubiquitous, achieving state-of-the-art results across a wide...
Network binarization (i.e., binary neural networks, BNNs) can efficiently compress deep neural netwo...
Deep neural networks have achieved impressive results in computer vision and machine learning. Unfor...
Binarization of feature representation is critical for Binarized Neural Networks (BNNs). Currently, ...