abstract: Deep neural networks (DNNs) have had tremendous success in a variety of statistical learning applications due to their vast expressive power. Most applications run DNNs on the cloud on parallelized architectures. There is a need for for efficient DNN inference on edge with low precision hardware and analog accelerators. To make trained models more robust for this setting, quantization and analog compute noise are modeled as weight space perturbations to DNNs and an information theoretic regularization scheme is used to penalize the KL-divergence between perturbed and unperturbed models. This regularizer has similarities to both natural gradient descent and knowledge distillation, but has the advantage of explicitly promot...
Deep learning has seen tremendous growth, largely fueled by more powerful computers, the availabilit...
Currently neural networks run as software, which typically requires expensive GPU resources. As the ...
Sampling theory investigates signal recovery from its partial information, and one of the simplest a...
First we present a proof that convolutional neural networks (CNNs) with max-norm regularization, max...
Deep neural networks have achieved state-of-the-art performance for various machine learning tasks i...
Deep generative neural networks such as the Deep Belief Network and Deep Boltzmann Machines have bee...
Energy-efficient deep neural network (DNN) accelerators are prone to non-idealities that degrade DNN...
Deep neural networks (DNNs) have surpassed human-level accuracy in various fields, including object ...
This work explores the impact of various design and training choices on the resilience of a neural n...
1This work seeks to answer the question: as the (near-) orthogonality of weights is found to be a fa...
Quantization of Convolutional Neural Networks (CNNs) is a common approach to ease the computational ...
In recent years, Deep Neural Networks (DNNs) have become an area of high interest due to it's ground...
Regularizing the gradient norm of the output of a neural network is a powerful technique, rediscover...
This book develops an effective theory approach to understanding deep neural networks of practical r...
Deep Neural Networks (DNNs) needs to be both efficient and robust for practical uses. Quantization a...
Deep learning has seen tremendous growth, largely fueled by more powerful computers, the availabilit...
Currently neural networks run as software, which typically requires expensive GPU resources. As the ...
Sampling theory investigates signal recovery from its partial information, and one of the simplest a...
First we present a proof that convolutional neural networks (CNNs) with max-norm regularization, max...
Deep neural networks have achieved state-of-the-art performance for various machine learning tasks i...
Deep generative neural networks such as the Deep Belief Network and Deep Boltzmann Machines have bee...
Energy-efficient deep neural network (DNN) accelerators are prone to non-idealities that degrade DNN...
Deep neural networks (DNNs) have surpassed human-level accuracy in various fields, including object ...
This work explores the impact of various design and training choices on the resilience of a neural n...
1This work seeks to answer the question: as the (near-) orthogonality of weights is found to be a fa...
Quantization of Convolutional Neural Networks (CNNs) is a common approach to ease the computational ...
In recent years, Deep Neural Networks (DNNs) have become an area of high interest due to it's ground...
Regularizing the gradient norm of the output of a neural network is a powerful technique, rediscover...
This book develops an effective theory approach to understanding deep neural networks of practical r...
Deep Neural Networks (DNNs) needs to be both efficient and robust for practical uses. Quantization a...
Deep learning has seen tremendous growth, largely fueled by more powerful computers, the availabilit...
Currently neural networks run as software, which typically requires expensive GPU resources. As the ...
Sampling theory investigates signal recovery from its partial information, and one of the simplest a...