Overparametrized Deep Neural Networks (DNNs) often achieve astounding performances, but may potentially result in severe generalization error. Recently, the relation between the sharpness of the loss landscape and the generalization error has been established by Foret et al. (2020), in which the Sharpness Aware Minimizer (SAM) was proposed to mitigate the degradation of the generalization. Unfortunately, SAM s computational cost is roughly double that of base optimizers, such as Stochastic Gradient Descent (SGD). This paper thus proposes Efficient Sharpness Aware Minimizer (ESAM), which boosts SAM s efficiency at no cost to its generalization performance. ESAM includes two novel and efficient training strategies-StochasticWeight Perturbatio...
The loss surface of an overparameterized neural network (NN) possesses many global minima of zero tr...
The general features of the optimization problem for the case of overparametrized nonlinear networks...
Recently, brain-inspired spiking neuron networks (SNNs) have attracted widespread research interest ...
Modern deep neural networks (DNNs) have achieved state-of-the-art performances but are typically ove...
To fully uncover the great potential of deep neural networks (DNNs), various learning algorithms hav...
Network quantization has gained increasing attention since it can significantly reduce the model siz...
Recently, flat-minima optimizers, which seek to find parameters in low loss neighborhoods, have been...
Energy-efficient deep neural network (DNN) accelerators are prone to non-idealities that degrade DNN...
How to train deep neural networks (DNNs) to generalize well is a central concern in deep learning, e...
In an effort to improve generalization in deep learning and automate the process of learning rate sc...
We consider Sharpness-Aware Minimization (SAM), a gradient-based optimization method for deep networ...
Optimization is the key component of deep learning. Increasing depth, which is vital for reaching a...
The success of deep learning has shown impressive empirical breakthroughs, but many theoretical ques...
In Theory IIb we characterize with a mix of theory and experiments the optimization of deep convolut...
In the last decade or so, deep learning has revolutionized entire domains of machine learning. Neura...
The loss surface of an overparameterized neural network (NN) possesses many global minima of zero tr...
The general features of the optimization problem for the case of overparametrized nonlinear networks...
Recently, brain-inspired spiking neuron networks (SNNs) have attracted widespread research interest ...
Modern deep neural networks (DNNs) have achieved state-of-the-art performances but are typically ove...
To fully uncover the great potential of deep neural networks (DNNs), various learning algorithms hav...
Network quantization has gained increasing attention since it can significantly reduce the model siz...
Recently, flat-minima optimizers, which seek to find parameters in low loss neighborhoods, have been...
Energy-efficient deep neural network (DNN) accelerators are prone to non-idealities that degrade DNN...
How to train deep neural networks (DNNs) to generalize well is a central concern in deep learning, e...
In an effort to improve generalization in deep learning and automate the process of learning rate sc...
We consider Sharpness-Aware Minimization (SAM), a gradient-based optimization method for deep networ...
Optimization is the key component of deep learning. Increasing depth, which is vital for reaching a...
The success of deep learning has shown impressive empirical breakthroughs, but many theoretical ques...
In Theory IIb we characterize with a mix of theory and experiments the optimization of deep convolut...
In the last decade or so, deep learning has revolutionized entire domains of machine learning. Neura...
The loss surface of an overparameterized neural network (NN) possesses many global minima of zero tr...
The general features of the optimization problem for the case of overparametrized nonlinear networks...
Recently, brain-inspired spiking neuron networks (SNNs) have attracted widespread research interest ...