© 2017 IEEE. Training deep neural networks is difficult for the pathological curvature problem. Re-parameterization is an effective way to relieve the problem by learning the curvature approximately or constraining the solutions of weights with good properties for optimization. This paper proposes to reparameterize the input weight of each neuron in deep neural networks by normalizing it with zero-mean and unit-norm, followed by a learnable scalar parameter to adjust the norm of the weight. This technique effectively stabilizes the distribution implicitly. Besides, it improves the conditioning of the optimization problem and thus accelerates the training of deep neural networks. It can be wrapped as a linear module in practice and plugged i...
Orthogonal matrix has shown advantages in training Recurrent Neural Networks (RNNs), but such matrix...
We introduce canonical weight normalization for convolutional neural networks. Inspired by the canon...
Recent years have witnessed the success of deep neural networks in dealing with a plenty of practica...
Abstract We present weight normalization: a reparameterization of the weight vectors in a neural net...
The highly non-linear nature of deep neural networks causes them to be susceptible to adversarial ex...
International audienceModern neural networks are over-parametrized. In particular, each rectified li...
The success of deep neural networks is in part due to the use of normalization layers. Normalization...
This paper proposes a set of new error criteria and a learning approach, called Adaptive Normalized ...
Optimization is the key component of deep learning. Increasing depth, which is vital for reaching a...
In the recent decade, deep neural networks have solved ever more complex tasks across many fronts in...
A new method of initializing the weights in deep neural networks is proposed. The method follows two...
How to train deep neural networks (DNNs) to generalize well is a central concern in deep learning, e...
Batch Normalization (BN) is an essential component of the Deep Neural Networks (DNNs) architectures....
It is a central problem in both statistics and computer science to understand the theoretical founda...
© Copyright 2016, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rig...
Orthogonal matrix has shown advantages in training Recurrent Neural Networks (RNNs), but such matrix...
We introduce canonical weight normalization for convolutional neural networks. Inspired by the canon...
Recent years have witnessed the success of deep neural networks in dealing with a plenty of practica...
Abstract We present weight normalization: a reparameterization of the weight vectors in a neural net...
The highly non-linear nature of deep neural networks causes them to be susceptible to adversarial ex...
International audienceModern neural networks are over-parametrized. In particular, each rectified li...
The success of deep neural networks is in part due to the use of normalization layers. Normalization...
This paper proposes a set of new error criteria and a learning approach, called Adaptive Normalized ...
Optimization is the key component of deep learning. Increasing depth, which is vital for reaching a...
In the recent decade, deep neural networks have solved ever more complex tasks across many fronts in...
A new method of initializing the weights in deep neural networks is proposed. The method follows two...
How to train deep neural networks (DNNs) to generalize well is a central concern in deep learning, e...
Batch Normalization (BN) is an essential component of the Deep Neural Networks (DNNs) architectures....
It is a central problem in both statistics and computer science to understand the theoretical founda...
© Copyright 2016, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rig...
Orthogonal matrix has shown advantages in training Recurrent Neural Networks (RNNs), but such matrix...
We introduce canonical weight normalization for convolutional neural networks. Inspired by the canon...
Recent years have witnessed the success of deep neural networks in dealing with a plenty of practica...