We introduce canonical weight normalization for convolutional neural networks. Inspired by the canonical tensor decomposition, we express the weight tensors in so-called canonical networks as scaled sums of outer vector products. In particular, we train network weights in the decomposed form, where scale weights are optimized separately for each mode. Additionally, similarly to weight normalization, we include a global scaling parameter. We study the initialization of the canonical form by running the power method and by drawing randomly from Gaussian or uniform distributions. Our results indicate that we can replace the power method with cheaper initializations drawn from standard distributions. The canonical re-parametrization leads to co...
The success of deep neural networks is in part due to the use of normalization layers. Normalization...
It is a central problem in both statistics and computer science to understand the theoretical founda...
Analysing Generalisation Error Bounds for Convolutional Neural Networks Abstract: Convolutional neur...
We introduce canonical weight normalization for convolutional neural networks. Inspired by the canon...
© 2017 IEEE. Training deep neural networks is difficult for the pathological curvature problem. Re-p...
International audienceModern neural networks are over-parametrized. In particular, each rectified li...
In the last years, deep neural networks have revolutionized machine learning tasks. However, the des...
In many information processing systems, it may be desirable to ensure that any change of the input, ...
Abstract We present weight normalization: a reparameterization of the weight vectors in a neural net...
Several methods of normalizing convolution kernels have been proposed in the literature to train con...
Neural networks have gained widespread use in many machine learning tasks due to their state-of-the-...
The convolutional neural network is a very important model of deep learning. It can help avoid the e...
In recent years, a variety of normalization methods have been proposed to help training neural netwo...
Modern iterations of deep learning models contain millions (billions) of unique parameters-each repr...
Data symmetries have been used to successfully learn robust and optimal representation either via au...
The success of deep neural networks is in part due to the use of normalization layers. Normalization...
It is a central problem in both statistics and computer science to understand the theoretical founda...
Analysing Generalisation Error Bounds for Convolutional Neural Networks Abstract: Convolutional neur...
We introduce canonical weight normalization for convolutional neural networks. Inspired by the canon...
© 2017 IEEE. Training deep neural networks is difficult for the pathological curvature problem. Re-p...
International audienceModern neural networks are over-parametrized. In particular, each rectified li...
In the last years, deep neural networks have revolutionized machine learning tasks. However, the des...
In many information processing systems, it may be desirable to ensure that any change of the input, ...
Abstract We present weight normalization: a reparameterization of the weight vectors in a neural net...
Several methods of normalizing convolution kernels have been proposed in the literature to train con...
Neural networks have gained widespread use in many machine learning tasks due to their state-of-the-...
The convolutional neural network is a very important model of deep learning. It can help avoid the e...
In recent years, a variety of normalization methods have been proposed to help training neural netwo...
Modern iterations of deep learning models contain millions (billions) of unique parameters-each repr...
Data symmetries have been used to successfully learn robust and optimal representation either via au...
The success of deep neural networks is in part due to the use of normalization layers. Normalization...
It is a central problem in both statistics and computer science to understand the theoretical founda...
Analysing Generalisation Error Bounds for Convolutional Neural Networks Abstract: Convolutional neur...