Several methods for training feed-forward neural networks require second order information from the Hessian matrix of the error function. Although it is possible to calculate the Hessian matrix exactly it is often not desirable because of the computation and memory requirements involved. Some learning techniques does, however, only need the Hessian matrix times a vector. This paper presents a method to calculate the Hessian matrix times a vector inO(N) time, whereN is the number of variables in the network. This is in the same order as the calculation of the gradient to the error function. The usefulness of this algorithm is demonstrated by improvement of existing learning techniques.
A statistically-based algorithm for pruning weights from feed-forward networks is presented. This a...
We derive two second-order algorithms, based on the conjugate gradient method, for online training o...
Efficiently approximating local curvature information of the loss function is a key tool for optimiz...
Several methods for training feed-forward neural networks require second order information from the...
Since the discovery of the back-propagation method, many modified and new algorithms have been propo...
We extend here a general mathematical model for feed-forward neural networks. Such a network is repr...
Introduction Training algorithms for Multilayer Perceptrons optimize the set of W weights and biase...
The Levenberg-Marquardt (LM) learning algorithm is a popular algorithm for training neural networks;...
For training fully-connected neural networks (FCNNs), we propose a practical approximate second-orde...
We analyse the dynamics of a number of second order on-line learning algorithms training multi-layer...
The Levenberg-Marquardt (LM) learning algorithm is a popular algo-rithm for training neural networks...
Just storing the Hessian H (the matrix of second derivatives a2E/aw, aw, of the error E with respec...
We propose a very simple, and well principled wayofcomputing the optimal step size in gradient desce...
Minimization methods for training feed-forward networks with Backpropagation are compared. Feedforwa...
A method and apparatus for supervised neural learning of time dependent trajectories exploits the co...
A statistically-based algorithm for pruning weights from feed-forward networks is presented. This a...
We derive two second-order algorithms, based on the conjugate gradient method, for online training o...
Efficiently approximating local curvature information of the loss function is a key tool for optimiz...
Several methods for training feed-forward neural networks require second order information from the...
Since the discovery of the back-propagation method, many modified and new algorithms have been propo...
We extend here a general mathematical model for feed-forward neural networks. Such a network is repr...
Introduction Training algorithms for Multilayer Perceptrons optimize the set of W weights and biase...
The Levenberg-Marquardt (LM) learning algorithm is a popular algorithm for training neural networks;...
For training fully-connected neural networks (FCNNs), we propose a practical approximate second-orde...
We analyse the dynamics of a number of second order on-line learning algorithms training multi-layer...
The Levenberg-Marquardt (LM) learning algorithm is a popular algo-rithm for training neural networks...
Just storing the Hessian H (the matrix of second derivatives a2E/aw, aw, of the error E with respec...
We propose a very simple, and well principled wayofcomputing the optimal step size in gradient desce...
Minimization methods for training feed-forward networks with Backpropagation are compared. Feedforwa...
A method and apparatus for supervised neural learning of time dependent trajectories exploits the co...
A statistically-based algorithm for pruning weights from feed-forward networks is presented. This a...
We derive two second-order algorithms, based on the conjugate gradient method, for online training o...
Efficiently approximating local curvature information of the loss function is a key tool for optimiz...