This paper proposes an improved stochastic second order learning algorithm for supervised neural network training. The proposed algorithm, named bounded stochastic diagonal Levenberg-Marquardt (B-SDLM), utilizes both gradient and curvature information to achieve fast convergence while requiring only minimal computational overhead than the stochastic gradient descent (SGD) method. B-SDLM has only a single hyperparameter as opposed to most other learning algorithms that suffer from the hyperparameter overfitting problem due to having more hyperparameters to be tuned. Experiments using the multilayer perceptron (MLP) and convolutional neural network (CNN) models have shown that B-SDLM outperforms other learning algorithms with regard to the cl...
Neural networks are an important class of highly flexible and powerful models inspired by the struct...
Three main requirements of a successful application of deep learning are the network architecture, a...
Since the discovery of the back-propagation method, many modified and new algorithms have been propo...
This paper proposes an efficient asynchronous stochastic second order learning algorithm for distrib...
Convolutional neural networks (CNNs) are a variant of deep neural networks (DNNs) optimized for visu...
In this dissertation, we are concerned with the advancement of optimization algorithms for training ...
© 2017 IEEE. Deep structured of Convolutional Neural Networks (CNN) has recently gained intense atte...
© 2017 IEEE. Optimization is important in neural networks to iteratively update weights for pattern ...
Deep neural network models can achieve greater performance in numerous machine learning tasks by rai...
This paper proposes a new family of algorithms for training neural networks (NNs). These...
In stochastic gradient descent (SGD) and its variants, the optimized gradient estimators may be as e...
We propose an asynchronous version of stochastic secondorder optimization algorithm for parallel dis...
Many connectionist learning algorithms consists of minimizing a cost of the form C(w) = E(J(z; w)) ...
Abstract. Recently, we proposed to transform the outputs of each hidden neu-ron in a multi-layer per...
Stochastic Gradient Descent algorithms (SGD) remain a popular optimizer for deep learning networks a...
Neural networks are an important class of highly flexible and powerful models inspired by the struct...
Three main requirements of a successful application of deep learning are the network architecture, a...
Since the discovery of the back-propagation method, many modified and new algorithms have been propo...
This paper proposes an efficient asynchronous stochastic second order learning algorithm for distrib...
Convolutional neural networks (CNNs) are a variant of deep neural networks (DNNs) optimized for visu...
In this dissertation, we are concerned with the advancement of optimization algorithms for training ...
© 2017 IEEE. Deep structured of Convolutional Neural Networks (CNN) has recently gained intense atte...
© 2017 IEEE. Optimization is important in neural networks to iteratively update weights for pattern ...
Deep neural network models can achieve greater performance in numerous machine learning tasks by rai...
This paper proposes a new family of algorithms for training neural networks (NNs). These...
In stochastic gradient descent (SGD) and its variants, the optimized gradient estimators may be as e...
We propose an asynchronous version of stochastic secondorder optimization algorithm for parallel dis...
Many connectionist learning algorithms consists of minimizing a cost of the form C(w) = E(J(z; w)) ...
Abstract. Recently, we proposed to transform the outputs of each hidden neu-ron in a multi-layer per...
Stochastic Gradient Descent algorithms (SGD) remain a popular optimizer for deep learning networks a...
Neural networks are an important class of highly flexible and powerful models inspired by the struct...
Three main requirements of a successful application of deep learning are the network architecture, a...
Since the discovery of the back-propagation method, many modified and new algorithms have been propo...