First-order methods such as stochastic gradient descent (SGD) have recently become popular optimization methods to train deep neural networks (DNNs) for good generalization; however, they need a long training time. Second-order methods which can lower the training time are scarcely used on account of their overpriced computing cost to obtain the second-order information. Thus, many works have approximated the Hessian matrix to cut the cost of computing while the approximate Hessian matrix has large deviation. In this paper, we explore the convexity of the Hessian matrix of partial parameters and propose the damped Newton stochastic gradient descent (DN-SGD) method and stochastic gradient descent damped Newton (SGD-DN) method to train DNNs f...
© 2017 IEEE. Deep structured of Convolutional Neural Networks (CNN) has recently gained intense atte...
Since the discovery of the back-propagation method, many modified and new algorithms have been propo...
In recent years, neural networks, as part of deep learning, became pop- ular because the ability to...
In this dissertation, we are concerned with the advancement of optimization algorithms for training ...
While first-order methods are popular for solving optimization problems that arise in large-scale de...
We propose a fast second-order method that can be used as a drop-in replacement for current deep lea...
Neural networks are an important class of highly flexible and powerful models inspired by the struct...
This paper proposes a new family of algorithms for training neural networks (NNs). These...
This paper proposes an improved stochastic second order learning algorithm for supervised neural net...
Physics-informed neural networks (PINNs) have effectively been demonstrated in solving forward and i...
Deep neural networks have achieved significant success in a number of challenging engineering proble...
Stochastic Gradient Descent algorithms (SGD) remain a popular optimizer for deep learning networks a...
Hessian-free (HF) optimization has been successfully used for training deep au-toencoders and recurr...
In the past decade, neural networks have demonstrated impressive performance in supervised learning....
Deep Learning learning has recently become one of the most predominantly used techniques in the fiel...
© 2017 IEEE. Deep structured of Convolutional Neural Networks (CNN) has recently gained intense atte...
Since the discovery of the back-propagation method, many modified and new algorithms have been propo...
In recent years, neural networks, as part of deep learning, became pop- ular because the ability to...
In this dissertation, we are concerned with the advancement of optimization algorithms for training ...
While first-order methods are popular for solving optimization problems that arise in large-scale de...
We propose a fast second-order method that can be used as a drop-in replacement for current deep lea...
Neural networks are an important class of highly flexible and powerful models inspired by the struct...
This paper proposes a new family of algorithms for training neural networks (NNs). These...
This paper proposes an improved stochastic second order learning algorithm for supervised neural net...
Physics-informed neural networks (PINNs) have effectively been demonstrated in solving forward and i...
Deep neural networks have achieved significant success in a number of challenging engineering proble...
Stochastic Gradient Descent algorithms (SGD) remain a popular optimizer for deep learning networks a...
Hessian-free (HF) optimization has been successfully used for training deep au-toencoders and recurr...
In the past decade, neural networks have demonstrated impressive performance in supervised learning....
Deep Learning learning has recently become one of the most predominantly used techniques in the fiel...
© 2017 IEEE. Deep structured of Convolutional Neural Networks (CNN) has recently gained intense atte...
Since the discovery of the back-propagation method, many modified and new algorithms have been propo...
In recent years, neural networks, as part of deep learning, became pop- ular because the ability to...