Training deep neural networks consumes increasing computational resource shares in many compute centers. Often, a brute force approach to obtain hyperparameter values is employed. Our goal is (1) to enhance this by enabling second-order optimization methods with fewer hyperparameters for large-scale neural networks and (2) to perform a survey of the performance optimizers for specific tasks to suggest users the best one for their problem. We introduce a novel second-order optimization method that requires the effect of the Hessian on a vector only and avoids the huge cost of explicitly setting up the Hessian for large-scale networks. We compare the proposed second-order method with two state-of-the-art optimizers on five representative ne...
Introduction Training algorithms for Multilayer Perceptrons optimize the set of W weights and biase...
We present an efficient block-diagonal approximation to the Gauss-Newton matrix for feedforward neur...
Spiking neural networks (SNNs) have achieved orders of magnitude improvement in terms of energy cons...
Training of convolutional neural networks is a high dimensional and a non-convex optimization proble...
Neural networks are an important class of highly flexible and powerful models inspired by the struct...
We propose a fast second-order method that can be used as a drop-in replacement for current deep lea...
Which numerical methods are ideal for training a neural network? In this report four different optim...
While first-order methods are popular for solving optimization problems that arise in large-scale de...
This paper presents a state-of-the-art overview on how to architect, design, and optimize Deep Neura...
Thesis (Ph.D.)--University of Washington, 2019The advent of deep neural networks has revolutionized ...
In this dissertation, we are concerned with the advancement of optimization algorithms for training ...
Neural networks, as part of deep learning, have become extremely pop- ular due to their ability to e...
Neural networks stand out from artificial intelligence because they can complete challenging tasks, ...
The simulation of biological neural networks (BNN) is essential to neuroscience. The complexity of t...
Hessian-based analysis/computation is widely used in scientific computing. However, due to the (inco...
Introduction Training algorithms for Multilayer Perceptrons optimize the set of W weights and biase...
We present an efficient block-diagonal approximation to the Gauss-Newton matrix for feedforward neur...
Spiking neural networks (SNNs) have achieved orders of magnitude improvement in terms of energy cons...
Training of convolutional neural networks is a high dimensional and a non-convex optimization proble...
Neural networks are an important class of highly flexible and powerful models inspired by the struct...
We propose a fast second-order method that can be used as a drop-in replacement for current deep lea...
Which numerical methods are ideal for training a neural network? In this report four different optim...
While first-order methods are popular for solving optimization problems that arise in large-scale de...
This paper presents a state-of-the-art overview on how to architect, design, and optimize Deep Neura...
Thesis (Ph.D.)--University of Washington, 2019The advent of deep neural networks has revolutionized ...
In this dissertation, we are concerned with the advancement of optimization algorithms for training ...
Neural networks, as part of deep learning, have become extremely pop- ular due to their ability to e...
Neural networks stand out from artificial intelligence because they can complete challenging tasks, ...
The simulation of biological neural networks (BNN) is essential to neuroscience. The complexity of t...
Hessian-based analysis/computation is widely used in scientific computing. However, due to the (inco...
Introduction Training algorithms for Multilayer Perceptrons optimize the set of W weights and biase...
We present an efficient block-diagonal approximation to the Gauss-Newton matrix for feedforward neur...
Spiking neural networks (SNNs) have achieved orders of magnitude improvement in terms of energy cons...