We present DANTE, a novel method for training neural networks using the alternating minimization principle. DANTE provides an alternate perspective to traditional gradient-based backpropagation techniques commonly used to train deep networks. It utilizes an adaptation of quasi-convexity to cast training a neural network as a bi-quasi-convex optimization problem. We show that for neural network configurations with both differentiable (e.g. sigmoid) and non-differentiable (e.g. ReLU) activation functions, we can perform the alternations effectively in this formulation. DANTE can also be extended to networks with multiple hidden layers. In experiments on standard datasets, neural networks trained using the proposed method were found to be prom...
Neural networks have been intensively studied as machine learning models and widely applied in vario...
We propose BlockProp, a neural network training algorithm. Unlike backpropagation, it does not rely ...
While deep neural networks (DNNs) have shown to be successful in several domains like computer visio...
This paper proposes a set of new error criteria and a learning approach, called Adaptive Normalized ...
In the recent decade, deep neural networks have solved ever more complex tasks across many fronts in...
We present a novel regularization approach to train neural networks that enjoys better generalizatio...
In this dissertation, we are concerned with the advancement of optimization algorithms for training ...
In this paper, the authors propose a new training algorithm which does not only rely upon the traini...
Classic algorithms and machine learning systems like neural networks are both abundant in everyday l...
We propose a new algorithm for training deep neural networks (DNNs) with binary weights. In particul...
Adversarial training has been shown to regularize deep neural networks in addition to increasing the...
Deep neural network architectures have recently produced excellent results in a variety of areas in ...
In modern supervised learning, many deep neural networks are able to interpolate the data: the empir...
Deep neural network architectures have recently produced excellent results in a variety of areas in ...
Abstract—A constrained-backpropagation training technique is presented to suppress interference and ...
Neural networks have been intensively studied as machine learning models and widely applied in vario...
We propose BlockProp, a neural network training algorithm. Unlike backpropagation, it does not rely ...
While deep neural networks (DNNs) have shown to be successful in several domains like computer visio...
This paper proposes a set of new error criteria and a learning approach, called Adaptive Normalized ...
In the recent decade, deep neural networks have solved ever more complex tasks across many fronts in...
We present a novel regularization approach to train neural networks that enjoys better generalizatio...
In this dissertation, we are concerned with the advancement of optimization algorithms for training ...
In this paper, the authors propose a new training algorithm which does not only rely upon the traini...
Classic algorithms and machine learning systems like neural networks are both abundant in everyday l...
We propose a new algorithm for training deep neural networks (DNNs) with binary weights. In particul...
Adversarial training has been shown to regularize deep neural networks in addition to increasing the...
Deep neural network architectures have recently produced excellent results in a variety of areas in ...
In modern supervised learning, many deep neural networks are able to interpolate the data: the empir...
Deep neural network architectures have recently produced excellent results in a variety of areas in ...
Abstract—A constrained-backpropagation training technique is presented to suppress interference and ...
Neural networks have been intensively studied as machine learning models and widely applied in vario...
We propose BlockProp, a neural network training algorithm. Unlike backpropagation, it does not rely ...
While deep neural networks (DNNs) have shown to be successful in several domains like computer visio...