In recent years, deep learning has been connected with optimal control as a way to define a notion of a continuous underlying learning problem. In this view, neural networks can be interpreted as a discretization of a parametric Ordinary Differential Equation which, in the limit, defines a continuous-depth neural network. The learning task then consists in finding the best ODE parameters for the problem under consideration, and their number increases with the accuracy of the time discretization. Although important steps have been taken to realize the advantages of such continuous formulations, most current learning techniques fix a discretization (i.e. the number of layers is fixed). In this work, we propose an iterative adaptive algorithm ...
Artificial Neural Networks are increasingly being used in complex real- world applications because m...
We consider recent work of Haber and Ruthotto 2017 and Chang et al. 2018, where deep learning neural...
The general features of the optimization problem for the case of overparametrized nonlinear networks...
We propose an optimal architecture for deep neural networks of given size. The optimal architecture ...
We briefly review recent work where deep learning neural networks have been interpreted as discretis...
We consider recent work of Haber and Ruthotto 2017 and Chang et al. 2018, where deep learning neural...
Continuous-depth neural networks can be viewed as deep limits of discrete neural networks whose dyna...
We perform a comprehensive numerical study of the effect of approximation-theoretical results for n...
Neural networks have been very successful in many applications; we often, however, lack a theoretica...
The past decade has seen increasing interest in applying Deep Learning (DL) to Computational Science...
In this paper, we study a regularised relaxed optimal control problem and, in particular, we are con...
A main puzzle of deep networks revolves around the absence of overfitting despite overparametrizatio...
The paper contains approximation guarantees for neural networks that are trained with gradient flow,...
This work features an original result linking approximation and optimization theory for deep learnin...
Conventional wisdom in deep learning states that increasing depth improves expressiveness but compli...
Artificial Neural Networks are increasingly being used in complex real- world applications because m...
We consider recent work of Haber and Ruthotto 2017 and Chang et al. 2018, where deep learning neural...
The general features of the optimization problem for the case of overparametrized nonlinear networks...
We propose an optimal architecture for deep neural networks of given size. The optimal architecture ...
We briefly review recent work where deep learning neural networks have been interpreted as discretis...
We consider recent work of Haber and Ruthotto 2017 and Chang et al. 2018, where deep learning neural...
Continuous-depth neural networks can be viewed as deep limits of discrete neural networks whose dyna...
We perform a comprehensive numerical study of the effect of approximation-theoretical results for n...
Neural networks have been very successful in many applications; we often, however, lack a theoretica...
The past decade has seen increasing interest in applying Deep Learning (DL) to Computational Science...
In this paper, we study a regularised relaxed optimal control problem and, in particular, we are con...
A main puzzle of deep networks revolves around the absence of overfitting despite overparametrizatio...
The paper contains approximation guarantees for neural networks that are trained with gradient flow,...
This work features an original result linking approximation and optimization theory for deep learnin...
Conventional wisdom in deep learning states that increasing depth improves expressiveness but compli...
Artificial Neural Networks are increasingly being used in complex real- world applications because m...
We consider recent work of Haber and Ruthotto 2017 and Chang et al. 2018, where deep learning neural...
The general features of the optimization problem for the case of overparametrized nonlinear networks...