Deep neural networks (DNNs) are powerful tools for compressing and distilling information. Their scale and complexity, often involving billions of inter-dependent parameters, render direct microscopic analysis difficult. Under such circumstances, a common strategy is to identify slow variables that average the erratic behavior of the fast microscopic variables. Here, we identify a similar separation of scales occurring in fully trained finitely over-parameterized deep convolutional neural networks (CNNs) and fully connected networks (FCNs). Specifically, we show that DNN layers couple only through the second moment (kernels) of their activations and pre-activations. Moreover, the latter fluctuates in a nearly Gaussian manner. For infinite w...
Choosing appropriate architectures and regular-ization strategies of deep networks is crucial to goo...
In this work, we establish the relation between optimal control and training deep Convolution Neural...
We analyze numerically the training dynamics of deep neural networks (DNN) by using methods develope...
Two distinct limits for deep learning have been derived as the network width h -> infinity, dependin...
In the recent years, Deep Neural Networks (DNNs) have managed to succeed at tasks that previously ap...
We analyze feature learning in infinite-width neural networks trained with gradient flow through a s...
Deep Learning architectures give brilliant results in a large variety of fields, but a comprehensive...
We show that the output of a (residual) convolutional neural network (CNN) with an appropriate prior...
A remarkable characteristic of overparameterized deep neural networks (DNNs) is that their accuracy ...
The successes of modern deep neural networks (DNNs) are founded on their ability to transform inputs...
We show that the output of a (residual) convolutional neural network (CNN) with an appropriate prior...
This article studies the infinite-width limit of deep feedforward neural networks whose weights are ...
This thesis characterizes the training process of deep neural networks. We are driven by two apparen...
We investigate deep Bayesian neural networks with Gaussian priors on the weights and ReLU-like nonli...
This book develops an effective theory approach to understanding deep neural networks of practical r...
Choosing appropriate architectures and regular-ization strategies of deep networks is crucial to goo...
In this work, we establish the relation between optimal control and training deep Convolution Neural...
We analyze numerically the training dynamics of deep neural networks (DNN) by using methods develope...
Two distinct limits for deep learning have been derived as the network width h -> infinity, dependin...
In the recent years, Deep Neural Networks (DNNs) have managed to succeed at tasks that previously ap...
We analyze feature learning in infinite-width neural networks trained with gradient flow through a s...
Deep Learning architectures give brilliant results in a large variety of fields, but a comprehensive...
We show that the output of a (residual) convolutional neural network (CNN) with an appropriate prior...
A remarkable characteristic of overparameterized deep neural networks (DNNs) is that their accuracy ...
The successes of modern deep neural networks (DNNs) are founded on their ability to transform inputs...
We show that the output of a (residual) convolutional neural network (CNN) with an appropriate prior...
This article studies the infinite-width limit of deep feedforward neural networks whose weights are ...
This thesis characterizes the training process of deep neural networks. We are driven by two apparen...
We investigate deep Bayesian neural networks with Gaussian priors on the weights and ReLU-like nonli...
This book develops an effective theory approach to understanding deep neural networks of practical r...
Choosing appropriate architectures and regular-ization strategies of deep networks is crucial to goo...
In this work, we establish the relation between optimal control and training deep Convolution Neural...
We analyze numerically the training dynamics of deep neural networks (DNN) by using methods develope...