We introduce a probability distribution, combined with an efficient sampling algorithm, for weights and biases of fully-connected neural networks. In a supervised learning context, no iterative optimization or gradient computations of internal network parameters are needed to obtain a trained network. The sampling is based on the idea of random feature models. However, instead of a data-agnostic distribution, e.g., a normal distribution, we use both the input and the output training data to sample shallow and deep networks. We prove that sampled networks are universal approximators. For Barron functions, we show that the $L^2$-approximation error of sampled shallow networks decreases with the square root of the number of neurons. Our sampli...
Understanding how feature learning affects generalization is among the foremost goals of modern deep...
This paper considers several aspects of random matrix universality in deep neural networks. Motivate...
Artificial neural networks are functions depending on a finite number of parameters typically encode...
Single-index models are a class of functions given by an unknown univariate ``link'' function applie...
Deep neural networks have had tremendous success in a wide range of applications where they achieve ...
We study the computational complexity of (deterministic or randomized) algorithms based on point sam...
Given any deep fully connected neural network, initialized with random Gaussian parameters, we bound...
We show that every $d$-dimensional probability distribution of bounded support can be generated thro...
The success of deep learning has shown impressive empirical breakthroughs, but many theoretical ques...
Deep neural networks train millions of parameters to achieve state-of-the-art performance on a wide ...
International audienceIn the context of supervised learning of a function by a neural network, we cl...
The ability of deep neural networks to generalise well even when they interpolate their training dat...
Understanding the impact of data structure on the computational tractability of learning is a key ch...
Although the neural network (NN) technique plays an important role in machine learning, understandin...
Training deep neural networks with the error backpropagation algorithm is considered implausible fro...
Understanding how feature learning affects generalization is among the foremost goals of modern deep...
This paper considers several aspects of random matrix universality in deep neural networks. Motivate...
Artificial neural networks are functions depending on a finite number of parameters typically encode...
Single-index models are a class of functions given by an unknown univariate ``link'' function applie...
Deep neural networks have had tremendous success in a wide range of applications where they achieve ...
We study the computational complexity of (deterministic or randomized) algorithms based on point sam...
Given any deep fully connected neural network, initialized with random Gaussian parameters, we bound...
We show that every $d$-dimensional probability distribution of bounded support can be generated thro...
The success of deep learning has shown impressive empirical breakthroughs, but many theoretical ques...
Deep neural networks train millions of parameters to achieve state-of-the-art performance on a wide ...
International audienceIn the context of supervised learning of a function by a neural network, we cl...
The ability of deep neural networks to generalise well even when they interpolate their training dat...
Understanding the impact of data structure on the computational tractability of learning is a key ch...
Although the neural network (NN) technique plays an important role in machine learning, understandin...
Training deep neural networks with the error backpropagation algorithm is considered implausible fro...
Understanding how feature learning affects generalization is among the foremost goals of modern deep...
This paper considers several aspects of random matrix universality in deep neural networks. Motivate...
Artificial neural networks are functions depending on a finite number of parameters typically encode...