This manuscript considers the problem of learning a random Gaussian network function using a fully connected network with frozen intermediate layers and trainable readout layer. This problem can be seen as a natural generalization of the widely studied random features model to deeper architectures. First, we prove Gaussian universality of the test error in a ridge regression setting where the learner and target networks share the same intermediate layers, and provide a sharp asymptotic formula for it. Establishing this result requires proving a deterministic equivalent for traces of the deep random features sample covariance matrices which can be of independent interest. Second, we conjecture the asymptotic Gaussian universality of the test...
We investigate the local spectral statistics of the loss surface Hessians of artificial neural netwo...
Regression models usually tend to recover a noisy signal in the form of a combination of regressors,...
We prove a non-asymptotic distribution-independent lower bound for the expected mean squared general...
We prove a universality theorem for learning with random features. Our result shows that, in terms o...
Understanding the impact of data structure on the computational tractability of learning is a key ch...
Kernel methods and neural networks are two important schemes in the supervised learning field. The t...
This paper considers several aspects of random matrix universality in deep neural networks. Motivate...
Choosing appropriate architectures and regularization strategies of deep networks is crucial to good...
Understanding how feature learning affects generalization is among the foremost goals of modern deep...
We compute precise asymptotic expressions for the learning curves of least squares random feature (R...
We propose semi-random features for nonlinear function approximation. The flexibility of semi-random...
We consider the problem of learning a target function corresponding to a deep, extensive-width, non-...
This paper considers several aspects of random matrix universality in deep neural networks. Motivate...
In this work, we provide a characterization of the feature-learning process in two-layer ReLU networ...
The reliability of deep learning algorithms is fundamentally challenged by the existence of adversar...
We investigate the local spectral statistics of the loss surface Hessians of artificial neural netwo...
Regression models usually tend to recover a noisy signal in the form of a combination of regressors,...
We prove a non-asymptotic distribution-independent lower bound for the expected mean squared general...
We prove a universality theorem for learning with random features. Our result shows that, in terms o...
Understanding the impact of data structure on the computational tractability of learning is a key ch...
Kernel methods and neural networks are two important schemes in the supervised learning field. The t...
This paper considers several aspects of random matrix universality in deep neural networks. Motivate...
Choosing appropriate architectures and regularization strategies of deep networks is crucial to good...
Understanding how feature learning affects generalization is among the foremost goals of modern deep...
We compute precise asymptotic expressions for the learning curves of least squares random feature (R...
We propose semi-random features for nonlinear function approximation. The flexibility of semi-random...
We consider the problem of learning a target function corresponding to a deep, extensive-width, non-...
This paper considers several aspects of random matrix universality in deep neural networks. Motivate...
In this work, we provide a characterization of the feature-learning process in two-layer ReLU networ...
The reliability of deep learning algorithms is fundamentally challenged by the existence of adversar...
We investigate the local spectral statistics of the loss surface Hessians of artificial neural netwo...
Regression models usually tend to recover a noisy signal in the form of a combination of regressors,...
We prove a non-asymptotic distribution-independent lower bound for the expected mean squared general...