From the sampling of data to the initialisation of parameters, randomness is ubiquitous in modern Machine Learning practice. Understanding the statistical fluctuations engendered by the different sources of randomness in prediction is therefore key to understanding robust generalisation. In this manuscript we develop a quantitative and rigorous theory for the study of fluctuations in an ensemble of generalised linear models trained on different, but correlated, features in high-dimensions. In particular, we provide a complete description of the asymptotic joint distribution of the empirical risk minimiser for generic convex loss and regularisation in the high-dimensional limit. Our result encompasses a rich set of classification and regress...
Benign overfitting, the phenomenon where interpolating models generalize well in the presence of noi...
With the deluge of digitized information in the Big Data era, massive datasets are becoming increasi...
We consider the problem of stochastic convex optimization with exp-concave losses using Empirical Ri...
The recent empirical success of machine learning in all fields involving data analysis has prompted ...
In these lecture notes we present different methods and concepts developed in statistical physics to...
We study generalization properties of random features (RF) regression in high dimensions optimized b...
Interpolators -- estimators that achieve zero training error -- have attracted growing attention in ...
see also: 2020 Generalisation error in learning with random features and the hidden manifold model, ...
We prove a non-asymptotic distribution-independent lower bound for the expected mean squared general...
We study generalised linear regression and classification for a synthetically generated dataset enco...
We prove a universality theorem for learning with random features. Our result shows that, in terms o...
We examine the necessity of interpolation in overparameterized models, that is, when achieving optim...
In this note, we provide an elementary analysis of the prediction error of ridge regression with ran...
Statistical measures for group fairness in machine learning reflect the gap in performance of algori...
This paper establishes a precise high-dimensional asymptotic theory for boosting on separable data, ...
Benign overfitting, the phenomenon where interpolating models generalize well in the presence of noi...
With the deluge of digitized information in the Big Data era, massive datasets are becoming increasi...
We consider the problem of stochastic convex optimization with exp-concave losses using Empirical Ri...
The recent empirical success of machine learning in all fields involving data analysis has prompted ...
In these lecture notes we present different methods and concepts developed in statistical physics to...
We study generalization properties of random features (RF) regression in high dimensions optimized b...
Interpolators -- estimators that achieve zero training error -- have attracted growing attention in ...
see also: 2020 Generalisation error in learning with random features and the hidden manifold model, ...
We prove a non-asymptotic distribution-independent lower bound for the expected mean squared general...
We study generalised linear regression and classification for a synthetically generated dataset enco...
We prove a universality theorem for learning with random features. Our result shows that, in terms o...
We examine the necessity of interpolation in overparameterized models, that is, when achieving optim...
In this note, we provide an elementary analysis of the prediction error of ridge regression with ran...
Statistical measures for group fairness in machine learning reflect the gap in performance of algori...
This paper establishes a precise high-dimensional asymptotic theory for boosting on separable data, ...
Benign overfitting, the phenomenon where interpolating models generalize well in the presence of noi...
With the deluge of digitized information in the Big Data era, massive datasets are becoming increasi...
We consider the problem of stochastic convex optimization with exp-concave losses using Empirical Ri...