Sketching and stochastic gradient methods are arguably the most common techniques to derive efficient large scale learning algorithms. In this paper, we investigate their application in the context of nonparametric statistical learning. More precisely, we study the estimator defined by stochastic gradient with mini batches and random features. The latter can be seen as form of nonlinear sketching and used to define approximate kernel methods. The considered estimator is not explicitly penalized/constrained and regularization is implicit. Indeed, our study highlights how different parameters, such as number of features, iterations, step-size and mini-batch size control the learning properties of the solutions. We do this by deriving optimal ...
In recent studies, the generalization properties for distributed learning and random features assume...
Thesis (Ph.D.)--University of Washington, 2019Tremendous advances in large scale machine learning an...
We study generalization properties of distributed algorithms in the setting of nonparametric regress...
SpotlightInternational audienceSketching and stochastic gradient methods are arguably the most commo...
We analyze the learning properties of the stochastic gradient method when multiple passes over the d...
We analyze the learning properties of the stochastic gradient method when multiple passes over the d...
We study generalization properties of random features (RF) regression in high dimensions optimized b...
We study the generalization properties of ridge regression with random features in the statistical l...
Stochastic gradient descent (SGD) and its variants have established themselves as the go-to algorith...
AbstractWe propose a stochastic gradient descent algorithm for learning the gradient of a regression...
Kernel methods and neural networks are two important schemes in the supervised learning field. The t...
This dissertation focuses on stochastic gradient learning for problems involving large data sets or ...
We consider the random-design least-squares regression problem within the reproducing kernel Hilbert...
Statistical inference, such as hypothesis testing and calculating a confidence interval, is an impor...
Stochastic descent methods (of the gradient and mirror varieties) have become increasingly popular i...
In recent studies, the generalization properties for distributed learning and random features assume...
Thesis (Ph.D.)--University of Washington, 2019Tremendous advances in large scale machine learning an...
We study generalization properties of distributed algorithms in the setting of nonparametric regress...
SpotlightInternational audienceSketching and stochastic gradient methods are arguably the most commo...
We analyze the learning properties of the stochastic gradient method when multiple passes over the d...
We analyze the learning properties of the stochastic gradient method when multiple passes over the d...
We study generalization properties of random features (RF) regression in high dimensions optimized b...
We study the generalization properties of ridge regression with random features in the statistical l...
Stochastic gradient descent (SGD) and its variants have established themselves as the go-to algorith...
AbstractWe propose a stochastic gradient descent algorithm for learning the gradient of a regression...
Kernel methods and neural networks are two important schemes in the supervised learning field. The t...
This dissertation focuses on stochastic gradient learning for problems involving large data sets or ...
We consider the random-design least-squares regression problem within the reproducing kernel Hilbert...
Statistical inference, such as hypothesis testing and calculating a confidence interval, is an impor...
Stochastic descent methods (of the gradient and mirror varieties) have become increasingly popular i...
In recent studies, the generalization properties for distributed learning and random features assume...
Thesis (Ph.D.)--University of Washington, 2019Tremendous advances in large scale machine learning an...
We study generalization properties of distributed algorithms in the setting of nonparametric regress...