There has been a surge of interest in developing robust estimators for models with heavy-tailed data in statistics and machine learning. This paper proposes a log-truncated M-estimator for a large family of statistical regressions and establishes its excess risk bound under the condition that the data have $(1+\varepsilon)$-th moment with $\varepsilon \in (0,1]$. With an additional assumption on the associated risk function, we obtain an $\ell_2$-error bound for the estimation. Our theorems are applied to establish robust M-estimators for concrete regressions. Besides convex regressions such as quantile regression and generalized linear models, many non-convex regressions can also be fit into our theorems, we focus on robust deep neural net...
International audienceWe propose new parametrizations for neural networks in order to estimate extre...
Many datasets are collected automatically, and are thus easily contaminated by outliers. In order to...
78We consider the problem of predicting as well as the best linear combination of d given functions ...
We consider the problem of constrained M-estimation when both explanatory and response variables hav...
29 pagesInternational audienceWe consider the problem of robustly predicting as well as the best lin...
42 pagesWe introduce a procedure for predictive conditional density estimation under logarithmic los...
We establish risk bounds for Regularized Empirical Risk Minimizers (RERM) when the loss is Lipschitz...
This paper provides some extended results on estimating parameter matrix of some regression models w...
The limiting distribution of M-estimators of the regression parameter in linear models is derived un...
The design of statistical estimators robust to outliers has been a mainstay of statistical research ...
This work provides test error bounds for iterative fixed point methods on linear predictors -- speci...
Data sets where the number of variables p is comparable to or larger than the number of observations...
A stylized feature of high-dimensional data is that many variables have heavy tails, and robust stat...
48 pages, 6 figuresInternational audienceWe introduce new estimators for robust machine learning bas...
A fundamental problem in modern high-dimensional data analysis involves efficiently inferring a set ...
International audienceWe propose new parametrizations for neural networks in order to estimate extre...
Many datasets are collected automatically, and are thus easily contaminated by outliers. In order to...
78We consider the problem of predicting as well as the best linear combination of d given functions ...
We consider the problem of constrained M-estimation when both explanatory and response variables hav...
29 pagesInternational audienceWe consider the problem of robustly predicting as well as the best lin...
42 pagesWe introduce a procedure for predictive conditional density estimation under logarithmic los...
We establish risk bounds for Regularized Empirical Risk Minimizers (RERM) when the loss is Lipschitz...
This paper provides some extended results on estimating parameter matrix of some regression models w...
The limiting distribution of M-estimators of the regression parameter in linear models is derived un...
The design of statistical estimators robust to outliers has been a mainstay of statistical research ...
This work provides test error bounds for iterative fixed point methods on linear predictors -- speci...
Data sets where the number of variables p is comparable to or larger than the number of observations...
A stylized feature of high-dimensional data is that many variables have heavy tails, and robust stat...
48 pages, 6 figuresInternational audienceWe introduce new estimators for robust machine learning bas...
A fundamental problem in modern high-dimensional data analysis involves efficiently inferring a set ...
International audienceWe propose new parametrizations for neural networks in order to estimate extre...
Many datasets are collected automatically, and are thus easily contaminated by outliers. In order to...
78We consider the problem of predicting as well as the best linear combination of d given functions ...