In learning theory, the training and test sets are assumed to be drawn from the same probability distribution. This assumption is also followed in practical situations, where matching the training and test distributions is considered desirable. Contrary to conventional wisdom, we show that mismatched training and test distributions in supervised learning can in fact outperform matched distributions in terms of the bottom line, the out-of-sample performance, independent of the target function in question. This surprising result has theoretical and algorithmic ramifications that we discuss
The estimated accuracy of a classifier is a random quantity with variability. A common practice in s...
Breakthroughs in machine learning are rapidly changing science and society, yet our fundamental unde...
Since out-of-distribution generalization is a generally ill-posed problem, various proxy targets (e....
In learning theory, the training and test sets are assumed to be drawn from the same probability dis...
Supervised machine learning techniques developed in the Probably Approximately Correct, Maximum A Po...
The Domain Adaptation problem in machine learning occurs when the distribution generating the test d...
In the first part of the thesis we explore three fundamental questions that arise naturally when we ...
In machine learning, we traditionally evaluate the performance of a single model, averaged over a co...
In the theory of supervised learning, the identical assumption, i.e. the training and test samples a...
In Machine Learning (ML), the x-covariate and its y-label may have di erent joint probability distr...
We study the interaction between input distributions, learning algo-rithms, and finite sample sizes ...
Machine learning algorithms typically assume that training and test examples are drawn from the same...
Research has not yet reached a consensus on why humans match probabilities instead of maximise in a ...
We study the interaction between input distributions, learning algorithms and finite sample sizes in...
Discriminative learning methods for classification perform well when training and test data are draw...
The estimated accuracy of a classifier is a random quantity with variability. A common practice in s...
Breakthroughs in machine learning are rapidly changing science and society, yet our fundamental unde...
Since out-of-distribution generalization is a generally ill-posed problem, various proxy targets (e....
In learning theory, the training and test sets are assumed to be drawn from the same probability dis...
Supervised machine learning techniques developed in the Probably Approximately Correct, Maximum A Po...
The Domain Adaptation problem in machine learning occurs when the distribution generating the test d...
In the first part of the thesis we explore three fundamental questions that arise naturally when we ...
In machine learning, we traditionally evaluate the performance of a single model, averaged over a co...
In the theory of supervised learning, the identical assumption, i.e. the training and test samples a...
In Machine Learning (ML), the x-covariate and its y-label may have di erent joint probability distr...
We study the interaction between input distributions, learning algo-rithms, and finite sample sizes ...
Machine learning algorithms typically assume that training and test examples are drawn from the same...
Research has not yet reached a consensus on why humans match probabilities instead of maximise in a ...
We study the interaction between input distributions, learning algorithms and finite sample sizes in...
Discriminative learning methods for classification perform well when training and test data are draw...
The estimated accuracy of a classifier is a random quantity with variability. A common practice in s...
Breakthroughs in machine learning are rapidly changing science and society, yet our fundamental unde...
Since out-of-distribution generalization is a generally ill-posed problem, various proxy targets (e....