Learning models that are robust to distribution shifts is a key concern in the context of their real-life applicability. Invariant Risk Minimization (IRM) is a popular framework that aims to learn robust models from multiple environments. The success of IRM requires an important assumption: the underlying causal mechanisms/features remain invariant across environments. When not satisfied, we show that IRM can over-constrain the predictor and to remedy this, we propose a relaxation via partial invariance. In this work, we theoretically highlight the sub-optimality of IRM and then demonstrate how learning from a partition of training domains can help improve invariant models. Several experiments, conducted both in linear settings as well as w...
Modern pretrained language models are critical components of NLP pipelines. Yet, they suffer from sp...
Assumptions about invariances or symmetries in data can significantly increase the predictive power ...
Learning invariant representations is a critical first step in a number of machine learning tasks. A...
Learning algorithms can perform poorly in unseen environments when they learnspurious correlations. ...
Domain generalization asks for models trained over a set of training environments to perform well in...
Despite impressive success in many tasks, deep learning models are shown to rely on spurious feature...
Due to spurious correlations, machine learning systems often fail to generalize to environments whos...
When the test distribution differs from the training distribution, machine learning models can perfo...
Learning invariant (causal) features for out-of-distribution (OOD) generalization have attracted ext...
Designing learning systems which are invariant to certain data transformations is critical in machin...
In this thesis, Invariance in Deep Representations, we propose novel solutions to the problem of lea...
One of the main goal of Artificial Intelligence is to develop models capable of providing valuable p...
Generalizing models for new unknown datasets is a common problem in machine learning. Algorithms tha...
One method for obtaining generalizable solutions to machine learning tasks when presented with diver...
In supervised machine learning, it is common practice to choose a loss function for learning predict...
Modern pretrained language models are critical components of NLP pipelines. Yet, they suffer from sp...
Assumptions about invariances or symmetries in data can significantly increase the predictive power ...
Learning invariant representations is a critical first step in a number of machine learning tasks. A...
Learning algorithms can perform poorly in unseen environments when they learnspurious correlations. ...
Domain generalization asks for models trained over a set of training environments to perform well in...
Despite impressive success in many tasks, deep learning models are shown to rely on spurious feature...
Due to spurious correlations, machine learning systems often fail to generalize to environments whos...
When the test distribution differs from the training distribution, machine learning models can perfo...
Learning invariant (causal) features for out-of-distribution (OOD) generalization have attracted ext...
Designing learning systems which are invariant to certain data transformations is critical in machin...
In this thesis, Invariance in Deep Representations, we propose novel solutions to the problem of lea...
One of the main goal of Artificial Intelligence is to develop models capable of providing valuable p...
Generalizing models for new unknown datasets is a common problem in machine learning. Algorithms tha...
One method for obtaining generalizable solutions to machine learning tasks when presented with diver...
In supervised machine learning, it is common practice to choose a loss function for learning predict...
Modern pretrained language models are critical components of NLP pipelines. Yet, they suffer from sp...
Assumptions about invariances or symmetries in data can significantly increase the predictive power ...
Learning invariant representations is a critical first step in a number of machine learning tasks. A...