Due to spurious correlations, machine learning systems often fail to generalize to environments whose distributions differ from the ones used at training time. Prior work addressing this, either explicitly or implicitly, attempted to find a data representation that has an invariant relationship with the target. This is done by leveraging a diverse set of training environments to reduce the effect of spurious features and build an invariant predictor. However, these methods have generalization guarantees only when both data representation and classifiers come from a linear model class. We propose invariant Causal Representation Learning (iCaRL), an approach that enables out-of-distribution (OOD) generalization in the nonlinear setting (i.e.,...
Learning the causal structure behind data is invaluable for improving generalization and obtaining h...
Learning high-level causal representations together with a causal model from unstructured low-level ...
Despite impressive success in many tasks, deep learning models are shown to rely on spurious feature...
An important problem in many domains is to predict how a system will respond to interventions. This ...
Learning algorithms can perform poorly in unseen environments when they learnspurious correlations. ...
The two fields of machine learning and graphical causality arose and are developed separately. Howev...
We propose to learn an invariant causal predictor that is robust to distributional shifts, in the su...
Generalization across environments is critical to the successful application of reinforcement learni...
In this thesis, Invariance in Deep Representations, we propose novel solutions to the problem of lea...
Causal representation learning (CRL) aims at identifying high-level causal variables from low-level ...
The popularity of machine learning in both academia and industry has experienced unparalleled growth...
Modern pretrained language models are critical components of NLP pipelines. Yet, they suffer from sp...
We explore relationships between machine learning (ML) and causal inference. We focus on improvement...
Learning models that offer robust out-of-distribution generalization and fast adaptation is a key ch...
Learning models that are robust to distribution shifts is a key concern in the context of their real...
Learning the causal structure behind data is invaluable for improving generalization and obtaining h...
Learning high-level causal representations together with a causal model from unstructured low-level ...
Despite impressive success in many tasks, deep learning models are shown to rely on spurious feature...
An important problem in many domains is to predict how a system will respond to interventions. This ...
Learning algorithms can perform poorly in unseen environments when they learnspurious correlations. ...
The two fields of machine learning and graphical causality arose and are developed separately. Howev...
We propose to learn an invariant causal predictor that is robust to distributional shifts, in the su...
Generalization across environments is critical to the successful application of reinforcement learni...
In this thesis, Invariance in Deep Representations, we propose novel solutions to the problem of lea...
Causal representation learning (CRL) aims at identifying high-level causal variables from low-level ...
The popularity of machine learning in both academia and industry has experienced unparalleled growth...
Modern pretrained language models are critical components of NLP pipelines. Yet, they suffer from sp...
We explore relationships between machine learning (ML) and causal inference. We focus on improvement...
Learning models that offer robust out-of-distribution generalization and fast adaptation is a key ch...
Learning models that are robust to distribution shifts is a key concern in the context of their real...
Learning the causal structure behind data is invaluable for improving generalization and obtaining h...
Learning high-level causal representations together with a causal model from unstructured low-level ...
Despite impressive success in many tasks, deep learning models are shown to rely on spurious feature...