Learning invariant (causal) features for out-of-distribution (OOD) generalization have attracted extensive attention recently, and among the proposals, invariant risk minimization (IRM) is a notable solution. In spite of its theoretical promise for linear regression, the challenges of using IRM in linear classification problems remain. By introducing the information bottleneck (IB) principle into the learning of IRM, the IB-IRM approach has demonstrated its power to solve these challenges. In this paper, we further improve IB-IRM from two aspects. First, we show that the key assumption of support overlap of invariant features used in IB-IRM guarantees OOD generalization, and it is still possible to achieve the optimal solution without this ...
Invariance-principle-based methods such as Invariant Risk Minimization (IRM), have recently emerged ...
Machine learning methods suffer from test-time performance degeneration when faced with out-of-distr...
Machine learning methods suffer from test-time performance degeneration when faced with out-of-distr...
Learning invariant (causal) features for out-of-distribution (OOD) generalization have attracted ext...
The invariance principle from causality is at the heart of notable approaches such as invariant risk...
Learning algorithms can perform poorly in unseen environments when they learnspurious correlations. ...
Despite impressive success in many tasks, deep learning models are shown to rely on spurious feature...
Due to spurious correlations, machine learning systems often fail to generalize to environments whos...
Learning models that are robust to distribution shifts is a key concern in the context of their real...
Due to the poor generalization performance of traditional empirical risk minimization (ERM) in the c...
International audienceA grand challenge in representation learning is the development of computation...
The Information Bottleneck (IB) method provides an insightful and principled approach for balancing ...
Out-of-Domain (OOD) generalization is a challenging problem in machine learning about learning a mod...
Generalizing models for new unknown datasets is a common problem in machine learning. Algorithms tha...
Domain generalization asks for models trained over a set of training environments to generalize well...
Invariance-principle-based methods such as Invariant Risk Minimization (IRM), have recently emerged ...
Machine learning methods suffer from test-time performance degeneration when faced with out-of-distr...
Machine learning methods suffer from test-time performance degeneration when faced with out-of-distr...
Learning invariant (causal) features for out-of-distribution (OOD) generalization have attracted ext...
The invariance principle from causality is at the heart of notable approaches such as invariant risk...
Learning algorithms can perform poorly in unseen environments when they learnspurious correlations. ...
Despite impressive success in many tasks, deep learning models are shown to rely on spurious feature...
Due to spurious correlations, machine learning systems often fail to generalize to environments whos...
Learning models that are robust to distribution shifts is a key concern in the context of their real...
Due to the poor generalization performance of traditional empirical risk minimization (ERM) in the c...
International audienceA grand challenge in representation learning is the development of computation...
The Information Bottleneck (IB) method provides an insightful and principled approach for balancing ...
Out-of-Domain (OOD) generalization is a challenging problem in machine learning about learning a mod...
Generalizing models for new unknown datasets is a common problem in machine learning. Algorithms tha...
Domain generalization asks for models trained over a set of training environments to generalize well...
Invariance-principle-based methods such as Invariant Risk Minimization (IRM), have recently emerged ...
Machine learning methods suffer from test-time performance degeneration when faced with out-of-distr...
Machine learning methods suffer from test-time performance degeneration when faced with out-of-distr...