Last layer re-training is sufficient for robustness to spurious correlations
Neural network classifiers can largely rely on simple spurious features, such as
backgrounds, to make predictions. However, even in these cases, we show that they still …
backgrounds, to make predictions. However, even in these cases, we show that they still …
Towards out-of-distribution generalization: A survey
Traditional machine learning paradigms are based on the assumption that both training and
test data follow the same statistical pattern, which is mathematically referred to as …
test data follow the same statistical pattern, which is mathematically referred to as …
Invariance principle meets information bottleneck for out-of-distribution generalization
The invariance principle from causality is at the heart of notable approaches such as
invariant risk minimization (IRM) that seek to address out-of-distribution (OOD) …
invariant risk minimization (IRM) that seek to address out-of-distribution (OOD) …
Learning causally invariant representations for out-of-distribution generalization on graphs
Despite recent success in using the invariance principle for out-of-distribution (OOD)
generalization on Euclidean data (eg, images), studies on graph data are still limited …
generalization on Euclidean data (eg, images), studies on graph data are still limited …
Pareto invariant risk minimization: Towards mitigating the optimization dilemma in out-of-distribution generalization
Recently, there has been a growing surge of interest in enabling machine learning systems
to generalize well to Out-of-Distribution (OOD) data. Most efforts are devoted to advancing …
to generalize well to Out-of-Distribution (OOD) data. Most efforts are devoted to advancing …
Explicit tradeoffs between adversarial and natural distributional robustness
M Moayeri, K Banihashem… - Advances in Neural …, 2022 - proceedings.neurips.cc
Several existing works study either adversarial or natural distributional robustness of deep
neural networks separately. In practice, however, models need to enjoy both types of …
neural networks separately. In practice, however, models need to enjoy both types of …
Vne: An effective method for improving deep representation by manipulating eigenvalue distribution
Since the introduction of deep learning, a wide scope of representation properties, such as
decorrelation, whitening, disentanglement, rank, isotropy, and mutual information, have …
decorrelation, whitening, disentanglement, rank, isotropy, and mutual information, have …
Iterative feature matching: Toward provable domain generalization with logarithmic environments
Abstract Domain generalization aims at performing well on unseen test environments with
data from a limited number of training environments. Despite a proliferation of proposed …
data from a limited number of training environments. Despite a proliferation of proposed …
Provable domain generalization via invariant-feature subspace recovery
Abstract Domain generalization asks for models trained over a set of training environments
to perform well in unseen test environments. Recently, a series of algorithms such as …
to perform well in unseen test environments. Recently, a series of algorithms such as …
Log: Active model adaptation for label-efficient ood generalization
This work discusses how to achieve worst-case Out-Of-Distribution (OOD) generalization for
a variety of distributions based on a relatively small labeling cost. The problem has broad …
a variety of distributions based on a relatively small labeling cost. The problem has broad …