Last layer re-training is sufficient for robustness to spurious correlations

P Kirichenko, P Izmailov, AG Wilson - arXiv preprint arXiv:2204.02937, 2022 - arxiv.org
Neural network classifiers can largely rely on simple spurious features, such as
backgrounds, to make predictions. However, even in these cases, we show that they still …

Towards out-of-distribution generalization: A survey

J Liu, Z Shen, Y He, X Zhang, R Xu, H Yu… - arXiv preprint arXiv …, 2021 - arxiv.org
Traditional machine learning paradigms are based on the assumption that both training and
test data follow the same statistical pattern, which is mathematically referred to as …

Invariance principle meets information bottleneck for out-of-distribution generalization

K Ahuja, E Caballero, D Zhang… - Advances in …, 2021 - proceedings.neurips.cc
The invariance principle from causality is at the heart of notable approaches such as
invariant risk minimization (IRM) that seek to address out-of-distribution (OOD) …

Learning causally invariant representations for out-of-distribution generalization on graphs

Y Chen, Y Zhang, Y Bian, H Yang… - Advances in …, 2022 - proceedings.neurips.cc
Despite recent success in using the invariance principle for out-of-distribution (OOD)
generalization on Euclidean data (eg, images), studies on graph data are still limited …

Pareto invariant risk minimization: Towards mitigating the optimization dilemma in out-of-distribution generalization

Y Chen, K Zhou, Y Bian, B Xie, B Wu, Y Zhang… - arXiv preprint arXiv …, 2022 - arxiv.org
Recently, there has been a growing surge of interest in enabling machine learning systems
to generalize well to Out-of-Distribution (OOD) data. Most efforts are devoted to advancing …

Explicit tradeoffs between adversarial and natural distributional robustness

M Moayeri, K Banihashem… - Advances in Neural …, 2022 - proceedings.neurips.cc
Several existing works study either adversarial or natural distributional robustness of deep
neural networks separately. In practice, however, models need to enjoy both types of …

Vne: An effective method for improving deep representation by manipulating eigenvalue distribution

J Kim, S Kang, D Hwang, J Shin… - Proceedings of the …, 2023 - openaccess.thecvf.com
Since the introduction of deep learning, a wide scope of representation properties, such as
decorrelation, whitening, disentanglement, rank, isotropy, and mutual information, have …

Iterative feature matching: Toward provable domain generalization with logarithmic environments

Y Chen, E Rosenfeld, M Sellke… - Advances in Neural …, 2022 - proceedings.neurips.cc
Abstract Domain generalization aims at performing well on unseen test environments with
data from a limited number of training environments. Despite a proliferation of proposed …

Provable domain generalization via invariant-feature subspace recovery

H Wang, H Si, B Li, H Zhao - International Conference on …, 2022 - proceedings.mlr.press
Abstract Domain generalization asks for models trained over a set of training environments
to perform well in unseen test environments. Recently, a series of algorithms such as …

Log: Active model adaptation for label-efficient ood generalization

JJ Shao, LZ Guo, XW Yang… - Advances in Neural …, 2022 - proceedings.neurips.cc
This work discusses how to achieve worst-case Out-Of-Distribution (OOD) generalization for
a variety of distributions based on a relatively small labeling cost. The problem has broad …