Trak: Attributing model behavior at scale

SM Park, K Georgiev, A Ilyas, G Leclerc… - arXiv preprint arXiv …, 2023 - arxiv.org
The goal of data attribution is to trace model predictions back to training data. Despite a long
line of work towards this goal, existing approaches to data attribution tend to force users to …

Describing differences in image sets with natural language

L Dunlap, Y Zhang, X Wang, R Zhong… - Proceedings of the …, 2024 - openaccess.thecvf.com
How do two sets of images differ? Discerning set-level differences is crucial for
understanding model behaviors and analyzing datasets yet manually sifting through …

Facts: First amplify correlations and then slice to discover bias

S Yenamandra, P Ramesh… - Proceedings of the …, 2023 - openaccess.thecvf.com
Computer vision datasets frequently contain spurious correlations between task-relevant
labels and (easy to learn) latent task-irrelevant attributes (eg context). Models trained on …

Similarity of neural network models: A survey of functional and representational measures

M Klabunde, T Schumacher, M Strohmaier… - arXiv preprint arXiv …, 2023 - arxiv.org
Measuring similarity of neural networks to understand and improve their behavior has
become an issue of great importance and research interest. In this survey, we provide a …

Understanding the detrimental class-level effects of data augmentation

P Kirichenko, M Ibrahim, R Balestriero… - Advances in …, 2024 - proceedings.neurips.cc
Data augmentation (DA) encodes invariance and provides implicit regularization critical to a
model's performance in image classification tasks. However, while DA improves average …

Bias-to-text: Debiasing unknown visual biases through language interpretation

Y Kim, S Mo, M Kim, K Lee, J Lee, J Shin - arXiv preprint arXiv:2301.11104, 2023 - arxiv.org
Biases in models pose a critical issue when deploying machine learning systems, but
diagnosing them in an explainable manner can be challenging. To address this, we …

DiG-IN: Diffusion Guidance for Investigating Networks-Uncovering Classifier Differences Neuron Visualisations and Visual Counterfactual Explanations

M Augustin, Y Neuhaus, M Hein - Proceedings of the IEEE …, 2024 - openaccess.thecvf.com
While deep learning has led to huge progress in complex image classification tasks like
ImageNet unexpected failure modes eg via spurious features call into question how reliably …

Decomposing and editing predictions by modeling model computation

H Shah, A Ilyas, A Madry - arXiv preprint arXiv:2404.11534, 2024 - arxiv.org
How does the internal computation of a machine learning model transform inputs into
predictions? In this paper, we introduce a task called component modeling that aims to …

The journey, not the destination: How data guides diffusion models

K Georgiev, J Vendrow, H Salman, SM Park… - arXiv preprint arXiv …, 2023 - arxiv.org
Diffusion models trained on large datasets can synthesize photo-realistic images of
remarkable quality and diversity. However, attributing these images back to the training data …

Data Debiasing with Datamodels (D3M): Improving Subgroup Robustness via Data Selection

S Jain, K Hamidieh, K Georgiev, A Ilyas… - arXiv preprint arXiv …, 2024 - arxiv.org
Machine learning models can fail on subgroups that are underrepresented during training.
While techniques such as dataset balancing can improve performance on underperforming …