LAS-AT: adversarial training with learnable attack strategy

X Jia, Y Zhang, B Wu, K Ma… - Proceedings of the …, 2022 - openaccess.thecvf.com
Adversarial training (AT) is always formulated as a minimax problem, of which the
performance depends on the inner optimization that involves the generation of adversarial …

Learning to augment distributions for out-of-distribution detection

Q Wang, Z Fang, Y Zhang, F Liu… - Advances in neural …, 2023 - proceedings.neurips.cc
Open-world classification systems should discern out-of-distribution (OOD) data whose
labels deviate from those of in-distribution (ID) cases, motivating recent studies in OOD …

Robust generalization against photon-limited corruptions via worst-case sharpness minimization

Z Huang, M Zhu, X Xia, L Shen, J Yu… - Proceedings of the …, 2023 - openaccess.thecvf.com
Robust generalization aims to tackle the most challenging data distributions which are rare
in the training set and contain severe noises, ie, photon-limited corruptions. Common …

Watermarking for out-of-distribution detection

Q Wang, F Liu, Y Zhang, J Zhang… - Advances in …, 2022 - proceedings.neurips.cc
Abstract Out-of-distribution (OOD) detection aims to identify OOD data based on
representations extracted from well-trained deep models. However, existing methods largely …

Harnessing out-of-distribution examples via augmenting content and style

Z Huang, X Xia, L Shen, B Han, M Gong… - arXiv preprint arXiv …, 2022 - arxiv.org
Machine learning models are vulnerable to Out-Of-Distribution (OOD) examples, and such a
problem has drawn much attention. However, current methods lack a full understanding of …

Better safe than sorry: Preventing delusive adversaries with adversarial training

L Tao, L Feng, J Yi, SJ Huang… - Advances in Neural …, 2021 - proceedings.neurips.cc
Delusive attacks aim to substantially deteriorate the test accuracy of the learning model by
slightly perturbing the features of correctly labeled training examples. By formalizing this …

The enemy of my enemy is my friend: Exploring inverse adversaries for improving adversarial training

J Dong, SM Moosavi-Dezfooli… - Proceedings of the …, 2023 - openaccess.thecvf.com
Although current deep learning techniques have yielded superior performance on various
computer vision tasks, yet they are still vulnerable to adversarial examples. Adversarial …

Certified robustness via dynamic margin maximization and improved lipschitz regularization

M Fazlyab, T Entesari, A Roy… - Advances in Neural …, 2024 - proceedings.neurips.cc
To improve the robustness of deep classifiers against adversarial perturbations, many
approaches have been proposed, such as designing new architectures with better …

Exploring and exploiting decision boundary dynamics for adversarial robustness

Y Xu, Y Sun, M Goldblum, T Goldstein… - arXiv preprint arXiv …, 2023 - arxiv.org
The robustness of a deep classifier can be characterized by its margins: the decision
boundary's distances to natural data points. However, it is unclear whether existing robust …

Defenses in adversarial machine learning: A survey

B Wu, S Wei, M Zhu, M Zheng, Z Zhu, M Zhang… - arXiv preprint arXiv …, 2023 - arxiv.org
Adversarial phenomenon has been widely observed in machine learning (ML) systems,
especially in those using deep neural networks, describing that ML systems may produce …