Enhancing fine-tuning based backdoor defense with sharpness-aware minimization

M Zhu, S Wei, L Shen, Y Fan… - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
Backdoor defense, which aims to detect or mitigate the effect of malicious triggers introduced
by attackers, is becoming increasingly critical for machine learning security and integrity …

Less is more: Fewer interpretable region via submodular subset selection

R Chen, H Zhang, S Liang, J Li, X Cao - arXiv preprint arXiv:2402.09164, 2024 - arxiv.org
Image attribution algorithms aim to identify important regions that are highly relevant to
model decisions. Although existing attribution solutions can effectively assign importance to …

Poisoned forgery face: Towards backdoor attacks on face forgery detection

J Liang, S Liang, A Liu, X Jia, J Kuang… - arXiv preprint arXiv …, 2024 - arxiv.org
The proliferation of face forgery techniques has raised significant concerns within society,
thereby motivating the development of face forgery detection methods. These methods aim …

Ensemble-based blackbox attacks on dense prediction

Z Cai, Y Tan, MS Asif - … of the IEEE/CVF conference on …, 2023 - openaccess.thecvf.com
We propose an approach for adversarial attacks on dense prediction models (such as object
detectors and segmentation). It is well known that the attacks generated by a single …

Privacy-enhancing face obfuscation guided by semantic-aware attribution maps

J Li, H Zhang, S Liang, P Dai… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Face recognition technology is increasingly being integrated into our daily life, eg Face ID.
With the advancement of machine learning algorithms, the personal information such as …

Does few-shot learning suffer from backdoor attacks?

X Liu, X Jia, J Gu, Y Xun, S Liang, X Cao - Proceedings of the AAAI …, 2024 - ojs.aaai.org
The field of few-shot learning (FSL) has shown promising results in scenarios where training
data is limited, but its vulnerability to backdoor attacks remains largely unexplored. We first …

Vl-trojan: Multimodal instruction backdoor attacks against autoregressive visual language models

J Liang, S Liang, M Luo, A Liu, D Han… - arXiv preprint arXiv …, 2024 - arxiv.org
Autoregressive Visual Language Models (VLMs) showcase impressive few-shot learning
capabilities in a multimodal context. Recently, multimodal instruction tuning has been …

Fast propagation is better: Accelerating single-step adversarial training via sampling subnetworks

X Jia, J Li, J Gu, Y Bai, X Cao - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
Adversarial training has shown promise in building robust models against adversarial
examples. A major drawback of adversarial training is the computational overhead …

Isolation and induction: Training robust deep neural networks against model stealing attacks

J Guo, X Zheng, A Liu, S Liang, Y Xiao, Y Wu… - Proceedings of the 31st …, 2023 - dl.acm.org
Despite the broad application of Machine Learning models as a Service (MLaaS), they are
vulnerable to model stealing attacks. These attacks can replicate the model functionality by …

Face Encryption via Frequency-Restricted Identity-Agnostic Attacks

X Dong, R Wang, S Liang, A Liu, L Jing - Proceedings of the 31st ACM …, 2023 - dl.acm.org
Billions of people are sharing their daily live images on social media everyday. However,
malicious collectors use deep face recognition systems to easily steal their biometric …