Poisoning web-scale training datasets is practical

N Carlini, M Jagielski, CA Choquette-Choo… - arXiv preprint arXiv …, 2023 - arxiv.org
Deep learning models are often trained on distributed, webscale datasets crawled from the
internet. In this paper, we introduce two new dataset poisoning attacks that intentionally …

Invisible backdoor attack with sample-specific triggers

Y Li, Y Li, B Wu, L Li, R He… - Proceedings of the IEEE …, 2021 - openaccess.thecvf.com
Recently, backdoor attacks pose a new security threat to the training process of deep neural
networks (DNNs). Attackers intend to inject hidden backdoors into DNNs, such that the …

Backdoor learning: A survey

Y Li, Y Jiang, Z Li, ST Xia - IEEE Transactions on Neural …, 2022 - ieeexplore.ieee.org
Backdoor attack intends to embed hidden backdoors into deep neural networks (DNNs), so
that the attacked models perform well on benign samples, whereas their predictions will be …

Backdoor defense via decoupling the training process

K Huang, Y Li, B Wu, Z Qin, K Ren - arXiv preprint arXiv:2202.03423, 2022 - arxiv.org
Recent studies have revealed that deep neural networks (DNNs) are vulnerable to backdoor
attacks, where attackers embed hidden backdoors in the DNN model by poisoning a few …

Better trigger inversion optimization in backdoor scanning

G Tao, G Shen, Y Liu, S An, Q Xu… - Proceedings of the …, 2022 - openaccess.thecvf.com
Backdoor attacks aim to cause misclassification of a subject model by stamping a trigger to
inputs. Backdoors could be injected through malicious training and naturally exist. Deriving …

Hidden trigger backdoor attack on {NLP} models via linguistic style manipulation

X Pan, M Zhang, B Sheng, J Zhu, M Yang - 31st USENIX Security …, 2022 - usenix.org
The vulnerability of deep neural networks (DNN) to backdoor (trojan) attacks is extensively
studied for the image domain. In a backdoor attack, a DNN is modified to exhibit expected …

Rethinking the trigger of backdoor attack

Y Li, T Zhai, B Wu, Y Jiang, Z Li, S Xia - arXiv preprint arXiv:2004.04692, 2020 - arxiv.org
Backdoor attack intends to inject hidden backdoor into the deep neural networks (DNNs),
such that the prediction of the infected model will be maliciously changed if the hidden …

Aeva: Black-box backdoor detection using adversarial extreme value analysis

J Guo, A Li, C Liu - arXiv preprint arXiv:2110.14880, 2021 - arxiv.org
Deep neural networks (DNNs) are proved to be vulnerable against backdoor attacks. A
backdoor is often embedded in the target DNNs through injecting a backdoor trigger into …

Model orthogonalization: Class distance hardening in neural networks for better security

G Tao, Y Liu, G Shen, Q Xu, S An… - … IEEE Symposium on …, 2022 - ieeexplore.ieee.org
The distance between two classes for a deep learning classifier can be measured by the
level of difficulty in flipping all (or majority of) samples in a class to the other. The class …

Few-shot backdoor defense using shapley estimation

J Guan, Z Tu, R He, D Tao - Proceedings of the IEEE/CVF …, 2022 - openaccess.thecvf.com
Deep neural networks have achieved impressive performance in a variety of tasks over the
last decade, such as autonomous driving, face recognition, and medical diagnosis …