Backdoor learning: A survey

Y Li, Y Jiang, Z Li, ST Xia - IEEE Transactions on Neural …, 2022 - ieeexplore.ieee.org
Backdoor attack intends to embed hidden backdoors into deep neural networks (DNNs), so
that the attacked models perform well on benign samples, whereas their predictions will be …

Toward transparent ai: A survey on interpreting the inner structures of deep neural networks

T Räuker, A Ho, S Casper… - 2023 ieee conference …, 2023 - ieeexplore.ieee.org
The last decade of machine learning has seen drastic increases in scale and capabilities.
Deep neural networks (DNNs) are increasingly being deployed in the real world. However …

A survey of neural trojan attacks and defenses in deep learning

J Wang, GM Hassan, N Akhtar - arXiv preprint arXiv:2202.07183, 2022 - arxiv.org
Artificial Intelligence (AI) relies heavily on deep learning-a technology that is becoming
increasingly popular in real-life applications of AI, even in the safety-critical and high-risk …

Position paper: Challenges and opportunities in topological deep learning

T Papamarkou, T Birdal, M Bronstein… - arXiv preprint arXiv …, 2024 - arxiv.org
Topological deep learning (TDL) is a rapidly evolving field that uses topological features to
understand and design deep learning models. This paper posits that TDL may complement …

Notable: Transferable backdoor attacks against prompt-based nlp models

K Mei, Z Li, Z Wang, Y Zhang, S Ma - arXiv preprint arXiv:2305.17826, 2023 - arxiv.org
Prompt-based learning is vulnerable to backdoor attacks. Existing backdoor attacks against
prompt-based models consider injecting backdoors into the entire embedding layers or word …

Attention-enhancing backdoor attacks against bert-based models

W Lyu, S Zheng, L Pang, H Ling, C Chen - arXiv preprint arXiv:2310.14480, 2023 - arxiv.org
Recent studies have revealed that\textit {Backdoor Attacks} can threaten the safety of natural
language processing (NLP) models. Investigating the strategies of backdoor attacks will help …

Defending against patch-based backdoor attacks on self-supervised learning

A Tejankar, M Sanjabi, Q Wang… - Proceedings of the …, 2023 - openaccess.thecvf.com
Recently, self-supervised learning (SSL) was shown to be vulnerable to patch-based data
poisoning backdoor attacks. It was shown that an adversary can poison a small part of the …

A study of the attention abnormality in trojaned berts

W Lyu, S Zheng, T Ma, C Chen - arXiv preprint arXiv:2205.08305, 2022 - arxiv.org
Trojan attacks raise serious security concerns. In this paper, we investigate the underlying
mechanism of Trojaned BERT models. We observe the attention focus drifting behavior of …

Defenses in adversarial machine learning: A survey

B Wu, S Wei, M Zhu, M Zheng, Z Zhu, M Zhang… - arXiv preprint arXiv …, 2023 - arxiv.org
Adversarial phenomenon has been widely observed in machine learning (ML) systems,
especially in those using deep neural networks, describing that ML systems may produce …

Backdoor attack and defense in federated generative adversarial network-based medical image synthesis

R Jin, X Li - Medical Image Analysis, 2023 - Elsevier
Deep Learning-based image synthesis techniques have been applied in healthcare
research for generating medical images to support open research and augment medical …