Review of artificial intelligence adversarial attack and defense technologies

S Qiu, Q Liu, S Zhou, C Wu - Applied Sciences, 2019 - mdpi.com
In recent years, artificial intelligence technologies have been widely used in computer
vision, natural language processing, automatic driving, and other fields. However, artificial …

Towards trustworthy and aligned machine learning: A data-centric survey with causality perspectives

H Liu, M Chaudhary, H Wang - arXiv preprint arXiv:2307.16851, 2023 - arxiv.org
The trustworthiness of machine learning has emerged as a critical topic in the field,
encompassing various applications and research areas such as robustness, security …

Countering adversarial images using input transformations

C Guo, M Rana, M Cisse, L Van Der Maaten - arXiv preprint arXiv …, 2017 - arxiv.org
This paper investigates strategies that defend against adversarial-example attacks on image-
classification systems by transforming the inputs before feeding them to the system …

Threat of adversarial attacks on deep learning in computer vision: A survey

N Akhtar, A Mian - Ieee Access, 2018 - ieeexplore.ieee.org
Deep learning is at the heart of the current rise of artificial intelligence. In the field of
computer vision, it has become the workhorse for applications ranging from self-driving cars …

Lemna: Explaining deep learning based security applications

W Guo, D Mu, J Xu, P Su, G Wang, X Xing - proceedings of the 2018 …, 2018 - dl.acm.org
While deep learning has shown a great potential in various domains, the lack of
transparency has limited its application in security or safety-critical areas. Existing research …

Mass-producing failures of multimodal systems with language models

S Tong, E Jones, J Steinhardt - Advances in Neural …, 2024 - proceedings.neurips.cc
Deployed multimodal models can fail in ways that evaluators did not anticipate. In order to
find these failures before deployment, we introduce MultiMon, a system that automatically …

Detecting adversarial image examples in deep neural networks with adaptive noise reduction

B Liang, H Li, M Su, X Li, W Shi… - IEEE Transactions on …, 2018 - ieeexplore.ieee.org
Recently, many studies have demonstrated deep neural network (DNN) classifiers can be
fooled by the adversarial example, which is crafted via introducing some perturbations into …

Training robust deep neural networks via adversarial noise propagation

A Liu, X Liu, H Yu, C Zhang, Q Liu… - IEEE Transactions on …, 2021 - ieeexplore.ieee.org
In practice, deep neural networks have been found to be vulnerable to various types of
noise, such as adversarial examples and corruption. Various adversarial defense methods …

Learn2perturb: an end-to-end feature perturbation learning to improve adversarial robustness

A Jeddi, MJ Shafiee, M Karg… - Proceedings of the …, 2020 - openaccess.thecvf.com
While deep neural networks have been achieving state-of-the-art performance across a
wide variety of applications, their vulnerability to adversarial attacks limits their widespread …

[HTML][HTML] A comprehensive survey of robust deep learning in computer vision

J Liu, Y Jin - Journal of Automation and Intelligence, 2023 - Elsevier
Deep learning has presented remarkable progress in various tasks. Despite the excellent
performance, deep learning models remain not robust, especially to well-designed …