Learning from noisy labels with deep neural networks: A survey

H Song, M Kim, D Park, Y Shin… - IEEE transactions on …, 2022 - ieeexplore.ieee.org
Deep learning has achieved remarkable success in numerous domains with help from large
amounts of big data. However, the quality of data labels is a concern because of the lack of …

Adversarial attacks and defenses in images, graphs and text: A review

H Xu, Y Ma, HC Liu, D Deb, H Liu, JL Tang… - International journal of …, 2020 - Springer
Deep neural networks (DNN) have achieved unprecedented success in numerous machine
learning tasks in various domains. However, the existence of adversarial examples raises …

Fit without fear: remarkable mathematical phenomena of deep learning through the prism of interpolation

M Belkin - Acta Numerica, 2021 - cambridge.org
In the past decade the mathematical theory of machine learning has lagged far behind the
triumphs of deep neural networks on practical challenges. However, the gap between theory …

Adversarial examples are not bugs, they are features

A Ilyas, S Santurkar, D Tsipras… - Advances in neural …, 2019 - proceedings.neurips.cc
Adversarial examples have attracted significant attention in machine learning, but the
reasons for their existence and pervasiveness remain unclear. We demonstrate that …

Certified adversarial robustness via randomized smoothing

J Cohen, E Rosenfeld, Z Kolter - international conference on …, 2019 - proceedings.mlr.press
We show how to turn any classifier that classifies well under Gaussian noise into a new
classifier that is certifiably robust to adversarial perturbations under the L2 norm. While this" …

Salient object detection in the deep learning era: An in-depth survey

W Wang, Q Lai, H Fu, J Shen, H Ling… - IEEE Transactions on …, 2021 - ieeexplore.ieee.org
As an essential problem in computer vision, salient object detection (SOD) has attracted an
increasing amount of research attention over the years. Recent advances in SOD are …

Robustness may be at odds with accuracy

D Tsipras, S Santurkar, L Engstrom, A Turner… - arXiv preprint arXiv …, 2018 - arxiv.org
We show that there may exist an inherent tension between the goal of adversarial
robustness and that of standard generalization. Specifically, training robust models may not …

Machine learning with adversaries: Byzantine tolerant gradient descent

P Blanchard, EM El Mhamdi… - Advances in neural …, 2017 - proceedings.neurips.cc
We study the resilience to Byzantine failures of distributed implementations of Stochastic
Gradient Descent (SGD). So far, distributed machine learning frameworks have largely …

Countering adversarial images using input transformations

C Guo, M Rana, M Cisse, L Van Der Maaten - arXiv preprint arXiv …, 2017 - arxiv.org
This paper investigates strategies that defend against adversarial-example attacks on image-
classification systems by transforming the inputs before feeding them to the system …

Adversarially robust generalization requires more data

L Schmidt, S Santurkar, D Tsipras… - Advances in neural …, 2018 - proceedings.neurips.cc
Abstract Machine learning models are often susceptible to adversarial perturbations of their
inputs. Even small perturbations can cause state-of-the-art classifiers with high" standard" …