A survey of adversarial defenses and robustness in nlp
In the past few years, it has become increasingly evident that deep neural networks are not
resilient enough to withstand adversarial perturbations in input data, leaving them …
resilient enough to withstand adversarial perturbations in input data, leaving them …
Robust natural language processing: Recent advances, challenges, and future directions
Recent natural language processing (NLP) techniques have accomplished high
performance on benchmark data sets, primarily due to the significant improvement in the …
performance on benchmark data sets, primarily due to the significant improvement in the …
Prompt as triggers for backdoor attack: Examining the vulnerability in language models
The prompt-based learning paradigm, which bridges the gap between pre-training and fine-
tuning, achieves state-of-the-art performance on several NLP tasks, particularly in few-shot …
tuning, achieves state-of-the-art performance on several NLP tasks, particularly in few-shot …
Defending against alignment-breaking attacks via robustly aligned llm
Recently, Large Language Models (LLMs) have made significant advancements and are
now widely used across various domains. Unfortunately, there has been a rising concern …
now widely used across various domains. Unfortunately, there has been a rising concern …
Towards improving adversarial training of NLP models
Adversarial training, a method for learning robust deep neural networks, constructs
adversarial examples during training. However, recent methods for generating NLP …
adversarial examples during training. However, recent methods for generating NLP …
Evaluating the robustness of neural language models to input perturbations
High-performance neural language models have obtained state-of-the-art results on a wide
range of Natural Language Processing (NLP) tasks. However, results for common …
range of Natural Language Processing (NLP) tasks. However, results for common …
Searching for an effective defender: Benchmarking defense against adversarial word substitution
Recent studies have shown that deep neural networks are vulnerable to intentionally crafted
adversarial examples, and various methods have been proposed to defend against …
adversarial examples, and various methods have been proposed to defend against …
How should pre-trained language models be fine-tuned towards adversarial robustness?
The fine-tuning of pre-trained language models has a great success in many NLP fields. Yet,
it is strikingly vulnerable to adversarial examples, eg, word substitution attacks using only …
it is strikingly vulnerable to adversarial examples, eg, word substitution attacks using only …
Transformer models used for text-based question answering systems
K Nassiri, M Akhloufi - Applied Intelligence, 2023 - Springer
The question answering system is frequently applied in the area of natural language
processing (NLP) because of the wide variety of applications. It consists of answering …
processing (NLP) because of the wide variety of applications. It consists of answering …
Prada: Practical black-box adversarial attacks against neural ranking models
Neural ranking models (NRMs) have shown remarkable success in recent years, especially
with pre-trained language models. However, deep neural models are notorious for their …
with pre-trained language models. However, deep neural models are notorious for their …