Mambair: A simple baseline for image restoration with state-space model
Recent years have seen significant advancements in image restoration, largely attributed to
the development of modern deep neural networks, such as CNNs and Transformers …
the development of modern deep neural networks, such as CNNs and Transformers …
Badclip: Dual-embedding guided backdoor attack on multimodal contrastive learning
While existing backdoor attacks have successfully infected multimodal contrastive learning
models such as CLIP they can be easily countered by specialized backdoor defenses for …
models such as CLIP they can be easily countered by specialized backdoor defenses for …
Backdoor Attacks and Defenses Targeting Multi-Domain AI Models: A Comprehensive Review
Since the emergence of security concerns in artificial intelligence (AI), there has been
significant attention devoted to the examination of backdoor attacks. Attackers can utilize …
significant attention devoted to the examination of backdoor attacks. Attackers can utilize …
Backdoorllm: A comprehensive benchmark for backdoor attacks on large language models
Generative Large Language Models (LLMs) have made significant strides across various
tasks, but they remain vulnerable to backdoor attacks, where specific triggers in the prompt …
tasks, but they remain vulnerable to backdoor attacks, where specific triggers in the prompt …
Test-time backdoor attacks on multimodal large language models
Backdoor attacks are commonly executed by contaminating training data, such that a trigger
can activate predetermined harmful effects during the test phase. In this work, we present …
can activate predetermined harmful effects during the test phase. In this work, we present …
Pointncbw: Towards dataset ownership verification for point clouds via negative clean-label backdoor watermark
Recently, point clouds have been widely used in computer vision, whereas their collection is
time-consuming and expensive. As such, point cloud datasets are the valuable intellectual …
time-consuming and expensive. As such, point cloud datasets are the valuable intellectual …
Parameter-efficient and memory-efficient tuning for vision transformer: a disentangled approach
Recent works on parameter-efficient transfer learning (PETL) show the potential to adapt a
pre-trained Vision Transformer to downstream recognition tasks with only a few learnable …
pre-trained Vision Transformer to downstream recognition tasks with only a few learnable …
Adversarial backdoor defense in clip
Multimodal contrastive pretraining, exemplified by models like CLIP, has been found to be
vulnerable to backdoor attacks. While current backdoor defense methods primarily employ …
vulnerable to backdoor attacks. While current backdoor defense methods primarily employ …
Energy-latency manipulation of multi-modal large language models via verbose samples
Despite the exceptional performance of multi-modal large language models (MLLMs), their
deployment requires substantial computational resources. Once malicious users induce …
deployment requires substantial computational resources. Once malicious users induce …
Adversarial robustness for visual grounding of multimodal large language models
Multi-modal Large Language Models (MLLMs) have recently achieved enhanced
performance across various vision-language tasks including visual grounding capabilities …
performance across various vision-language tasks including visual grounding capabilities …