Prompt-specific poisoning attacks on text-to-image generative models

S Shan, W Ding, J Passananti, H Zheng… - arXiv preprint arXiv …, 2023 - arxiv.org
Data poisoning attacks manipulate training data to introduce unexpected behaviors into
machine learning models at training time. For text-to-image generative models with massive …

Badmerging: Backdoor attacks against model merging

J Zhang, J Chi, Z Li, K Cai, Y Zhang… - Proceedings of the 2024 on …, 2024 - dl.acm.org
Fine-tuning pre-trained models for downstream tasks has led to a proliferation of open-
sourced task-specific models. Recently, Model Merging (MM) has emerged as an effective …

Ssl-cleanse: Trojan detection and mitigation in self-supervised learning

M Zheng, J Xue, Z Wang, X Chen, Q Lou… - … on Computer Vision, 2025 - Springer
Self-supervised learning (SSL) is a prevalent approach for encoding data representations.
Using a pre-trained SSL image encoder and subsequently training a downstream classifier …

Nightshade: Prompt-Specific Poisoning Attacks on Text-to-Image Generative Models

S Shan, W Ding, J Passananti, S Wu… - 2024 IEEE Symposium …, 2024 - computer.org
Trained on billions of images, diffusion-based text-to-image models seem impervious to
traditional data poisoning attacks, which typically require poison samples approaching 20 …

Semantic Shield: Defending Vision-Language Models Against Backdooring and Poisoning via Fine-grained Knowledge Alignment

AM Ishmam, C Thomas - … of the IEEE/CVF Conference on …, 2024 - openaccess.thecvf.com
In recent years there has been enormous interest in vision-language models trained using
self-supervised objectives. However the use of large-scale datasets scraped from the web …

Transtroj: Transferable backdoor attacks to pre-trained models via embedding indistinguishability

H Wang, T Xiang, S Guo, J He, H Liu… - arXiv preprint arXiv …, 2024 - arxiv.org
Pre-trained models (PTMs) are extensively utilized in various downstream tasks. Adopting
untrusted PTMs may suffer from backdoor attacks, where the adversary can compromise the …

FCert: Certifiably Robust Few-Shot Classification in the Era of Foundation Models

Y Wang, W Zou, J Jia - arXiv preprint arXiv:2404.08631, 2024 - arxiv.org
Few-shot classification with foundation models (eg, CLIP, DINOv2, PaLM-2) enables users
to build an accurate classifier with a few labeled training samples (called support samples) …

Exploring the Vulnerability of Self-supervised Monocular Depth Estimation Models

R Hou, K Mo, Y Long, N Li, Y Rao - Information Sciences, 2024 - Elsevier
Recent advancements in deep learning have substantially boosted the performance of
monocular depth estimation (MDE), an essential component in fully-vision-based …

Backdoor Contrastive Learning via Bi-level Trigger Optimization

W Sun, X Zhang, H Lu, Y Chen, T Wang… - arXiv preprint arXiv …, 2024 - arxiv.org
Contrastive Learning (CL) has attracted enormous attention due to its remarkable capability
in unsupervised representation learning. However, recent works have revealed the …

[图书][B] Secure and Private Large Transformers

M Zheng - 2023 - search.proquest.com
Deep Learning's integration into critical sectors like autonomous vehicles and healthcare
diagnosis underscores the necessity for creating learning methods that are safe, secure …