Precurious: How innocent pre-trained language models turn into privacy traps

R Liu, T Wang, Y Cao, L Xiong - Proceedings of the 2024 on ACM …, 2024 - dl.acm.org
The pre-training and fine-tuning paradigm has demonstrated its effectiveness and has
become the standard approach for tailoring language models to various tasks. Currently …

Forget to flourish: Leveraging machine-unlearning on pretrained language models for privacy leakage

MRU Rashid, J Liu, T Koike-Akino, S Mehnaz… - arXiv preprint arXiv …, 2024 - arxiv.org
Fine-tuning large language models on private data for downstream applications poses
significant privacy risks in potentially exposing sensitive information. Several popular …

Rag-thief: Scalable extraction of private data from retrieval-augmented generation applications with agent-based attacks

C Jiang, X Pan, G Hong, C Bao, M Yang - arXiv preprint arXiv:2411.14110, 2024 - arxiv.org
While large language models (LLMs) have achieved notable success in generative tasks,
they still face limitations, such as lacking up-to-date knowledge and producing …

Generative AI model privacy: a survey

Y Liu, J Huang, Y Li, D Wang, B Xiao - Artificial Intelligence Review, 2025 - Springer
The rapid progress of generative AI models has yielded substantial breakthroughs in AI,
facilitating the generation of realistic synthetic data across various modalities. However …

ExpShield: Safeguarding Web Text from Unauthorized Crawling and Language Modeling Exploitation

R Liu, T Tran, T Wang, H Hu, S Wang… - arXiv preprint arXiv …, 2024 - arxiv.org
As large language models (LLMs) increasingly depend on web-scraped datasets, concerns
over unauthorized use of copyrighted or personal content for training have intensified …

Injecting Undetectable Backdoors in Deep Learning and Language Models

A Kalavasis, A Karbasi, A Oikonomou, K Sotiraki… - arXiv preprint arXiv …, 2024 - arxiv.org
As ML models become increasingly complex and integral to high-stakes domains such as
finance and healthcare, they also become more susceptible to sophisticated adversarial …

Pre-trained Encoder Inference: Revealing Upstream Encoders In Downstream Machine Learning Services

S Fu, X Sun, K Qing, T Zheng, D Wang - arXiv preprint arXiv:2408.02814, 2024 - arxiv.org
Though pre-trained encoders can be easily accessed online to build downstream machine
learning (ML) services quickly, various attacks have been designed to compromise the …

Sales Whisperer: A Human-Inconspicuous Attack on LLM Brand Recommendations

W Lin, A Gerchanovsky, O Akgul, L Bauer… - arXiv preprint arXiv …, 2024 - arxiv.org
Large language model (LLM) users might rely on others (eg, prompting services), to write
prompts. However, the risks of trusting prompts written by others remain unstudied. In this …

Meanings and Feelings of Large Language Models: Observability of Latent States in Generative AI

TY Liu, S Soatto, M Marchi, P Chaudhari… - arXiv preprint arXiv …, 2024 - arxiv.org
We tackle the question of whether Large Language Models (LLMs), viewed as dynamical
systems with state evolving in the embedding space of symbolic tokens, are observable …

Privacy in Fine-tuning Large Language Models: Attacks, Defenses, and Future Directions

H Du, S Liu, L Zheng, Y Cao, A Nakamura… - arXiv preprint arXiv …, 2024 - arxiv.org
Fine-tuning has emerged as a critical process in leveraging Large Language Models (LLMs)
for specific downstream tasks, enabling these models to achieve state-of-the-art …