Precurious: How innocent pre-trained language models turn into privacy traps
The pre-training and fine-tuning paradigm has demonstrated its effectiveness and has
become the standard approach for tailoring language models to various tasks. Currently …
become the standard approach for tailoring language models to various tasks. Currently …
Forget to flourish: Leveraging machine-unlearning on pretrained language models for privacy leakage
Fine-tuning large language models on private data for downstream applications poses
significant privacy risks in potentially exposing sensitive information. Several popular …
significant privacy risks in potentially exposing sensitive information. Several popular …
Rag-thief: Scalable extraction of private data from retrieval-augmented generation applications with agent-based attacks
While large language models (LLMs) have achieved notable success in generative tasks,
they still face limitations, such as lacking up-to-date knowledge and producing …
they still face limitations, such as lacking up-to-date knowledge and producing …
Generative AI model privacy: a survey
The rapid progress of generative AI models has yielded substantial breakthroughs in AI,
facilitating the generation of realistic synthetic data across various modalities. However …
facilitating the generation of realistic synthetic data across various modalities. However …
ExpShield: Safeguarding Web Text from Unauthorized Crawling and Language Modeling Exploitation
As large language models (LLMs) increasingly depend on web-scraped datasets, concerns
over unauthorized use of copyrighted or personal content for training have intensified …
over unauthorized use of copyrighted or personal content for training have intensified …
Injecting Undetectable Backdoors in Deep Learning and Language Models
As ML models become increasingly complex and integral to high-stakes domains such as
finance and healthcare, they also become more susceptible to sophisticated adversarial …
finance and healthcare, they also become more susceptible to sophisticated adversarial …
Pre-trained Encoder Inference: Revealing Upstream Encoders In Downstream Machine Learning Services
Though pre-trained encoders can be easily accessed online to build downstream machine
learning (ML) services quickly, various attacks have been designed to compromise the …
learning (ML) services quickly, various attacks have been designed to compromise the …
Sales Whisperer: A Human-Inconspicuous Attack on LLM Brand Recommendations
Large language model (LLM) users might rely on others (eg, prompting services), to write
prompts. However, the risks of trusting prompts written by others remain unstudied. In this …
prompts. However, the risks of trusting prompts written by others remain unstudied. In this …
Meanings and Feelings of Large Language Models: Observability of Latent States in Generative AI
We tackle the question of whether Large Language Models (LLMs), viewed as dynamical
systems with state evolving in the embedding space of symbolic tokens, are observable …
systems with state evolving in the embedding space of symbolic tokens, are observable …
Privacy in Fine-tuning Large Language Models: Attacks, Defenses, and Future Directions
Fine-tuning has emerged as a critical process in leveraging Large Language Models (LLMs)
for specific downstream tasks, enabling these models to achieve state-of-the-art …
for specific downstream tasks, enabling these models to achieve state-of-the-art …