Privacy in large language models: Attacks, defenses and future directions
The advancement of large language models (LLMs) has significantly enhanced the ability to
effectively tackle various downstream NLP tasks and unify these tasks into generative …
effectively tackle various downstream NLP tasks and unify these tasks into generative …
Llm-based edge intelligence: A comprehensive survey on architectures, applications, security and trustworthiness
The integration of Large Language Models (LLMs) and Edge Intelligence (EI) introduces a
groundbreaking paradigm for intelligent edge devices. With their capacity for human-like …
groundbreaking paradigm for intelligent edge devices. With their capacity for human-like …
The instruction hierarchy: Training llms to prioritize privileged instructions
Today's LLMs are susceptible to prompt injections, jailbreaks, and other attacks that allow
adversaries to overwrite a model's original instructions with their own malicious prompts. In …
adversaries to overwrite a model's original instructions with their own malicious prompts. In …
Injecagent: Benchmarking indirect prompt injections in tool-integrated large language model agents
Recent work has embodied LLMs as agents, allowing them to access tools, perform actions,
and interact with external content (eg, emails or websites). However, external content …
and interact with external content (eg, emails or websites). However, external content …
Eia: Environmental injection attack on generalist web agents for privacy leakage
Generalist web agents have evolved rapidly and demonstrated remarkable potential.
However, there are unprecedented safety risks associated with these them, which are nearly …
However, there are unprecedented safety risks associated with these them, which are nearly …
VLMGuard: Defending VLMs against Malicious Prompts via Unlabeled Data
Vision-language models (VLMs) are essential for contextual understanding of both visual
and textual information. However, their vulnerability to adversarially manipulated inputs …
and textual information. However, their vulnerability to adversarially manipulated inputs …
PROMPTFUZZ: Harnessing Fuzzing Techniques for Robust Testing of Prompt Injection in LLMs
Large Language Models (LLMs) have gained widespread use in various applications due to
their powerful capability to generate human-like text. However, prompt injection attacks …
their powerful capability to generate human-like text. However, prompt injection attacks …
AgentDojo: A Dynamic Environment to Evaluate Attacks and Defenses for LLM Agents
AI agents aim to solve complex tasks by combining text-based reasoning with external tool
calls. Unfortunately, AI agents are vulnerable to prompt injection attacks where data returned …
calls. Unfortunately, AI agents are vulnerable to prompt injection attacks where data returned …
Adversarial Search Engine Optimization for Large Language Models
Large Language Models (LLMs) are increasingly used in applications where the model
selects from competing third-party content, such as in LLM-powered search engines or …
selects from competing third-party content, such as in LLM-powered search engines or …