Red-Teaming for Generative AI: Silver Bullet or Security Theater?
In response to rising concerns surrounding the safety, security, and trustworthiness of
Generative AI (GenAI) models, practitioners and regulators alike have pointed to AI red …
Generative AI (GenAI) models, practitioners and regulators alike have pointed to AI red …
Jailbreaker in jail: Moving target defense for large language models
Large language models (LLMs), known for their capability in understanding and following
instructions, are vulnerable to adversarial attacks. Researchers have found that current …
instructions, are vulnerable to adversarial attacks. Researchers have found that current …
Beyond Boundaries: A Comprehensive Survey of Transferable Attacks on AI Systems
Artificial Intelligence (AI) systems such as autonomous vehicles, facial recognition, and
speech recognition systems are increasingly integrated into our daily lives. However …
speech recognition systems are increasingly integrated into our daily lives. However …
Multi-turn hidden backdoor in large language model-powered chatbot models
Large Language Model (LLM)-powered chatbot services like GPTs, simulating human-to-
human conversation via machine-generated text, are used in numerous fields. They are …
human conversation via machine-generated text, are used in numerous fields. They are …
DynamicFL: Balancing Communication Dynamics and Client Manipulation for Federated Learning
Federated Learning (FL) is a distributed machine learning (ML) paradigm, aiming to train a
global model by exploiting the decentralized data across millions of edge devices …
global model by exploiting the decentralized data across millions of edge devices …
Chain of Attack: a Semantic-Driven Contextual Multi-Turn attacker for LLM
X Yang, X Tang, S Hu, J Han - arXiv preprint arXiv:2405.05610, 2024 - arxiv.org
Large language models (LLMs) have achieved remarkable performance in various natural
language processing tasks, especially in dialogue systems. However, LLM may also pose …
language processing tasks, especially in dialogue systems. However, LLM may also pose …
A Systematic Review of Toxicity in Large Language Models: Definitions, Datasets, Detectors, Detoxification Methods and Challenges
G Villate-Castillo, JDS Lorente, BS Urquijo - 2024 - researchsquare.com
The emergence of the transformer architecture has ushered in a new era of possibilities,
showcasing remarkable capabilities in generative tasks exemplified by models like GPT4o …
showcasing remarkable capabilities in generative tasks exemplified by models like GPT4o …
IntentObfuscator: A Jailbreaking Method via Confusing LLM with Prompts
In the era of Large Language Models (LLMs), developers establish content review
conditions to comply with legal, policy, and societal requirements, aiming to prevent the …
conditions to comply with legal, policy, and societal requirements, aiming to prevent the …
The Personification of ChatGPT (GPT-4)—Understanding Its Personality and Adaptability
L Stöckli, L Joho, F Lehner, T Hanne - Information, 2024 - mdpi.com
Thanks to the publication of ChatGPT, Artificial Intelligence is now basically accessible and
usable to all internet users. The technology behind it can be used in many chatbots …
usable to all internet users. The technology behind it can be used in many chatbots …
Protecting Activity Sensing Data Privacy Using Hierarchical Information Dissociation
Smartphones and wearable devices have been integrated into our daily lives, offering
personalized services. However, many apps become overprivileged as their collected …
personalized services. However, many apps become overprivileged as their collected …