[HTML][HTML] A survey on large language model (llm) security and privacy: The good, the bad, and the ugly
Abstract Large Language Models (LLMs), such as ChatGPT and Bard, have revolutionized
natural language understanding and generation. They possess deep language …
natural language understanding and generation. They possess deep language …
Trustllm: Trustworthiness in large language models
Large language models (LLMs), exemplified by ChatGPT, have gained considerable
attention for their excellent natural language processing capabilities. Nonetheless, these …
attention for their excellent natural language processing capabilities. Nonetheless, these …
Instructions as backdoors: Backdoor vulnerabilities of instruction tuning for large language models
We investigate security concerns of the emergent instruction tuning paradigm, that models
are trained on crowdsourced datasets with task instructions to achieve superior …
are trained on crowdsourced datasets with task instructions to achieve superior …
[HTML][HTML] Position: TrustLLM: Trustworthiness in large language models
Large language models (LLMs) have gained considerable attention for their excellent
natural language processing capabilities. Nonetheless, these LLMs present many …
natural language processing capabilities. Nonetheless, these LLMs present many …
Cognitive overload: Jailbreaking large language models with overloaded logical thinking
While large language models (LLMs) have demonstrated increasing power, they have also
given rise to a wide range of harmful behaviors. As representatives, jailbreak attacks can …
given rise to a wide range of harmful behaviors. As representatives, jailbreak attacks can …
Learning to poison large language models during instruction tuning
The advent of Large Language Models (LLMs) has marked significant achievements in
language processing and reasoning capabilities. Despite their advancements, LLMs face …
language processing and reasoning capabilities. Despite their advancements, LLMs face …
Hijacking large language models via adversarial in-context learning
Y Qiang - 2024 - search.proquest.com
In-context learning (ICL) has emerged as a powerful paradigm leveraging LLMs for specific
downstream tasks by utilizing labeled examples as demonstrations in the precondition …
downstream tasks by utilizing labeled examples as demonstrations in the precondition …
Mitigating backdoor threats to large language models: Advancement and challenges
The advancement of Large Language Models (LLMs) has significantly impacted various
domains, including Web search, healthcare, and software development. However, as these …
domains, including Web search, healthcare, and software development. However, as these …
Rethinking Backdoor Detection Evaluation for Language Models
Backdoor attacks, in which a model behaves maliciously when given an attacker-specified
trigger, pose a major security risk for practitioners who depend on publicly released …
trigger, pose a major security risk for practitioners who depend on publicly released …
Combating Security and Privacy Issues in the Era of Large Language Models
This tutorial seeks to provide a systematic summary of risks and vulnerabilities in security,
privacy and copyright aspects of large language models (LLMs), and most recent solutions …
privacy and copyright aspects of large language models (LLMs), and most recent solutions …