Securing large language models: Addressing bias, misinformation, and prompt attacks

B Peng, K Chen, M Li, P Feng, Z Bi, J Liu… - arXiv preprint arXiv …, 2024 - arxiv.org
Large Language Models (LLMs) demonstrate impressive capabilities across various fields,
yet their increasing use raises critical security concerns. This article reviews recent literature …

Instruction backdoor attacks against customized {LLMs}

R Zhang, H Li, R Wen, W Jiang, Y Zhang… - 33rd USENIX Security …, 2024 - usenix.org
The increasing demand for customized Large Language Models (LLMs) has led to the
development of solutions like GPTs. These solutions facilitate tailored LLM creation via …

Transformers and large language models for efficient intrusion detection systems: A comprehensive survey

H Kheddar - arXiv preprint arXiv:2408.07583, 2024 - arxiv.org
With significant advancements in Transformers LLMs, NLP has extended its reach into many
research fields due to its enhanced capabilities in text generation and user interaction. One …

A survey of backdoor attacks and defenses on large language models: Implications for security measures

S Zhao, M Jia, Z Guo, L Gan, X Xu, X Wu, J Fu… - arXiv preprint arXiv …, 2024 - arxiv.org
Large Language Models (LLMs), which bridge the gap between human language
understanding and complex problem-solving, achieve state-of-the-art performance on …

Defending against weight-poisoning backdoor attacks for parameter-efficient fine-tuning

S Zhao, L Gan, LA Tuan, J Fu, L Lyu, M Jia… - arXiv preprint arXiv …, 2024 - arxiv.org
Recently, various parameter-efficient fine-tuning (PEFT) strategies for application to
language models have been proposed and successfully implemented. However, this raises …

Artwork protection against neural style transfer using locally adaptive adversarial color attack

Z Guo, J Dong, Y Qian, K Wang, W Li, Z Guo… - ECAI 2024, 2024 - ebooks.iospress.nl
Neural style transfer (NST) generates new images by combining the style of one image with
the content of another. However, unauthorized NST can exploit artwork, raising concerns …

Enhancing federated semi-supervised learning with out-of-distribution filtering amidst class mismatches

J Jin, F Ni, S Dai, K Li, B Hong - Journal of Computer Technology …, 2024 - suaspress.org
Federated Learning (FL) has gained prominence as a method for training models on edge
computing devices, enabling the preservation of data privacy by eliminating the need to …

Compromising embodied agents with contextual backdoor attacks

A Liu, Y Zhou, X Liu, T Zhang, S Liang, J Wang… - arXiv preprint arXiv …, 2024 - arxiv.org
Large language models (LLMs) have transformed the development of embodied
intelligence. By providing a few contextual demonstrations, developers can utilize the …

Safeguarding Large Language Models: A Survey

Y Dong, R Mu, Y Zhang, S Sun, T Zhang, C Wu… - arXiv preprint arXiv …, 2024 - arxiv.org
In the burgeoning field of Large Language Models (LLMs), developing a robust safety
mechanism, colloquially known as" safeguards" or" guardrails", has become imperative to …

When llms meet cybersecurity: A systematic literature review

J Zhang, H Bu, H Wen, Y Chen, L Li, H Zhu - arXiv preprint arXiv …, 2024 - arxiv.org
The rapid advancements in large language models (LLMs) have opened new avenues
across various fields, including cybersecurity, which faces an ever-evolving threat landscape …