Securing large language models: Addressing bias, misinformation, and prompt attacks

B Peng, K Chen, M Li, P Feng, Z Bi, J Liu… - arXiv preprint arXiv …, 2024 - arxiv.org
Large Language Models (LLMs) demonstrate impressive capabilities across various fields,
yet their increasing use raises critical security concerns. This article reviews recent literature …

[HTML][HTML] A survey on the use of large language models (llms) in fake news

E Papageorgiou, C Chronis, I Varlamis, Y Himeur - Future Internet, 2024 - mdpi.com
The proliferation of fake news and fake profiles on social media platforms poses significant
threats to information integrity and societal trust. Traditional detection methods, including …

Preference tuning with human feedback on language, speech, and vision tasks: A survey

GI Winata, H Zhao, A Das, W Tang, DD Yao… - arXiv preprint arXiv …, 2024 - arxiv.org
Preference tuning is a crucial process for aligning deep generative models with human
preferences. This survey offers a thorough overview of recent advancements in preference …

Catching chameleons: Detecting evolving disinformation generated using large language models

B Jiang, C Zhao, Z Tan, H Liu - arXiv preprint arXiv:2406.17992, 2024 - arxiv.org
Despite recent advancements in detecting disinformation generated by large language
models (LLMs), current efforts overlook the ever-evolving nature of this disinformation. In this …

Generative monoculture in large language models

F Wu, E Black, V Chandrasekaran - arXiv preprint arXiv:2407.02209, 2024 - arxiv.org
We introduce {\em generative monoculture}, a behavior observed in large language models
(LLMs) characterized by a significant narrowing of model output diversity relative to …

Model attribution in llm-generated disinformation: A domain generalization approach with supervised contrastive learning

A Beigi, Z Tan, N Mudiam, C Chen… - 2024 IEEE 11th …, 2024 - ieeexplore.ieee.org
Model attribution for LLM-generated disinformation poses a significant challenge in
understanding its origins and mitigating its spread. This task is especially challenging …

Seeing Through AI's Lens: Enhancing Human Skepticism Towards LLM-Generated Fake News

N Ayoobi, S Shahriar, A Mukherjee - … of the 35th ACM Conference on …, 2024 - dl.acm.org
LLMs offer valuable capabilities, yet they can be utilized by malicious users to disseminate
deceptive information and generate fake news. The growing prevalence of LLMs poses …

Safe+ Safe= Unsafe? Exploring How Safe Images Can Be Exploited to Jailbreak Large Vision-Language Models

C Cui, G Deng, A Zhang, J Zheng, Y Li, L Gao… - arXiv preprint arXiv …, 2024 - arxiv.org
Recent advances in Large Vision-Language Models (LVLMs) have showcased strong
reasoning abilities across multiple modalities, achieving significant breakthroughs in various …

Humanizing the Machine: Proxy Attacks to Mislead LLM Detectors

T Wang, Y Chen, Z Liu, Z Chen, H Chen… - arXiv preprint arXiv …, 2024 - arxiv.org
The advent of large language models (LLMs) has revolutionized the field of text generation,
producing outputs that closely mimic human-like writing. Although academic and industrial …

Cross-attention multi-perspective fusion network based fake news censorship

W Zhang, M Zhang, Z Bao, Z Wang - Neurocomputing, 2025 - Elsevier
Current fake news censorship models mostly use only one single semantic perspective,
which contains insufficient information and may result in biases. However, news inherently …