Combating misinformation in the age of llms: Opportunities and challenges

C Chen, K Shu - AI Magazine, 2024 - Wiley Online Library
Misinformation such as fake news and rumors is a serious threat for information ecosystems
and public trust. The emergence of large language models (LLMs) has great potential to …

Explainable generative ai (genxai): A survey, conceptualization, and research agenda

J Schneider - Artificial Intelligence Review, 2024 - Springer
Generative AI (GenAI) represents a shift from AI's ability to “recognize” to its ability to
“generate” solutions for a wide range of tasks. As generated solutions and applications grow …

Siren's song in the AI ocean: a survey on hallucination in large language models

Y Zhang, Y Li, L Cui, D Cai, L Liu, T Fu… - arXiv preprint arXiv …, 2023 - arxiv.org
While large language models (LLMs) have demonstrated remarkable capabilities across a
range of downstream tasks, a significant concern revolves around their propensity to exhibit …

Explainability for large language models: A survey

H Zhao, H Chen, F Yang, N Liu, H Deng, H Cai… - ACM Transactions on …, 2024 - dl.acm.org
Large language models (LLMs) have demonstrated impressive capabilities in natural
language processing. However, their internal mechanisms are still unclear and this lack of …

[PDF][PDF] Calibration and correctness of language models for code

C Spiess, D Gros, KS Pai, M Pradel… - arXiv preprint arXiv …, 2024 - software-lab.org
Machine learning models are widely used, but can also often be wrong. Users would benefit
from a reliable indication of whether a given output from a given model should be trusted, so …

Decomposing uncertainty for large language models through input clarification ensembling

B Hou, Y Liu, K Qian, J Andreas, S Chang… - arXiv preprint arXiv …, 2023 - arxiv.org
Uncertainty decomposition refers to the task of decomposing the total uncertainty of a model
into data (aleatoric) uncertainty, resulting from the inherent complexity or ambiguity of the …

A new era in llm security: Exploring security concerns in real-world llm-based systems

F Wu, N Zhang, S Jha, P McDaniel, C Xiao - arXiv preprint arXiv …, 2024 - arxiv.org
Large Language Model (LLM) systems are inherently compositional, with individual LLM
serving as the core foundation with additional layers of objects such as plugins, sandbox …

Think twice before assure: Confidence estimation for large language models through reflection on multiple answers

M Li, W Wang, F Feng, F Zhu, Q Wang… - arXiv preprint arXiv …, 2024 - arxiv.org
Confidence estimation aiming to evaluate output trustability is crucial for the application of
large language models (LLM), especially the black-box ones. Existing confidence estimation …

Benchmarking llms via uncertainty quantification

F Ye, M Yang, J Pang, L Wang, DF Wong… - arXiv preprint arXiv …, 2024 - arxiv.org
The proliferation of open-source Large Language Models (LLMs) from various institutions
has highlighted the urgent need for comprehensive evaluation methods. However, current …

A Survey of Confidence Estimation and Calibration in Large Language Models

J Geng, F Cai, Y Wang, H Koeppl… - Proceedings of the …, 2024 - aclanthology.org
Large language models (LLMs) have demonstrated remarkable capabilities across a wide
range of tasks in various domains. Despite their impressive performance, they can be …