Challenges and applications of large language models

J Kaddour, J Harris, M Mozes, H Bradley… - arXiv preprint arXiv …, 2023 - arxiv.org
Large Language Models (LLMs) went from non-existent to ubiquitous in the machine
learning discourse within a few years. Due to the fast pace of the field, it is difficult to identify …

Survey of hallucination in natural language generation

Z Ji, N Lee, R Frieske, T Yu, D Su, Y Xu, E Ishii… - ACM Computing …, 2023 - dl.acm.org
Natural Language Generation (NLG) has improved exponentially in recent years thanks to
the development of sequence-to-sequence deep learning technologies such as Transformer …

A survey on hallucination in large language models: Principles, taxonomy, challenges, and open questions

L Huang, W Yu, W Ma, W Zhong, Z Feng… - ACM Transactions on …, 2023 - dl.acm.org
The emergence of large language models (LLMs) has marked a significant breakthrough in
natural language processing (NLP), fueling a paradigm shift in information acquisition …

Halueval: A large-scale hallucination evaluation benchmark for large language models

J Li, X Cheng, WX Zhao, JY Nie, JR Wen - arXiv preprint arXiv:2305.11747, 2023 - arxiv.org
Large language models (LLMs), such as ChatGPT, are prone to generate hallucinations, ie,
content that conflicts with the source or cannot be verified by the factual knowledge. To …

Trustworthy LLMs: A survey and guideline for evaluating large language models' alignment

Y Liu, Y Yao, JF Ton, X Zhang, RGH Cheng… - arXiv preprint arXiv …, 2023 - arxiv.org
Ensuring alignment, which refers to making models behave in accordance with human
intentions [1, 2], has become a critical task before deploying large language models (LLMs) …

Factuality enhanced language models for open-ended text generation

N Lee, W Ping, P Xu, M Patwary… - Advances in …, 2022 - proceedings.neurips.cc
Pretrained language models (LMs) are susceptible to generate text with nonfactual
information. In this work, we measure and improve the factual accuracy of large-scale LMs …

Hallucination is inevitable: An innate limitation of large language models

Z Xu, S Jain, M Kankanhalli - arXiv preprint arXiv:2401.11817, 2024 - arxiv.org
Hallucination has been widely recognized to be a significant drawback for large language
models (LLMs). There have been many works that attempt to reduce the extent of …

Automatic evaluation of attribution by large language models

X Yue, B Wang, Z Chen, K Zhang, Y Su… - arXiv preprint arXiv …, 2023 - arxiv.org
A recent focus of large language model (LLM) development, as exemplified by generative
search engines, is to incorporate external references to generate and support its claims …

A culturally sensitive test to evaluate nuanced gpt hallucination

TR McIntosh, T Liu, T Susnjak, P Watters… - IEEE Transactions …, 2023 - ieeexplore.ieee.org
The Generative Pre-trained Transformer (GPT) models, renowned for generating human-like
text, occasionally produce “hallucinations”-outputs that diverge from human expectations …

" kelly is a warm person, joseph is a role model": Gender biases in llm-generated reference letters

Y Wan, G Pu, J Sun, A Garimella, KW Chang… - arXiv preprint arXiv …, 2023 - arxiv.org
Large Language Models (LLMs) have recently emerged as an effective tool to assist
individuals in writing various types of content, including professional documents such as …