A Survey of Confidence Estimation and Calibration in Large Language Models
Large language models (LLMs) have demonstrated remarkable capabilities across a wide
range of tasks in various domains. Despite their impressive performance, they can be …
range of tasks in various domains. Despite their impressive performance, they can be …
Factcheck-bench: Fine-grained evaluation benchmark for automatic fact-checkers
The increased use of large language models (LLMs) across a variety of real-world
applications calls for mechanisms to verify the factual accuracy of their outputs. In this work …
applications calls for mechanisms to verify the factual accuracy of their outputs. In this work …
Adapting fake news detection to the era of large language models
In the age of large language models (LLMs) and the widespread adoption of AI-driven
content creation, the landscape of information dissemination has witnessed a paradigm shift …
content creation, the landscape of information dissemination has witnessed a paradigm shift …
Large language models as reliable knowledge bases?
The NLP community has recently shown a growing interest in leveraging Large Language
Models (LLMs) for knowledge-intensive tasks, viewing LLMs as potential knowledge bases …
Models (LLMs) for knowledge-intensive tasks, viewing LLMs as potential knowledge bases …
Generative large language models in automated fact-checking: A survey
The dissemination of false information on online platforms presents a serious societal
challenge. While manual fact-checking remains crucial, Large Language Models (LLMs) …
challenge. While manual fact-checking remains crucial, Large Language Models (LLMs) …
" Sorry, Come Again?" Prompting--Enhancing Comprehension and Diminishing Hallucination with [PAUSE]-injected Optimal Paraphrasing
Hallucination has emerged as the most vulnerable aspect of contemporary Large Language
Models (LLMs). In this paper, we introduce the Sorry, Come Again (SCA) prompting, aimed …
Models (LLMs). In this paper, we introduce the Sorry, Come Again (SCA) prompting, aimed …
Medico: Towards Hallucination Detection and Correction with Multi-source Evidence Fusion
As we all know, hallucinations prevail in Large Language Models (LLMs), where the
generated content is coherent but factually incorrect, which inflicts a heavy blow on the …
generated content is coherent but factually incorrect, which inflicts a heavy blow on the …
OpenFactCheck: A Unified Framework for Factuality Evaluation of LLMs
The increased use of large language models (LLMs) across a variety of real-world
applications calls for mechanisms to verify the factual accuracy of their outputs. Difficulties lie …
applications calls for mechanisms to verify the factual accuracy of their outputs. Difficulties lie …
OpenFactCheck: A Unified Framework for Factuality Evaluation of LLMs
The increased use of large language models (LLMs) across a variety of real-world
applications calls for automatic tools to check the factual accuracy of their outputs, as LLMs …
applications calls for automatic tools to check the factual accuracy of their outputs, as LLMs …
Loki: An Open-Source Tool for Fact Verification
We introduce Loki, an open-source tool designed to address the growing problem of
misinformation. Loki adopts a human-centered approach, striking a balance between the …
misinformation. Loki adopts a human-centered approach, striking a balance between the …