Survey of hallucination in natural language generation

Z Ji, N Lee, R Frieske, T Yu, D Su, Y Xu, E Ishii… - ACM Computing …, 2023 - dl.acm.org
Natural Language Generation (NLG) has improved exponentially in recent years thanks to
the development of sequence-to-sequence deep learning technologies such as Transformer …

A survey of evaluation metrics used for NLG systems

AB Sai, AK Mohankumar, MM Khapra - ACM Computing Surveys (CSUR …, 2022 - dl.acm.org
In the last few years, a large number of automatic evaluation metrics have been proposed for
evaluating Natural Language Generation (NLG) systems. The rapid development and …

Siren's song in the AI ocean: a survey on hallucination in large language models

Y Zhang, Y Li, L Cui, D Cai, L Liu, T Fu… - arXiv preprint arXiv …, 2023 - arxiv.org
While large language models (LLMs) have demonstrated remarkable capabilities across a
range of downstream tasks, a significant concern revolves around their propensity to exhibit …

Palm: Scaling language modeling with pathways

A Chowdhery, S Narang, J Devlin, M Bosma… - Journal of Machine …, 2023 - jmlr.org
Large language models have been shown to achieve remarkable performance across a
variety of natural language tasks using few-shot learning, which drastically reduces the …

Graph neural networks for natural language processing: A survey

L Wu, Y Chen, K Shen, X Guo, H Gao… - … and Trends® in …, 2023 - nowpublishers.com
Deep learning has become the dominant approach in addressing various tasks in Natural
Language Processing (NLP). Although text inputs are typically represented as a sequence …

Spot: Better frozen model adaptation through soft prompt transfer

T Vu, B Lester, N Constant, R Al-Rfou, D Cer - arXiv preprint arXiv …, 2021 - arxiv.org
There has been growing interest in parameter-efficient methods to apply pre-trained
language models to downstream tasks. Building on the Prompt Tuning approach of Lester et …

Ext5: Towards extreme multi-task scaling for transfer learning

V Aribandi, Y Tay, T Schuster, J Rao, HS Zheng… - arXiv preprint arXiv …, 2021 - arxiv.org
Despite the recent success of multi-task learning and transfer learning for natural language
processing (NLP), few works have systematically studied the effect of scaling up the number …

TPLinker: Single-stage joint extraction of entities and relations through token pair linking

Y Wang, B Yu, Y Zhang, T Liu, H Zhu, L Sun - arXiv preprint arXiv …, 2020 - arxiv.org
Extracting entities and relations from unstructured text has attracted increasing attention in
recent years but remains challenging, due to the intrinsic difficulty in identifying overlapping …

A novel cascade binary tagging framework for relational triple extraction

Z Wei, J Su, Y Wang, Y Tian, Y Chang - arXiv preprint arXiv:1909.03227, 2019 - arxiv.org
Extracting relational triples from unstructured text is crucial for large-scale knowledge graph
construction. However, few existing works excel in solving the overlapping triple problem …

Onerel: Joint entity and relation extraction with one module in one step

YM Shang, H Huang, X Mao - Proceedings of the AAAI conference on …, 2022 - ojs.aaai.org
Joint entity and relation extraction is an essential task in natural language processing and
knowledge graph construction. Existing approaches usually decompose the joint extraction …