Survey of hallucination in natural language generation

Z Ji, N Lee, R Frieske, T Yu, D Su, Y Xu, E Ishii… - ACM Computing …, 2023 - dl.acm.org
Natural Language Generation (NLG) has improved exponentially in recent years thanks to
the development of sequence-to-sequence deep learning technologies such as Transformer …

A survey of deep learning for mathematical reasoning

P Lu, L Qiu, W Yu, S Welleck, KW Chang - arXiv preprint arXiv:2212.10535, 2022 - arxiv.org
Mathematical reasoning is a fundamental aspect of human intelligence and is applicable in
various fields, including science, engineering, finance, and everyday life. The development …

Measuring and improving consistency in pretrained language models

Y Elazar, N Kassner, S Ravfogel… - Transactions of the …, 2021 - direct.mit.edu
Consistency of a model—that is, the invariance of its behavior under meaning-preserving
alternations in its input—is a highly desirable property in natural language processing. In …

Reasoning or reciting? exploring the capabilities and limitations of language models through counterfactual tasks

Z Wu, L Qiu, A Ross, E Akyürek, B Chen… - arXiv preprint arXiv …, 2023 - arxiv.org
The impressive performance of recent language models across a wide range of tasks
suggests that they possess a degree of abstract reasoning skills. Are these skills general …

Lift: Language-interfaced fine-tuning for non-language machine learning tasks

T Dinh, Y Zeng, R Zhang, Z Lin… - Advances in …, 2022 - proceedings.neurips.cc
Fine-tuning pretrained language models (LMs) without making any architectural changes
has become a norm for learning various language downstream tasks. However, for non …

mgpt: Few-shot learners go multilingual

O Shliazhko, A Fenogenova, M Tikhonova… - arXiv preprint arXiv …, 2022 - arxiv.org
Recent studies report that autoregressive language models can successfully solve many
NLP tasks via zero-and few-shot learning paradigms, which opens up new possibilities for …

Language models: past, present, and future

H Li - Communications of the ACM, 2022 - dl.acm.org
Language models: past, present, and future Page 1 56 COMMUNICATIONS OF THE ACM |
JULY 2022 | VOL. 65 | NO. 7 contributed articles NATURAL LANGUAGE PROCESSING (NLP) …

Representing numbers in NLP: a survey and a vision

A Thawani, J Pujara, PA Szekely, F Ilievski - arXiv preprint arXiv …, 2021 - arxiv.org
NLP systems rarely give special consideration to numbers found in text. This starkly
contrasts with the consensus in neuroscience that, in the brain, numbers are represented …

Geollm: Extracting geospatial knowledge from large language models

R Manvi, S Khanna, G Mai, M Burke, D Lobell… - arXiv preprint arXiv …, 2023 - arxiv.org
The application of machine learning (ML) in a range of geospatial tasks is increasingly
common but often relies on globally available covariates such as satellite imagery that can …

FiNER: Financial numeric entity recognition for XBRL tagging

L Loukas, M Fergadiotis, I Chalkidis… - arXiv preprint arXiv …, 2022 - arxiv.org
Publicly traded companies are required to submit periodic reports with eXtensive Business
Reporting Language (XBRL) word-level tags. Manually tagging the reports is tedious and …