Language model behavior: A comprehensive survey

TA Chang, BK Bergen - Computational Linguistics, 2024 - direct.mit.edu
Transformer language models have received widespread public attention, yet their
generated text is often surprising even to NLP researchers. In this survey, we discuss over …

Revisiting out-of-distribution robustness in nlp: Benchmarks, analysis, and LLMs evaluations

L Yuan, Y Chen, G Cui, H Gao, F Zou… - Advances in …, 2023 - proceedings.neurips.cc
This paper reexamines the research on out-of-distribution (OOD) robustness in the field of
NLP. We find that the distribution shift settings in previous studies commonly lack adequate …

Embers of autoregression: Understanding large language models through the problem they are trained to solve

RT McCoy, S Yao, D Friedman, M Hardy… - arXiv preprint arXiv …, 2023 - arxiv.org
The widespread adoption of large language models (LLMs) makes it important to recognize
their strengths and limitations. We argue that in order to develop a holistic understanding of …

Glue-x: Evaluating natural language understanding models from an out-of-distribution generalization perspective

L Yang, S Zhang, L Qin, Y Li, Y Wang, H Liu… - arXiv preprint arXiv …, 2022 - arxiv.org
Pre-trained language models (PLMs) are known to improve the generalization performance
of natural language understanding models by leveraging large amounts of data during the …

Uncertainty in natural language generation: From theory to applications

J Baan, N Daheim, E Ilia, D Ulmer, HS Li… - arXiv preprint arXiv …, 2023 - arxiv.org
Recent advances of powerful Language Models have allowed Natural Language
Generation (NLG) to emerge as an important technology that can not only perform traditional …

Cross-lingual consistency of factual knowledge in multilingual language models

J Qi, R Fernández, A Bisazza - arXiv preprint arXiv:2310.10378, 2023 - arxiv.org
Multilingual large-scale Pretrained Language Models (PLMs) have been shown to store
considerable amounts of factual knowledge, but large variations are observed across …

Mind the instructions: a holistic evaluation of consistency and interactions in prompt-based learning

L Weber, E Bruni, D Hupkes - arXiv preprint arXiv:2310.13486, 2023 - arxiv.org
Finding the best way of adapting pre-trained language models to a task is a big challenge in
current NLP. Just like the previous generation of task-tuned models (TT), models that are …

Characterizing mechanisms for factual recall in language models

Q Yu, J Merullo, E Pavlick - arXiv preprint arXiv:2310.15910, 2023 - arxiv.org
Language Models (LMs) often must integrate facts they memorized in pretraining with new
information that appears in a given context. These two sources can disagree, causing …

Out-of-distribution generalization in natural language processing: Past, present, and future

L Yang, Y Song, X Ren, C Lyu, Y Wang… - Proceedings of the …, 2023 - aclanthology.org
Abstract Machine learning (ML) systems in natural language processing (NLP) face
significant challenges in generalizing to out-of-distribution (OOD) data, where the test …

Kg-gpt: A general framework for reasoning on knowledge graphs using large language models

J Kim, Y Kwon, Y Jo, E Choi - arXiv preprint arXiv:2310.11220, 2023 - arxiv.org
While large language models (LLMs) have made considerable advancements in
understanding and generating unstructured text, their application in structured data remains …