Language model behavior: A comprehensive survey
Transformer language models have received widespread public attention, yet their
generated text is often surprising even to NLP researchers. In this survey, we discuss over …
generated text is often surprising even to NLP researchers. In this survey, we discuss over …
Revisiting out-of-distribution robustness in nlp: Benchmarks, analysis, and LLMs evaluations
This paper reexamines the research on out-of-distribution (OOD) robustness in the field of
NLP. We find that the distribution shift settings in previous studies commonly lack adequate …
NLP. We find that the distribution shift settings in previous studies commonly lack adequate …
Embers of autoregression: Understanding large language models through the problem they are trained to solve
The widespread adoption of large language models (LLMs) makes it important to recognize
their strengths and limitations. We argue that in order to develop a holistic understanding of …
their strengths and limitations. We argue that in order to develop a holistic understanding of …
Glue-x: Evaluating natural language understanding models from an out-of-distribution generalization perspective
Pre-trained language models (PLMs) are known to improve the generalization performance
of natural language understanding models by leveraging large amounts of data during the …
of natural language understanding models by leveraging large amounts of data during the …
Uncertainty in natural language generation: From theory to applications
Recent advances of powerful Language Models have allowed Natural Language
Generation (NLG) to emerge as an important technology that can not only perform traditional …
Generation (NLG) to emerge as an important technology that can not only perform traditional …
Cross-lingual consistency of factual knowledge in multilingual language models
Multilingual large-scale Pretrained Language Models (PLMs) have been shown to store
considerable amounts of factual knowledge, but large variations are observed across …
considerable amounts of factual knowledge, but large variations are observed across …
Mind the instructions: a holistic evaluation of consistency and interactions in prompt-based learning
Finding the best way of adapting pre-trained language models to a task is a big challenge in
current NLP. Just like the previous generation of task-tuned models (TT), models that are …
current NLP. Just like the previous generation of task-tuned models (TT), models that are …
Characterizing mechanisms for factual recall in language models
Language Models (LMs) often must integrate facts they memorized in pretraining with new
information that appears in a given context. These two sources can disagree, causing …
information that appears in a given context. These two sources can disagree, causing …
Out-of-distribution generalization in natural language processing: Past, present, and future
Abstract Machine learning (ML) systems in natural language processing (NLP) face
significant challenges in generalizing to out-of-distribution (OOD) data, where the test …
significant challenges in generalizing to out-of-distribution (OOD) data, where the test …
Kg-gpt: A general framework for reasoning on knowledge graphs using large language models
While large language models (LLMs) have made considerable advancements in
understanding and generating unstructured text, their application in structured data remains …
understanding and generating unstructured text, their application in structured data remains …