Symbols and grounding in large language models
E Pavlick - … Transactions of the Royal Society A, 2023 - royalsocietypublishing.org
Large language models (LLMs) are one of the most impressive achievements of artificial
intelligence in recent years. However, their relevance to the study of language more broadly …
intelligence in recent years. However, their relevance to the study of language more broadly …
Semantic structure in deep learning
E Pavlick - Annual Review of Linguistics, 2022 - annualreviews.org
Deep learning has recently come to dominate computational linguistics, leading to claims of
human-level performance in a range of language processing tasks. Like much previous …
human-level performance in a range of language processing tasks. Like much previous …
[HTML][HTML] Pre-trained models: Past, present and future
Large-scale pre-trained models (PTMs) such as BERT and GPT have recently achieved
great success and become a milestone in the field of artificial intelligence (AI). Owing to …
great success and become a milestone in the field of artificial intelligence (AI). Owing to …
A primer in BERTology: What we know about how BERT works
A Rogers, O Kovaleva, A Rumshisky - Transactions of the Association …, 2021 - direct.mit.edu
Transformer-based models have pushed state of the art in many areas of NLP, but our
understanding of what is behind their success is still limited. This paper is the first survey of …
understanding of what is behind their success is still limited. This paper is the first survey of …
Measuring and improving consistency in pretrained language models
Consistency of a model—that is, the invariance of its behavior under meaning-preserving
alternations in its input—is a highly desirable property in natural language processing. In …
alternations in its input—is a highly desirable property in natural language processing. In …
On the systematicity of probing contextualized word representations: The case of hypernymy in BERT
Contextualized word representations have become a driving force in NLP, motivating
widespread interest in understanding their capabilities and the mechanisms by which they …
widespread interest in understanding their capabilities and the mechanisms by which they …
Combining pre-trained language models and structured knowledge
P Colon-Hernandez, C Havasi, J Alonso… - arXiv preprint arXiv …, 2021 - arxiv.org
In recent years, transformer-based language models have achieved state of the art
performance in various NLP benchmarks. These models are able to extract mostly …
performance in various NLP benchmarks. These models are able to extract mostly …
COMPS: Conceptual minimal pair sentences for testing robust property knowledge and its inheritance in pre-trained language models
A characteristic feature of human semantic cognition is its ability to not only store and
retrieve the properties of concepts observed through experience, but to also facilitate the …
retrieve the properties of concepts observed through experience, but to also facilitate the …
The world of an octopus: How reporting bias influences a language model's perception of color
C Paik, S Aroca-Ouellette, A Roncone… - arXiv preprint arXiv …, 2021 - arxiv.org
Recent work has raised concerns about the inherent limitations of text-only pretraining. In
this paper, we first demonstrate that reporting bias, the tendency of people to not state the …
this paper, we first demonstrate that reporting bias, the tendency of people to not state the …
Acquiring and modeling abstract commonsense knowledge via conceptualization
Conceptualization, or viewing entities and situations as instances of abstract concepts in
mind and making inferences based on that, is a vital component in human intelligence for …
mind and making inferences based on that, is a vital component in human intelligence for …