Dissociating language and thought in large language models

K Mahowald, AA Ivanova, IA Blank, N Kanwisher… - Trends in Cognitive …, 2024 - cell.com
Large language models (LLMs) have come closest among all models to date to mastering
human language, yet opinions about their linguistic and cognitive capabilities remain split …

Winoground: Probing vision and language models for visio-linguistic compositionality

T Thrush, R Jiang, M Bartolo, A Singh… - Proceedings of the …, 2022 - openaccess.thecvf.com
We present a novel task and dataset for evaluating the ability of vision and language models
to conduct visio-linguistic compositional reasoning, which we call Winoground. Given two …

The defeat of the Winograd schema challenge

V Kocijan, E Davis, T Lukasiewicz, G Marcus… - Artificial Intelligence, 2023 - Elsevier
Abstract The Winograd Schema Challenge—a set of twin sentences involving pronoun
reference disambiguation that seem to require the use of commonsense knowledge—was …

On the paradox of learning to reason from data

H Zhang, LH Li, T Meng, KW Chang… - arXiv preprint arXiv …, 2022 - arxiv.org
Logical reasoning is needed in a wide range of NLP tasks. Can a BERT model be trained
end-to-end to solve logical reasoning problems presented in natural language? We attempt …

State-of-the-art generalisation research in NLP: a taxonomy and review

D Hupkes, M Giulianelli, V Dankers, M Artetxe… - arXiv preprint arXiv …, 2022 - arxiv.org
The ability to generalise well is one of the primary desiderata of natural language
processing (NLP). Yet, what'good generalisation'entails and how it should be evaluated is …

Event knowledge in large language models: the gap between the impossible and the unlikely

C Kauf, AA Ivanova, G Rambelli, E Chersoni… - Cognitive …, 2023 - Wiley Online Library
Word co‐occurrence patterns in language corpora contain a surprising amount of
conceptual knowledge. Large language models (LLMs), trained to predict words in context …

A systematic investigation of commonsense knowledge in large language models

XL Li, A Kuncoro, J Hoffmann, CM d'Autume… - arXiv preprint arXiv …, 2021 - arxiv.org
Language models (LMs) trained on large amounts of data have shown impressive
performance on many NLP tasks under the zero-shot and few-shot setup. Here we aim to …

Dimensions of commonsense knowledge

F Ilievski, A Oltramari, K Ma, B Zhang… - Knowledge-Based …, 2021 - Elsevier
Commonsense knowledge is essential for many AI applications, including those in natural
language processing, visual processing, and planning. Consequently, many sources that …

Measuring causal effects of data statistics on language model'sfactual'predictions

Y Elazar, N Kassner, S Ravfogel, A Feder… - arXiv preprint arXiv …, 2022 - arxiv.org
Large amounts of training data are one of the major reasons for the high performance of
state-of-the-art NLP models. But what exactly in the training data causes a model to make a …

ASER: Towards large-scale commonsense knowledge acquisition via higher-order selectional preference over eventualities

H Zhang, X Liu, H Pan, H Ke, J Ou, T Fang, Y Song - Artificial Intelligence, 2022 - Elsevier
Commonsense knowledge acquisition and reasoning have long been a core artificial
intelligence problem. However, in the past, there has been a lack of scalable methods to …