Human-like systematic generalization through a meta-learning neural network

BM Lake, M Baroni - Nature, 2023 - nature.com
The power of human language and thought arises from systematic compositionality—the
algebraic ability to understand and produce novel combinations from known components …

Least-to-most prompting enables complex reasoning in large language models

D Zhou, N Schärli, L Hou, J Wei, N Scales… - arXiv preprint arXiv …, 2022 - arxiv.org
Chain-of-thought prompting has demonstrated remarkable performance on various natural
language reasoning tasks. However, it tends to perform poorly on tasks which requires …

Compositional semantic parsing with large language models

A Drozdov, N Schärli, E Akyürek, N Scales… - The Eleventh …, 2022 - openreview.net
Humans can reason compositionally when presented with new tasks. Previous research
shows that appropriate prompting techniques enable large language models (LLMs) to …

Randomized positional encodings boost length generalization of transformers

A Ruoss, G Delétang, T Genewein… - arXiv preprint arXiv …, 2023 - arxiv.org
Transformers have impressive generalization capabilities on tasks with a fixed context
length. However, they fail to generalize to sequences of arbitrary length, even for seemingly …

How to reuse and compose knowledge for a lifetime of tasks: A survey on continual learning and functional composition

JA Mendez, E Eaton - arXiv preprint arXiv:2207.07730, 2022 - arxiv.org
A major goal of artificial intelligence (AI) is to create an agent capable of acquiring a general
understanding of the world. Such an agent would require the ability to continually …

Compositional generalization and natural language variation: Can a semantic parsing approach handle both?

P Shaw, MW Chang, P Pasupat… - arXiv preprint arXiv …, 2020 - arxiv.org
Sequence-to-sequence models excel at handling natural language variation, but have been
shown to struggle with out-of-distribution compositional generalization. This has motivated …

The devil is in the detail: Simple tricks improve systematic generalization of transformers

R Csordás, K Irie, J Schmidhuber - arXiv preprint arXiv:2108.12284, 2021 - arxiv.org
Recently, many datasets have been proposed to test the systematic generalization ability of
neural networks. The companion baseline Transformers, typically trained with default hyper …

Improving compositional generalization with latent structure and data augmentation

L Qiu, P Shaw, P Pasupat, PK Nowak, T Linzen… - arXiv preprint arXiv …, 2021 - arxiv.org
Generic unstructured neural networks have been shown to struggle on out-of-distribution
compositional generalization. Compositional data augmentation via example recombination …

Span-based semantic parsing for compositional generalization

J Herzig, J Berant - arXiv preprint arXiv:2009.06040, 2020 - arxiv.org
Despite the success of sequence-to-sequence (seq2seq) models in semantic parsing,
recent work has shown that they fail in compositional generalization, ie, the ability to …

Making transformers solve compositional tasks

S Ontanon, J Ainslie, V Cvicek, Z Fisher - arXiv preprint arXiv:2108.04378, 2021 - arxiv.org
Several studies have reported the inability of Transformer models to generalize
compositionally, a key type of generalization in many NLP tasks such as semantic parsing …