Dissociating language and thought in large language models

K Mahowald, AA Ivanova, IA Blank, N Kanwisher… - Trends in Cognitive …, 2024 - cell.com
Large language models (LLMs) have come closest among all models to date to mastering
human language, yet opinions about their linguistic and cognitive capabilities remain split …

Embers of autoregression: Understanding large language models through the problem they are trained to solve

RT McCoy, S Yao, D Friedman, M Hardy… - arXiv preprint arXiv …, 2023 - arxiv.org
The widespread adoption of large language models (LLMs) makes it important to recognize
their strengths and limitations. We argue that in order to develop a holistic understanding of …

What formal languages can transformers express? a survey

L Strobl, W Merrill, G Weiss, D Chiang… - Transactions of the …, 2024 - direct.mit.edu
As transformers have gained prominence in natural language processing, some researchers
have investigated theoretically what problems they can and cannot solve, by treating …

Boardgameqa: A dataset for natural language reasoning with contradictory information

M Kazemi, Q Yuan, D Bhatia, N Kim… - Advances in …, 2024 - proceedings.neurips.cc
Automated reasoning with unstructured natural text is a key requirement for many potential
applications of NLP and for developing robust AI systems. Recently, Language Models …

Are language models more like libraries or like librarians? Bibliotechnism, the novel reference problem, and the attitudes of LLMs

H Lederman, K Mahowald - Transactions of the Association for …, 2024 - direct.mit.edu
Are LLMs cultural technologies like photocopiers or printing presses, which transmit
information but cannot create new content? A challenge for this idea, which we call …

Generative AI: Overview, economic impact, and applications in asset management

M Luk - Economic Impact, and Applications in Asset …, 2023 - papers.ssrn.com
This paper provides a comprehensive overview of the evolution and latest advancements in
Generative AI models, alongside their economic impact and applications in asset …

The illusion of state in state-space models

W Merrill, J Petty, A Sabharwal - arXiv preprint arXiv:2404.08819, 2024 - arxiv.org
State-space models (SSMs) have emerged as a potential alternative architecture for building
large language models (LLMs) compared to the previously ubiquitous transformer …

Fine-tuning enhances existing mechanisms: A case study on entity tracking

N Prakash, TR Shaham, T Haklay, Y Belinkov… - arXiv preprint arXiv …, 2024 - arxiv.org
Fine-tuning on generalized tasks such as instruction following, code generation, and
mathematics has been shown to enhance language models' performance on a range of …

In-context Learning Generalizes, But Not Always Robustly: The Case of Syntax

A Mueller, A Webson, J Petty, T Linzen - arXiv preprint arXiv:2311.07811, 2023 - arxiv.org
In-context learning (ICL) is now a common method for supervising large language models
(LLMs): given labeled examples in the input context, the LLM learns to perform the task …

Monitoring latent world states in language models with propositional probes

J Feng, S Russell, J Steinhardt - arXiv preprint arXiv:2406.19501, 2024 - arxiv.org
Language models are susceptible to bias, sycophancy, backdoors, and other tendencies
that lead to unfaithful responses to the input context. Interpreting internal states of language …