Dissociating language and thought in large language models
Large language models (LLMs) have come closest among all models to date to mastering
human language, yet opinions about their linguistic and cognitive capabilities remain split …
human language, yet opinions about their linguistic and cognitive capabilities remain split …
Embers of autoregression: Understanding large language models through the problem they are trained to solve
The widespread adoption of large language models (LLMs) makes it important to recognize
their strengths and limitations. We argue that in order to develop a holistic understanding of …
their strengths and limitations. We argue that in order to develop a holistic understanding of …
What formal languages can transformers express? a survey
As transformers have gained prominence in natural language processing, some researchers
have investigated theoretically what problems they can and cannot solve, by treating …
have investigated theoretically what problems they can and cannot solve, by treating …
Boardgameqa: A dataset for natural language reasoning with contradictory information
Automated reasoning with unstructured natural text is a key requirement for many potential
applications of NLP and for developing robust AI systems. Recently, Language Models …
applications of NLP and for developing robust AI systems. Recently, Language Models …
Are language models more like libraries or like librarians? Bibliotechnism, the novel reference problem, and the attitudes of LLMs
H Lederman, K Mahowald - Transactions of the Association for …, 2024 - direct.mit.edu
Are LLMs cultural technologies like photocopiers or printing presses, which transmit
information but cannot create new content? A challenge for this idea, which we call …
information but cannot create new content? A challenge for this idea, which we call …
Generative AI: Overview, economic impact, and applications in asset management
M Luk - Economic Impact, and Applications in Asset …, 2023 - papers.ssrn.com
This paper provides a comprehensive overview of the evolution and latest advancements in
Generative AI models, alongside their economic impact and applications in asset …
Generative AI models, alongside their economic impact and applications in asset …
The illusion of state in state-space models
State-space models (SSMs) have emerged as a potential alternative architecture for building
large language models (LLMs) compared to the previously ubiquitous transformer …
large language models (LLMs) compared to the previously ubiquitous transformer …
Fine-tuning enhances existing mechanisms: A case study on entity tracking
Fine-tuning on generalized tasks such as instruction following, code generation, and
mathematics has been shown to enhance language models' performance on a range of …
mathematics has been shown to enhance language models' performance on a range of …
In-context Learning Generalizes, But Not Always Robustly: The Case of Syntax
In-context learning (ICL) is now a common method for supervising large language models
(LLMs): given labeled examples in the input context, the LLM learns to perform the task …
(LLMs): given labeled examples in the input context, the LLM learns to perform the task …
Monitoring latent world states in language models with propositional probes
Language models are susceptible to bias, sycophancy, backdoors, and other tendencies
that lead to unfaithful responses to the input context. Interpreting internal states of language …
that lead to unfaithful responses to the input context. Interpreting internal states of language …