Natural language reasoning, a survey

F Yu, H Zhang, P Tiwari, B Wang - ACM Computing Surveys, 2023 - dl.acm.org
This survey paper proposes a clearer view of natural language reasoning in the field of
Natural Language Processing (NLP), both conceptually and practically. Conceptually, we …

Augmented language models: a survey

G Mialon, R Dessì, M Lomeli, C Nalmpantis… - arXiv preprint arXiv …, 2023 - arxiv.org
This survey reviews works in which language models (LMs) are augmented with reasoning
skills and the ability to use tools. The former is defined as decomposing a potentially …

Measuring and narrowing the compositionality gap in language models

O Press, M Zhang, S Min, L Schmidt, NA Smith… - arXiv preprint arXiv …, 2022 - arxiv.org
We investigate the ability of language models to perform compositional reasoning tasks
where the overall solution depends on correctly composing the answers to sub-problems …

Toolqa: A dataset for llm question answering with external tools

Y Zhuang, Y Yu, K Wang, H Sun… - Advances in Neural …, 2024 - proceedings.neurips.cc
Abstract Large Language Models (LLMs) have demonstrated impressive performance in
various NLP tasks, but they still suffer from challenges such as hallucination and weak …

“What it wants me to say”: Bridging the abstraction gap between end-user programmers and code-generating large language models

MX Liu, A Sarkar, C Negreanu, B Zorn… - Proceedings of the …, 2023 - dl.acm.org
Code-generating large language models map natural language to code. However, only a
small portion of the infinite space of naturalistic utterances is effective at guiding code …

Take a step back: Evoking reasoning via abstraction in large language models

HS Zheng, S Mishra, X Chen, HT Cheng… - arXiv preprint arXiv …, 2023 - arxiv.org
We present Step-Back Prompting, a simple prompting technique that enables LLMs to do
abstractions to derive high-level concepts and first principles from instances containing …

Self-discover: Large language models self-compose reasoning structures

P Zhou, J Pujara, X Ren, X Chen, HT Cheng… - arXiv preprint arXiv …, 2024 - arxiv.org
We introduce SELF-DISCOVER, a general framework for LLMs to self-discover the task-
intrinsic reasoning structures to tackle complex reasoning problems that are challenging for …

Exploring question decomposition for zero-shot VQA

Z Khan, VK BG, S Schulter… - Advances in Neural …, 2024 - proceedings.neurips.cc
Visual question answering (VQA) has traditionally been treated as a single-step task where
each question receives the same amount of effort, unlike natural human question-answering …

An LLM compiler for parallel function calling

S Kim, S Moon, R Tabrizi, N Lee, MW Mahoney… - arXiv preprint arXiv …, 2023 - arxiv.org
Large Language Models (LLMs) have shown remarkable results on various complex
reasoning benchmarks. The reasoning capabilities of LLMs enable them to execute function …

Instruction tuned models are quick learners

H Gupta, SA Sawant, S Mishra, M Nakamura… - arXiv preprint arXiv …, 2023 - arxiv.org
Instruction tuning of language models has demonstrated the ability to enhance model
generalization to unseen tasks via in-context learning using a few examples. However …