Qa dataset explosion: A taxonomy of nlp resources for question answering and reading comprehension
Alongside huge volumes of research on deep learning models in NLP in the recent years,
there has been much work on benchmark datasets needed to track modeling progress …
there has been much work on benchmark datasets needed to track modeling progress …
A survey of deep learning for mathematical reasoning
Mathematical reasoning is a fundamental aspect of human intelligence and is applicable in
various fields, including science, engineering, finance, and everyday life. The development …
various fields, including science, engineering, finance, and everyday life. The development …
Augmented language models: a survey
This survey reviews works in which language models (LMs) are augmented with reasoning
skills and the ability to use tools. The former is defined as decomposing a potentially …
skills and the ability to use tools. The former is defined as decomposing a potentially …
Solving quantitative reasoning problems with language models
A Lewkowycz, A Andreassen… - Advances in …, 2022 - proceedings.neurips.cc
Abstract Language models have achieved remarkable performance on a wide range of
tasks that require natural language understanding. Nevertheless, state-of-the-art models …
tasks that require natural language understanding. Nevertheless, state-of-the-art models …
Reasoning with language model prompting: A survey
Reasoning, as an essential ability for complex problem-solving, can provide back-end
support for various real-world applications, such as medical diagnosis, negotiation, etc. This …
support for various real-world applications, such as medical diagnosis, negotiation, etc. This …
Cross-task generalization via natural language crowdsourcing instructions
Humans (eg, crowdworkers) have a remarkable ability in solving different tasks, by simply
reading textual instructions that define them and looking at a few examples. Despite the …
reading textual instructions that define them and looking at a few examples. Despite the …
Cheap and quick: Efficient vision-language instruction tuning for large language models
Recently, growing interest has been aroused in extending the multimodal capability of large
language models (LLMs), eg, vision-language (VL) learning, which is regarded as the next …
language models (LLMs), eg, vision-language (VL) learning, which is regarded as the next …
How does GPT-2 compute greater-than?: Interpreting mathematical abilities in a pre-trained language model
Pre-trained language models can be surprisingly adept at tasks they were not explicitly
trained on, but how they implement these capabilities is poorly understood. In this paper, we …
trained on, but how they implement these capabilities is poorly understood. In this paper, we …
Mammoth: Building math generalist models through hybrid instruction tuning
We introduce MAmmoTH, a series of open-source large language models (LLMs)
specifically tailored for general math problem-solving. The MAmmoTH models are trained on …
specifically tailored for general math problem-solving. The MAmmoTH models are trained on …
Folio: Natural language reasoning with first-order logic
Large language models (LLMs) have achieved remarkable performance on a variety of
natural language understanding tasks. However, existing benchmarks are inadequate in …
natural language understanding tasks. However, existing benchmarks are inadequate in …