A comprehensive overview of large language models
Large Language Models (LLMs) have recently demonstrated remarkable capabilities in
natural language processing tasks and beyond. This success of LLMs has led to a large …
natural language processing tasks and beyond. This success of LLMs has led to a large …
A survey on data selection for language models
A major factor in the recent success of large language models is the use of enormous and
ever-growing text datasets for unsupervised pre-training. However, naively training a model …
ever-growing text datasets for unsupervised pre-training. However, naively training a model …
Metamath: Bootstrap your own mathematical questions for large language models
Large language models (LLMs) have pushed the limits of natural language understanding
and exhibited excellent problem-solving ability. Despite the great success, most existing …
and exhibited excellent problem-solving ability. Despite the great success, most existing …
Wizardmath: Empowering mathematical reasoning for large language models via reinforced evol-instruct
Large language models (LLMs), such as GPT-4, have shown remarkable performance in
natural language processing (NLP) tasks, including challenging mathematical reasoning …
natural language processing (NLP) tasks, including challenging mathematical reasoning …
Self-play fine-tuning converts weak language models to strong language models
Harnessing the power of human-annotated data through Supervised Fine-Tuning (SFT) is
pivotal for advancing Large Language Models (LLMs). In this paper, we delve into the …
pivotal for advancing Large Language Models (LLMs). In this paper, we delve into the …
Math-shepherd: Verify and reinforce llms step-by-step without human annotations
In this paper, we present an innovative process-oriented math process reward model called
Math-shepherd, which assigns a reward score to each step of math problem solutions. The …
Math-shepherd, which assigns a reward score to each step of math problem solutions. The …
Tora: A tool-integrated reasoning agent for mathematical problem solving
Large language models have made significant progress in various language tasks, yet they
still struggle with complex mathematics. In this paper, we propose ToRA a series of Tool …
still struggle with complex mathematics. In this paper, we propose ToRA a series of Tool …
Mammoth: Building math generalist models through hybrid instruction tuning
We introduce MAmmoTH, a series of open-source large language models (LLMs)
specifically tailored for general math problem-solving. The MAmmoTH models are trained on …
specifically tailored for general math problem-solving. The MAmmoTH models are trained on …
Alphazero-like tree-search can guide large language model decoding and training
Recent works like Tree-of-Thought (ToT) and Reasoning via Planning (RAP) aim to augment
the reasoning capabilities of LLMs by using tree-search algorithms to guide multi-step …
the reasoning capabilities of LLMs by using tree-search algorithms to guide multi-step …
Instructerc: Reforming emotion recognition in conversation with a retrieval multi-task llms framework
The development of emotion recognition in dialogue (ERC) has been consistently hindered
by the complexity of pipeline designs, leading to ERC models that often overfit to specific …
by the complexity of pipeline designs, leading to ERC models that often overfit to specific …