Unleashing the potential of prompt engineering in Large Language Models: a comprehensive review

B Chen, Z Zhang, N Langrené, S Zhu - arXiv preprint arXiv:2310.14735, 2023 - arxiv.org
This paper delves into the pivotal role of prompt engineering in unleashing the capabilities
of Large Language Models (LLMs). Prompt engineering is the process of structuring input …

Scaling relationship on learning mathematical reasoning with large language models

Z Yuan, H Yuan, C Li, G Dong, K Lu, C Tan… - arXiv preprint arXiv …, 2023 - arxiv.org
Mathematical reasoning is a challenging task for large language models (LLMs), while the
scaling relationship of it with respect to LLM capacity is under-explored. In this paper, we …

Wizardmath: Empowering mathematical reasoning for large language models via reinforced evol-instruct

H Luo, Q Sun, C Xu, P Zhao, J Lou, C Tao… - arXiv preprint arXiv …, 2023 - arxiv.org
Large language models (LLMs), such as GPT-4, have shown remarkable performance in
natural language processing (NLP) tasks, including challenging mathematical reasoning …

Math-shepherd: Verify and reinforce llms step-by-step without human annotations

P Wang, L Li, Z Shao, R Xu, D Dai, Y Li… - Proceedings of the …, 2024 - aclanthology.org
In this paper, we present an innovative process-oriented math process reward model called
Math-shepherd, which assigns a reward score to each step of math problem solutions. The …

Tora: A tool-integrated reasoning agent for mathematical problem solving

Z Gou, Z Shao, Y Gong, Y Shen, Y Yang… - arXiv preprint arXiv …, 2023 - arxiv.org
Large language models have made significant progress in various language tasks, yet they
still struggle with complex mathematics. In this paper, we propose ToRA a series of Tool …

Active prompting with chain-of-thought for large language models

S Diao, P Wang, Y Lin, R Pan, X Liu… - arXiv preprint arXiv …, 2023 - arxiv.org
The increasing scale of large language models (LLMs) brings emergent abilities to various
complex tasks requiring reasoning, such as arithmetic and commonsense reasoning. It is …

Better zero-shot reasoning with role-play prompting

A Kong, S Zhao, H Chen, Q Li, Y Qin, R Sun… - arXiv preprint arXiv …, 2023 - arxiv.org
Modern large language models (LLMs), such as ChatGPT, exhibit a remarkable capacity for
role-playing, enabling them to embody not only human characters but also non-human …

Cumulative reasoning with large language models

Y Zhang, J Yang, Y Yuan, ACC Yao - arXiv preprint arXiv:2308.04371, 2023 - arxiv.org
While language models are powerful and versatile, they often fail to address highly complex
problems. This is because solving complex problems requires deliberate thinking, which has …

Solving challenging math word problems using gpt-4 code interpreter with code-based self-verification

A Zhou, K Wang, Z Lu, W Shi, S Luo, Z Qin, S Lu… - arXiv preprint arXiv …, 2023 - arxiv.org
Recent progress in large language models (LLMs) like GPT-4 and PaLM-2 has brought
significant advancements in addressing math reasoning problems. In particular, OpenAI's …

Mathematical language models: A survey

W Liu, H Hu, J Zhou, Y Ding, J Li, J Zeng, M He… - arXiv preprint arXiv …, 2023 - arxiv.org
In recent years, there has been remarkable progress in leveraging Language Models (LMs),
encompassing Pre-trained Language Models (PLMs) and Large-scale Language Models …