When Can LLMs Actually Correct Their Own Mistakes? A Critical Survey of Self-Correction of LLMs
Self-correction is an approach to improving responses from large language models (LLMs)
by refining the responses using LLMs during inference. Prior work has proposed various self …
by refining the responses using LLMs during inference. Prior work has proposed various self …
Natural language generation and understanding of big code for AI-assisted programming: A review
MF Wong, S Guo, CN Hang, SW Ho, CW Tan - Entropy, 2023 - mdpi.com
This paper provides a comprehensive review of the literature concerning the utilization of
Natural Language Processing (NLP) techniques, with a particular focus on transformer …
Natural Language Processing (NLP) techniques, with a particular focus on transformer …
Self-refine: Iterative refinement with self-feedback
Like humans, large language models (LLMs) do not always generate the best output on their
first try. Motivated by how humans refine their written text, we introduce Self-Refine, an …
first try. Motivated by how humans refine their written text, we introduce Self-Refine, an …
Reflexion: Language agents with verbal reinforcement learning
Large language models (LLMs) have been increasingly used to interact with external
environments (eg, games, compilers, APIs) as goal-driven agents. However, it remains …
environments (eg, games, compilers, APIs) as goal-driven agents. However, it remains …
Codet5+: Open code large language models for code understanding and generation
Large language models (LLMs) pretrained on vast source code have achieved prominent
progress in code intelligence. However, existing code LLMs have two main limitations in …
progress in code intelligence. However, existing code LLMs have two main limitations in …
Critic: Large language models can self-correct with tool-interactive critiquing
Recent developments in large language models (LLMs) have been impressive. However,
these models sometimes show inconsistencies and problematic behavior, such as …
these models sometimes show inconsistencies and problematic behavior, such as …
Trustworthy LLMs: A survey and guideline for evaluating large language models' alignment
Ensuring alignment, which refers to making models behave in accordance with human
intentions [1, 2], has become a critical task before deploying large language models (LLMs) …
intentions [1, 2], has become a critical task before deploying large language models (LLMs) …
Codet: Code generation with generated tests
The task of generating code solutions for a given programming problem can benefit from the
use of pre-trained language models such as Codex, which can produce multiple diverse …
use of pre-trained language models such as Codex, which can produce multiple diverse …
Cognitive architectures for language agents
Recent efforts have incorporated large language models (LLMs) with external resources (eg,
the Internet) or internal control flows (eg, prompt chaining) for tasks requiring grounding or …
the Internet) or internal control flows (eg, prompt chaining) for tasks requiring grounding or …
Large language models meet nl2code: A survey
The task of generating code from a natural language description, or NL2Code, is considered
a pressing and significant challenge in code intelligence. Thanks to the rapid development …
a pressing and significant challenge in code intelligence. Thanks to the rapid development …