Large language models for data annotation: A survey
Data annotation generally refers to the labeling or generating of raw data with relevant
information, which could be used for improving the efficacy of machine learning models. The …
information, which could be used for improving the efficacy of machine learning models. The …
The llama 3 herd of models
Modern artificial intelligence (AI) systems are powered by foundation models. This paper
presents a new set of foundation models, called Llama 3. It is a herd of language models …
presents a new set of foundation models, called Llama 3. It is a herd of language models …
Large language models for mathematical reasoning: Progresses and challenges
Mathematical reasoning serves as a cornerstone for assessing the fundamental cognitive
capabilities of human intelligence. In recent times, there has been a notable surge in the …
capabilities of human intelligence. In recent times, there has been a notable surge in the …
Personal llm agents: Insights and survey about the capability, efficiency and security
Since the advent of personal computing devices, intelligent personal assistants (IPAs) have
been one of the key technologies that researchers and engineers have focused on, aiming …
been one of the key technologies that researchers and engineers have focused on, aiming …
Understanding the planning of LLM agents: A survey
As Large Language Models (LLMs) have shown significant intelligence, the progress to
leverage LLMs as planning modules of autonomous agents has attracted more attention …
leverage LLMs as planning modules of autonomous agents has attracted more attention …
Internal consistency and self-feedback in large language models: A survey
Large language models (LLMs) often exhibit deficient reasoning or generate hallucinations.
To address these, studies prefixed with" Self-" such as Self-Consistency, Self-Improve, and …
To address these, studies prefixed with" Self-" such as Self-Consistency, Self-Improve, and …
Smaug: Fixing failure modes of preference optimisation with dpo-positive
Direct Preference Optimisation (DPO) is effective at significantly improving the performance
of large language models (LLMs) on downstream tasks such as reasoning, summarisation …
of large language models (LLMs) on downstream tasks such as reasoning, summarisation …
Common 7b language models already possess strong math capabilities
Mathematical capabilities were previously believed to emerge in common language models
only at a very large scale or require extensive math-related pre-training. This paper shows …
only at a very large scale or require extensive math-related pre-training. This paper shows …
Think twice before assure: Confidence estimation for large language models through reflection on multiple answers
Confidence estimation aiming to evaluate output trustability is crucial for the application of
large language models (LLM), especially the black-box ones. Existing confidence estimation …
large language models (LLM), especially the black-box ones. Existing confidence estimation …
V-star: Training verifiers for self-taught reasoners
Common self-improvement approaches for large language models (LLMs), such as STaR
(Zelikman et al., 2022), iteratively fine-tune LLMs on self-generated solutions to improve …
(Zelikman et al., 2022), iteratively fine-tune LLMs on self-generated solutions to improve …