Lever: Learning to verify language-to-code generation with execution
The advent of large language models trained on code (code LLMs) has led to significant
progress in language-to-code generation. State-of-the-art approaches in this area combine …
progress in language-to-code generation. State-of-the-art approaches in this area combine …
On the importance and applicability of pre-training for federated learning
Pre-training is prevalent in nowadays deep learning to improve the learned model's
performance. However, in the literature on federated learning (FL), neural networks are …
performance. However, in the literature on federated learning (FL), neural networks are …
Large language models are versatile decomposers: Decomposing evidence and questions for table-based reasoning
Table-based reasoning has shown remarkable progress in a wide range of table-based
tasks. It is a challenging task, which requires reasoning over both free-form natural language …
tasks. It is a challenging task, which requires reasoning over both free-form natural language …
Retrieval-augmented generation for ai-generated content: A survey
The development of Artificial Intelligence Generated Content (AIGC) has been facilitated by
advancements in model algorithms, scalable foundation model architectures, and the …
advancements in model algorithms, scalable foundation model architectures, and the …
Gptuner: A manual-reading database tuning system via gpt-guided bayesian optimization
Modern database management systems (DBMS) expose hundreds of configurable knobs to
control system behaviours. Determining the appropriate values for these knobs to improve …
control system behaviours. Determining the appropriate values for these knobs to improve …
Enhancing few-shot text-to-sql capabilities of large language models: A study on prompt design strategies
In-context learning (ICL) has emerged as a new approach to various natural language
processing tasks, utilizing large language models (LLMs) to make predictions based on …
processing tasks, utilizing large language models (LLMs) to make predictions based on …
Chain-of-table: Evolving tables in the reasoning chain for table understanding
Table-based reasoning with large language models (LLMs) is a promising direction to tackle
many table understanding tasks, such as table-based question answering and fact …
many table understanding tasks, such as table-based question answering and fact …
Large language models are versatile decomposers: Decompose evidence and questions for table-based reasoning
Table-based reasoning has shown remarkable progress in combining deep models with
discrete reasoning, which requires reasoning over both free-form natural language (NL) …
discrete reasoning, which requires reasoning over both free-form natural language (NL) …
AutoTQA: Towards Autonomous Tabular Question Answering through Multi-Agent Large Language Models
With the growing significance of data analysis, several studies aim to provide precise
answers to users' natural language questions from tables, a task referred to as tabular …
answers to users' natural language questions from tables, a task referred to as tabular …
TaPERA: Enhancing faithfulness and interpretability in long-form table QA by content planning and execution-based reasoning
Abstract Long-form Table Question Answering (LFTQA) requires systems to generate
paragraph long and complex answers to questions over tabular data. While Large language …
paragraph long and complex answers to questions over tabular data. While Large language …