Large language models for software engineering: A systematic literature review

X Hou, Y Zhao, Y Liu, Z Yang, K Wang, L Li… - ACM Transactions on …, 2023 - dl.acm.org
Large Language Models (LLMs) have significantly impacted numerous domains, including
Software Engineering (SE). Many recent publications have explored LLMs applied to …

A survey on rag meeting llms: Towards retrieval-augmented large language models

W Fan, Y Ding, L Ning, S Wang, H Li, D Yin… - Proceedings of the 30th …, 2024 - dl.acm.org
As one of the most advanced techniques in AI, Retrieval-Augmented Generation (RAG) can
offer reliable and up-to-date external knowledge, providing huge convenience for numerous …

Repocoder: Repository-level code completion through iterative retrieval and generation

F Zhang, B Chen, Y Zhang, J Keung, J Liu… - arXiv preprint arXiv …, 2023 - arxiv.org
The task of repository-level code completion is to continue writing the unfinished code based
on a broader context of the repository. While for automated code completion tools, it is …

Large language models meet nl2code: A survey

D Zan, B Chen, F Zhang, D Lu, B Wu, B Guan… - arXiv preprint arXiv …, 2022 - arxiv.org
The task of generating code from a natural language description, or NL2Code, is considered
a pressing and significant challenge in code intelligence. Thanks to the rapid development …

Pangu-coder2: Boosting large language models for code with ranking feedback

B Shen, J Zhang, T Chen, D Zan, B Geng, A Fu… - arXiv preprint arXiv …, 2023 - arxiv.org
Large Language Models for Code (Code LLM) are flourishing. New and powerful models
are released on a weekly basis, demonstrating remarkable performance on the code …

Deep learning based code generation methods: A literature review

Z Yang, S Chen, C Gao, Z Li, G Li, R Lv - arXiv preprint arXiv:2303.01056, 2023 - arxiv.org
Code Generation aims at generating relevant code fragments according to given natural
language descriptions. In the process of software development, there exist a large number of …

Monitor-guided decoding of code LMs with static analysis of repository context

LA Agrawal, A Kanade, N Goyal… - Advances in Neural …, 2024 - proceedings.neurips.cc
Abstract Language models of code (LMs) work well when the surrounding code provides
sufficient context. This is not true when it becomes necessary to use types, functionality or …

Retrieval-augmented generation for ai-generated content: A survey

P Zhao, H Zhang, Q Yu, Z Wang, Y Geng, F Fu… - arXiv preprint arXiv …, 2024 - arxiv.org
The development of Artificial Intelligence Generated Content (AIGC) has been facilitated by
advancements in model algorithms, scalable foundation model architectures, and the …

Toolcoder: Teach code generation models to use api search tools

K Zhang, H Zhang, G Li, J Li, Z Li, Z Jin - arXiv preprint arXiv:2305.04032, 2023 - arxiv.org
Automatically generating source code from natural language descriptions has been a
growing field of research in recent years. However, current large-scale code generation …

Bamboo: A comprehensive benchmark for evaluating long text modeling capacities of large language models

Z Dong, T Tang, J Li, WX Zhao, JR Wen - arXiv preprint arXiv:2309.13345, 2023 - arxiv.org
Large language models (LLMs) have achieved dramatic proficiency over NLP tasks with
normal length. Recently, multiple studies have committed to extending the context length …