“What it wants me to say”: Bridging the abstraction gap between end-user programmers and code-generating large language models

MX Liu, A Sarkar, C Negreanu, B Zorn… - Proceedings of the …, 2023 - dl.acm.org
Code-generating large language models map natural language to code. However, only a
small portion of the infinite space of naturalistic utterances is effective at guiding code …

Analyzing the performance of large language models on code summarization

R Haldar, J Hockenmaier - arXiv preprint arXiv:2404.08018, 2024 - arxiv.org
Large language models (LLMs) such as Llama 2 perform very well on tasks that involve both
natural language and source code, particularly code summarization and code generation …

Cat-probing: A metric-based approach to interpret how pre-trained models for programming language attend code structure

N Chen, Q Sun, R Zhu, X Li, X Lu, M Gao - arXiv preprint arXiv:2210.04633, 2022 - arxiv.org
Code pre-trained models (CodePTMs) have recently demonstrated significant success in
code intelligence. To interpret these models, some probing methods have been applied …

Analyzing declarative deployment code with large language models

G Lanciano, M Stein, V Hilt, T Cucinotta - CLOSER, 2023 - ricerca.sns.it
In the cloud-native era, developers have at their disposal an unprecedented landscape of
services to build scalable distributed systems. The DevOps paradigm emerged as a …

Learning Program Representations with a Tree-Structured Transformer

W Wang, K Zhang, G Li, S Liu, A Li… - … on Software Analysis …, 2023 - ieeexplore.ieee.org
Learning vector representations for programs is a critical step in applying deep learning
techniques for program understanding tasks. Various neural network models are proposed …

Prompt sensitivity of language model for solving programming problems

A Shirafuji, T Ito, M Morishita… - New Trends in …, 2022 - ebooks.iospress.nl
A popular language model that can solve introductory programming problems, OpenAI's
Codex, has drawn much attention not only in the natural language processing field but also …

An extensive study of the structure features in transformer-based code semantic summarization

K Yang, X Mao, S Wang, Y Qin, T Zhang… - 2023 IEEE/ACM 31st …, 2023 - ieeexplore.ieee.org
Transformers are now widely utilized in code intelligence tasks. To better fit highly structured
source code, various structure information is passed into Transformer, such as positional …

REPOEXEC: Evaluate Code Generation with a Repository-Level Executable Benchmark

NL Hai, DM Nguyen, NDQ Bui - arXiv preprint arXiv:2406.11927, 2024 - arxiv.org
The ability of CodeLLMs to generate executable and functionally correct code at the\textit
{repository-level scale} remains largely unexplored. We introduce\methodnamews, a novel …

Empirical Studies of Parameter Efficient Methods for Large Language Models of Code and Knowledge Transfer to R

A Esmaeili, I Saberi, FH Fard - arXiv preprint arXiv:2405.01553, 2024 - arxiv.org
Recently, Large Langauge Models (LLMs) have gained a lot of attention in the Software
Engineering (SE) community. LLMs or their variants pre-trained on code are used for many …

A Critical Study of What Code-LLMs (Do Not) Learn

A Anand, S Verma, K Narasimhan, M Mezini - arXiv preprint arXiv …, 2024 - arxiv.org
Large Language Models trained on code corpora (code-LLMs) have demonstrated
impressive performance in various coding assistance tasks. However, despite their …