“What it wants me to say”: Bridging the abstraction gap between end-user programmers and code-generating large language models
Code-generating large language models map natural language to code. However, only a
small portion of the infinite space of naturalistic utterances is effective at guiding code …
small portion of the infinite space of naturalistic utterances is effective at guiding code …
Analyzing the performance of large language models on code summarization
R Haldar, J Hockenmaier - arXiv preprint arXiv:2404.08018, 2024 - arxiv.org
Large language models (LLMs) such as Llama 2 perform very well on tasks that involve both
natural language and source code, particularly code summarization and code generation …
natural language and source code, particularly code summarization and code generation …
Cat-probing: A metric-based approach to interpret how pre-trained models for programming language attend code structure
Code pre-trained models (CodePTMs) have recently demonstrated significant success in
code intelligence. To interpret these models, some probing methods have been applied …
code intelligence. To interpret these models, some probing methods have been applied …
Analyzing declarative deployment code with large language models
In the cloud-native era, developers have at their disposal an unprecedented landscape of
services to build scalable distributed systems. The DevOps paradigm emerged as a …
services to build scalable distributed systems. The DevOps paradigm emerged as a …
Learning Program Representations with a Tree-Structured Transformer
Learning vector representations for programs is a critical step in applying deep learning
techniques for program understanding tasks. Various neural network models are proposed …
techniques for program understanding tasks. Various neural network models are proposed …
Prompt sensitivity of language model for solving programming problems
A Shirafuji, T Ito, M Morishita… - New Trends in …, 2022 - ebooks.iospress.nl
A popular language model that can solve introductory programming problems, OpenAI's
Codex, has drawn much attention not only in the natural language processing field but also …
Codex, has drawn much attention not only in the natural language processing field but also …
An extensive study of the structure features in transformer-based code semantic summarization
Transformers are now widely utilized in code intelligence tasks. To better fit highly structured
source code, various structure information is passed into Transformer, such as positional …
source code, various structure information is passed into Transformer, such as positional …
REPOEXEC: Evaluate Code Generation with a Repository-Level Executable Benchmark
The ability of CodeLLMs to generate executable and functionally correct code at the\textit
{repository-level scale} remains largely unexplored. We introduce\methodnamews, a novel …
{repository-level scale} remains largely unexplored. We introduce\methodnamews, a novel …
Empirical Studies of Parameter Efficient Methods for Large Language Models of Code and Knowledge Transfer to R
Recently, Large Langauge Models (LLMs) have gained a lot of attention in the Software
Engineering (SE) community. LLMs or their variants pre-trained on code are used for many …
Engineering (SE) community. LLMs or their variants pre-trained on code are used for many …
A Critical Study of What Code-LLMs (Do Not) Learn
Large Language Models trained on code corpora (code-LLMs) have demonstrated
impressive performance in various coding assistance tasks. However, despite their …
impressive performance in various coding assistance tasks. However, despite their …