Llm4vuln: A unified evaluation framework for decoupling and enhancing llms' vulnerability reasoning

Y Sun, D Wu, Y Xue, H Liu, W Ma, L Zhang… - arXiv preprint arXiv …, 2024 - arxiv.org
Large language models (LLMs) have demonstrated significant potential in various tasks,
including vulnerability detection. However, current efforts in this area are preliminary, lacking …

Clover: Clo sed-Loop Ver ifiable Code Generation

C Sun, Y Sheng, O Padon, C Barrett - International Symposium on AI …, 2024 - Springer
The use of large language models for code generation is a rapidly growing trend in software
development. However, without effective methods for ensuring the correctness of generated …

Selfpico: Self-guided partial code execution with llms

Z Xue, Z Gao, S Wang, X Hu, X Xia, S Li - Proceedings of the 33rd ACM …, 2024 - dl.acm.org
Code executability plays a vital role in software debugging and testing (eg, detecting runtime
exceptions or assertion violations). However, code execution, especially partial or arbitrary …

If At First You Don't Succeed, Try, Try, Again...? Insights and LLM-informed Tooling for Detecting Retry Bugs in Software Systems

BA Stoica, U Sethi, Y Su, C Zhou, S Lu, J Mace… - Proceedings of the …, 2024 - dl.acm.org
Retry---the re-execution of a task on failure---is a common mechanism to enable resilient
software systems. Yet, despite its commonality and long history, retry remains difficult to …

Transformer-based models are not yet perfect at learning to emulate structural recursion

D Zhang, C Tigges, Z Zhang, S Biderman… - arXiv preprint arXiv …, 2024 - arxiv.org
This paper investigates the ability of transformer-based models to learn structural recursion
from examples. Recursion is a universal concept in both natural and formal languages …

[PDF][PDF] SemCoder: Training Code Language Models with Comprehensive Semantics Reasoning

Y Ding, J Peng, MJ Min, G Kaiser, J Yang… - arXiv preprint arXiv …, 2024 - openreview.net
Abstract Code Large Language Models (Code LLMs) have excelled at tasks like code
completion but often miss deeper semantics such as execution effects and dynamic states …

Symmetry-Preserving Program Representations for Learning Code Semantics

K Pei, W Li, Q Jin, S Liu, S Geng, L Cavallaro… - arXiv preprint arXiv …, 2023 - arxiv.org
Large Language Models (LLMs) have shown promise in automated program reasoning, a
crucial aspect of many security tasks. However, existing LLM architectures for code are often …

Code Representation Pre-training with Complements from Program Executions

J Huang, J Zhao, Y Rong, Y Guo, Y He… - arXiv preprint arXiv …, 2023 - arxiv.org
Large language models (LLMs) for natural language processing have been grafted onto
programming language modeling for advancing code intelligence. Although it can be …

[PDF][PDF] Reasoning runtime behavior of a program with llm: How far are we?

J Chen, Z Pan, X Hu, Z Li, G Li, X Xia - arXiv preprint cs.SE …, 2024 - ginolzh.github.io
Large language models for code (ie, code LLMs) have shown strong code understanding
and generation capabilities. To evaluate the capabilities of code LLMs in various aspects …

Codeart: Better code models by attention regularization when symbols are lacking

Z Su, X Xu, Z Huang, Z Zhang, Y Ye, J Huang… - Proceedings of the …, 2024 - dl.acm.org
Transformer based code models have impressive performance in many software
engineering tasks. However, their effectiveness degrades when symbols are missing or not …