Skip to content

Commit 6eaac27

Browse files
author
Miltos Allamanis
committed
Add two recent papers.
1 parent 3798b5f commit 6eaac27

File tree

2 files changed

+24
-0
lines changed

2 files changed

+24
-0
lines changed
Lines changed: 12 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,12 @@
1+
---
2+
layout: publication
3+
title: "Learning to Reduce False Positives in Analytic Bug Detectors"
4+
authors: Anant Kharkar, Roshanak Zilouchian Moghaddam, Matthew Jin, Xiaoyu Liu, Xin Shi, Colin Clement, Neel Sundaresan
5+
conference: ICSE
6+
year: 2022
7+
bibkey: kharkar2022learning
8+
additional_links:
9+
- {name: "ArXiV", url: "https://arxiv.org/abs/2203.09907"}
10+
tags: ["Transformer", "static analysis"]
11+
---
12+
Due to increasingly complex software design and rapid iterative development, code defects and security vulnerabilities are prevalent in modern software. In response, programmers rely on static analysis tools to regularly scan their codebases and find potential bugs. In order to maximize coverage, however, these tools generally tend to report a significant number of false positives, requiring developers to manually verify each warning. To address this problem, we propose a Transformer-based learning approach to identify false positive bug warnings. We demonstrate that our models can improve the precision of static analysis by 17.5%. In addition, we validated the generalizability of this approach across two major bug types: null dereference and resource leak.

_publications/lu2022reacc.markdown

Lines changed: 12 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,12 @@
1+
---
2+
layout: publication
3+
title: "ReACC: A Retrieval-Augmented Code Completion Framework"
4+
authors: Shuai Lu, Nan Duan, Hojae Han, Daya Guo, Seung-won Hwang, Alexey Svyatkovskiy
5+
conference:
6+
year: 2022
7+
bibkey: lu2022reacc
8+
additional_links:
9+
- {name: "ArXiV", url: "https://arxiv.org/abs/2203.07722"}
10+
tags: ["Transformer", "autocomplete"]
11+
---
12+
Code completion, which aims to predict the following code token(s) according to the code context, can improve the productivity of software development. Recent work has proved that statistical language modeling with transformers can greatly improve the performance in the code completion task via learning from large-scale source code datasets. However, current approaches focus only on code context within the file or project, i.e. internal context. Our distinction is utilizing "external" context, inspired by human behaviors of copying from the related code snippets when writing code. Specifically, we propose a retrieval-augmented code completion framework, leveraging both lexical copying and referring to code with similar semantics by retrieval. We adopt a stage-wise training approach that combines a source code retriever and an auto-regressive language model for programming language. We evaluate our approach in the code completion task in Python and Java programming languages, achieving a state-of-the-art performance on CodeXGLUE benchmark.

0 commit comments

Comments
 (0)