Deep learning-based software engineering: progress, challenges, and opportunities
Researchers have recently achieved significant advances in deep learning techniques,
which in turn has substantially advanced other research disciplines, such as natural …
which in turn has substantially advanced other research disciplines, such as natural …
Contrastive self-supervised learning: review, progress, challenges and future research directions
In the last decade, deep supervised learning has had tremendous success. However, its
flaws, such as its dependency on manual and costly annotations on large datasets and …
flaws, such as its dependency on manual and costly annotations on large datasets and …
Reacc: A retrieval-augmented code completion framework
Code completion, which aims to predict the following code token (s) according to the code
context, can improve the productivity of software development. Recent work has proved that …
context, can improve the productivity of software development. Recent work has proved that …
Path-sensitive code embedding via contrastive learning for software vulnerability detection
Machine learning and its promising branch deep learning have shown success in a wide
range of application domains. Recently, much effort has been expended on applying deep …
range of application domains. Recently, much effort has been expended on applying deep …
Syncobert: Syntax-guided multi-modal contrastive pre-training for code representation
Code representation learning, which aims to encode the semantics of source code into
distributed vectors, plays an important role in recent deep-learning-based models for code …
distributed vectors, plays an important role in recent deep-learning-based models for code …
Bridging pre-trained models and downstream tasks for source code understanding
With the great success of pre-trained models, the pretrain-then-finetune paradigm has been
widely adopted on downstream tasks for source code understanding. However, compared to …
widely adopted on downstream tasks for source code understanding. However, compared to …
Contrabert: Enhancing code pre-trained models via contrastive learning
Large-scale pre-trained models such as CodeBERT, GraphCodeBERT have earned
widespread attention from both academia and industry. Attributed to the superior ability in …
widespread attention from both academia and industry. Attributed to the superior ability in …
Selfapr: Self-supervised program repair with test execution diagnostics
Learning-based program repair has achieved good results in a recent series of papers. Yet,
we observe that the related work fails to repair some bugs because of a lack of knowledge …
we observe that the related work fails to repair some bugs because of a lack of knowledge …
[PDF][PDF] Unifying the perspectives of nlp and software engineering: A survey on language models for code
Z Zhang, C Chen, B Liu, C Liao, Z Gong… - arxiv preprint arxiv …, 2023 - simg.baai.ac.cn
In this work we systematically review the recent advancements in code processing with
language models, covering 50+ models, 30+ evaluation tasks, 170+ datasets, and 700 …
language models, covering 50+ models, 30+ evaluation tasks, 170+ datasets, and 700 …
Traced: Execution-aware pre-training for source code
Most existing pre-trained language models for source code focus on learning the static code
text, typically augmented with static code structures (abstract syntax tree, dependency …
text, typically augmented with static code structures (abstract syntax tree, dependency …