Deep learning-based software engineering: progress, challenges, and opportunities

X Chen, X Hu, Y Huang, H Jiang, W Ji, Y Jiang… - Science China …, 2025 - Springer
Researchers have recently achieved significant advances in deep learning techniques,
which in turn has substantially advanced other research disciplines, such as natural …

Contrastive self-supervised learning: review, progress, challenges and future research directions

P Kumar, P Rawat, S Chauhan - International Journal of Multimedia …, 2022 - Springer
In the last decade, deep supervised learning has had tremendous success. However, its
flaws, such as its dependency on manual and costly annotations on large datasets and …

Reacc: A retrieval-augmented code completion framework

S Lu, N Duan, H Han, D Guo, S Hwang… - arxiv preprint arxiv …, 2022 - arxiv.org
Code completion, which aims to predict the following code token (s) according to the code
context, can improve the productivity of software development. Recent work has proved that …

Path-sensitive code embedding via contrastive learning for software vulnerability detection

X Cheng, G Zhang, H Wang, Y Sui - Proceedings of the 31st ACM …, 2022 - dl.acm.org
Machine learning and its promising branch deep learning have shown success in a wide
range of application domains. Recently, much effort has been expended on applying deep …

Syncobert: Syntax-guided multi-modal contrastive pre-training for code representation

X Wang, Y Wang, F Mi, P Zhou, Y Wan, X Liu… - arxiv preprint arxiv …, 2021 - arxiv.org
Code representation learning, which aims to encode the semantics of source code into
distributed vectors, plays an important role in recent deep-learning-based models for code …

Bridging pre-trained models and downstream tasks for source code understanding

D Wang, Z Jia, S Li, Y Yu, Y **ong, W Dong… - Proceedings of the 44th …, 2022 - dl.acm.org
With the great success of pre-trained models, the pretrain-then-finetune paradigm has been
widely adopted on downstream tasks for source code understanding. However, compared to …

Contrabert: Enhancing code pre-trained models via contrastive learning

S Liu, B Wu, X **e, G Meng, Y Liu - 2023 IEEE/ACM 45th …, 2023 - ieeexplore.ieee.org
Large-scale pre-trained models such as CodeBERT, GraphCodeBERT have earned
widespread attention from both academia and industry. Attributed to the superior ability in …

Selfapr: Self-supervised program repair with test execution diagnostics

H Ye, M Martinez, X Luo, T Zhang… - Proceedings of the 37th …, 2022 - dl.acm.org
Learning-based program repair has achieved good results in a recent series of papers. Yet,
we observe that the related work fails to repair some bugs because of a lack of knowledge …

[PDF][PDF] Unifying the perspectives of nlp and software engineering: A survey on language models for code

Z Zhang, C Chen, B Liu, C Liao, Z Gong… - arxiv preprint arxiv …, 2023 - simg.baai.ac.cn
In this work we systematically review the recent advancements in code processing with
language models, covering 50+ models, 30+ evaluation tasks, 170+ datasets, and 700 …

Traced: Execution-aware pre-training for source code

Y Ding, B Steenhoek, K Pei, G Kaiser, W Le… - Proceedings of the 46th …, 2024 - dl.acm.org
Most existing pre-trained language models for source code focus on learning the static code
text, typically augmented with static code structures (abstract syntax tree, dependency …