Large language models for software engineering: A systematic literature review

X Hou, Y Zhao, Y Liu, Z Yang, K Wang, L Li… - ACM Transactions on …, 2024 - dl.acm.org
Large Language Models (LLMs) have significantly impacted numerous domains, including
Software Engineering (SE). Many recent publications have explored LLMs applied to …

Towards an understanding of large language models in software engineering tasks

Z Zheng, K Ning, Q Zhong, J Chen, W Chen… - Empirical Software …, 2025 - Springer
Abstract Large Language Models (LLMs) have drawn widespread attention and research
due to their astounding performance in text generation and reasoning tasks. Derivative …

Dual-interactive fusion for code-mixed deep representation learning in tag recommendation

L Li, P Wang, X Zheng, Q **e, X Tao, JD Velásquez - Information Fusion, 2023 - Elsevier
Automatic tagging on software information sites is a tag recommendation service. It aims to
recommend content-based tags for a software object to help developers make distinctions …

Unveiling code pre-trained models: Investigating syntax and semantics capacities

W Ma, S Liu, M Zhao, X **e, W Wang, Q Hu… - ACM Transactions on …, 2024 - dl.acm.org
Code models have made significant advancements in code intelligence by encoding
knowledge about programming languages. While previous studies have explored the …

Towards efficient fine-tuning of pre-trained code models: An experimental study and beyond

E Shi, Y Wang, H Zhang, L Du, S Han… - Proceedings of the …, 2023 - dl.acm.org
Recently, fine-tuning pre-trained code models such as CodeBERT on downstream tasks has
achieved great success in many software testing and analysis tasks. While effective and …

Diet code is healthy: Simplifying programs for pre-trained models of code

Z Zhang, H Zhang, B Shen, X Gu - Proceedings of the 30th ACM Joint …, 2022 - dl.acm.org
Pre-trained code representation models such as CodeBERT have demonstrated superior
performance in a variety of software engineering tasks, yet they are often heavy in …

Prompt-tuned code language model as a neural knowledge base for type inference in statically-typed partial code

Q Huang, Z Yuan, Z **ng, X Xu, L Zhu… - Proceedings of the 37th …, 2022 - dl.acm.org
Partial code usually involves non-fully-qualified type names (non-FQNs) and undeclared
receiving objects. Resolving the FQNs of these non-FQN types and undeclared receiving …

CODE-MVP: Learning to represent source code from multiple views with contrastive pre-training

X Wang, Y Wang, Y Wan, J Wang, P Zhou, L Li… - arxiv preprint arxiv …, 2022 - arxiv.org
Recent years have witnessed increasing interest in code representation learning, which
aims to represent the semantics of source code into distributed vectors. Currently, various …

AST-Probe: Recovering abstract syntax trees from hidden representations of pre-trained language models

JA Hernández López, M Weyssow… - Proceedings of the 37th …, 2022 - dl.acm.org
The objective of pre-trained language models is to learn contextual representations of
textual data. Pre-trained language models have become mainstream in natural language …

Graph neural networks for vulnerability detection: A counterfactual explanation

Z Chu, Y Wan, Q Li, Y Wu, H Zhang, Y Sui… - Proceedings of the 33rd …, 2024 - dl.acm.org
Vulnerability detection is crucial for ensuring the security and reliability of software systems.
Recently, Graph Neural Networks (GNNs) have emerged as a prominent code embedding …