Large language models for software engineering: A systematic literature review

X Hou, Y Zhao, Y Liu, Z Yang, K Wang, L Li… - ACM Transactions on …, 2024 - dl.acm.org
Large Language Models (LLMs) have significantly impacted numerous domains, including
Software Engineering (SE). Many recent publications have explored LLMs applied to …

A survey on deep learning for software engineering

Y Yang, X **a, D Lo, J Grundy - ACM Computing Surveys (CSUR), 2022 - dl.acm.org
In 2006, Geoffrey Hinton proposed the concept of training “Deep Neural Networks (DNNs)”
and an improved model training method to break the bottleneck of neural network …

The programmer's assistant: Conversational interaction with a large language model for software development

SI Ross, F Martinez, S Houde, M Muller… - Proceedings of the 28th …, 2023 - dl.acm.org
Large language models (LLMs) have recently been applied in software engineering to
perform tasks such as translating code between programming languages, generating code …

Communicative agents for software development

C Qian, X Cong, C Yang, W Chen, Y Su, J Xu… - arxiv preprint arxiv …, 2023 - arxiv.org
Software engineering is a domain characterized by intricate decision-making processes,
often relying on nuanced intuition and consultation. Recent advancements in deep learning …

Is ChatGPT the ultimate programming assistant--how far is it?

H Tian, W Lu, TO Li, X Tang, SC Cheung… - arxiv preprint arxiv …, 2023 - arxiv.org
Recently, the ChatGPT LLM has received great attention: it can be used as a bot for
discussing source code, prompting it to suggest changes, provide descriptions or even …

Graph neural networks: foundation, frontiers and applications

L Wu, P Cui, J Pei, L Zhao, X Guo - … of the 28th ACM SIGKDD Conference …, 2022 - dl.acm.org
The field of graph neural networks (GNNs) has seen rapid and incredible strides over the
recent years. Graph neural networks, also known as deep learning on graphs, graph …

Codexglue: A machine learning benchmark dataset for code understanding and generation

S Lu, D Guo, S Ren, J Huang, A Svyatkovskiy… - arxiv preprint arxiv …, 2021 - arxiv.org
Benchmark datasets have a significant impact on accelerating research in programming
language tasks. In this paper, we introduce CodeXGLUE, a benchmark dataset to foster …

Large language models meet nl2code: A survey

D Zan, B Chen, F Zhang, D Lu, B Wu, B Guan… - arxiv preprint arxiv …, 2022 - arxiv.org
The task of generating code from a natural language description, or NL2Code, is considered
a pressing and significant challenge in code intelligence. Thanks to the rapid development …

No more fine-tuning? an experimental evaluation of prompt tuning in code intelligence

C Wang, Y Yang, C Gao, Y Peng, H Zhang… - Proceedings of the 30th …, 2022 - dl.acm.org
Pre-trained models have been shown effective in many code intelligence tasks. These
models are pre-trained on large-scale unlabeled corpus and then fine-tuned in downstream …

A transformer-based approach for source code summarization

WU Ahmad, S Chakraborty, B Ray… - arxiv preprint arxiv …, 2020 - arxiv.org
Generating a readable summary that describes the functionality of a program is known as
source code summarization. In this task, learning code representation by modeling the …