Large language models for software engineering: A systematic literature review
Large Language Models (LLMs) have significantly impacted numerous domains, including
Software Engineering (SE). Many recent publications have explored LLMs applied to …
Software Engineering (SE). Many recent publications have explored LLMs applied to …
Towards an understanding of large language models in software engineering tasks
Abstract Large Language Models (LLMs) have drawn widespread attention and research
due to their astounding performance in text generation and reasoning tasks. Derivative …
due to their astounding performance in text generation and reasoning tasks. Derivative …
Dual-interactive fusion for code-mixed deep representation learning in tag recommendation
Automatic tagging on software information sites is a tag recommendation service. It aims to
recommend content-based tags for a software object to help developers make distinctions …
recommend content-based tags for a software object to help developers make distinctions …
Unveiling code pre-trained models: Investigating syntax and semantics capacities
Code models have made significant advancements in code intelligence by encoding
knowledge about programming languages. While previous studies have explored the …
knowledge about programming languages. While previous studies have explored the …
Towards efficient fine-tuning of pre-trained code models: An experimental study and beyond
Recently, fine-tuning pre-trained code models such as CodeBERT on downstream tasks has
achieved great success in many software testing and analysis tasks. While effective and …
achieved great success in many software testing and analysis tasks. While effective and …
Diet code is healthy: Simplifying programs for pre-trained models of code
Pre-trained code representation models such as CodeBERT have demonstrated superior
performance in a variety of software engineering tasks, yet they are often heavy in …
performance in a variety of software engineering tasks, yet they are often heavy in …
Prompt-tuned code language model as a neural knowledge base for type inference in statically-typed partial code
Partial code usually involves non-fully-qualified type names (non-FQNs) and undeclared
receiving objects. Resolving the FQNs of these non-FQN types and undeclared receiving …
receiving objects. Resolving the FQNs of these non-FQN types and undeclared receiving …
CODE-MVP: Learning to represent source code from multiple views with contrastive pre-training
Recent years have witnessed increasing interest in code representation learning, which
aims to represent the semantics of source code into distributed vectors. Currently, various …
aims to represent the semantics of source code into distributed vectors. Currently, various …
AST-Probe: Recovering abstract syntax trees from hidden representations of pre-trained language models
The objective of pre-trained language models is to learn contextual representations of
textual data. Pre-trained language models have become mainstream in natural language …
textual data. Pre-trained language models have become mainstream in natural language …
Graph neural networks for vulnerability detection: A counterfactual explanation
Vulnerability detection is crucial for ensuring the security and reliability of software systems.
Recently, Graph Neural Networks (GNNs) have emerged as a prominent code embedding …
Recently, Graph Neural Networks (GNNs) have emerged as a prominent code embedding …