Pitfalls in language models for code intelligence: A taxonomy and survey

X She, Y Liu, Y Zhao, Y He, L Li… - arxiv preprint arxiv …, 2023 - arxiv.org
Modern language models (LMs) have been successfully employed in source code
generation and understanding, leading to a significant increase in research focused on …

Privacy in large language models: Attacks, defenses and future directions

H Li, Y Chen, J Luo, J Wang, H Peng, Y Kang… - arxiv preprint arxiv …, 2023 - arxiv.org
The advancement of large language models (LLMs) has significantly enhanced the ability to
effectively tackle various downstream NLP tasks and unify these tasks into generative …

A survey on large language models for software engineering

Q Zhang, C Fang, Y **e, Y Zhang, Y Yang… - arxiv preprint arxiv …, 2023 - arxiv.org
Software Engineering (SE) is the systematic design, development, maintenance, and
management of software applications underpinning the digital infrastructure of our modern …

Rethinking membership inference attacks against transfer learning

C Wu, J Chen, Q Fang, K He, Z Zhao… - IEEE Transactions …, 2024 - ieeexplore.ieee.org
Transfer learning, successful in knowledge translation across related tasks, faces a
substantial privacy threat from membership inference attacks (MIAs). These attacks, despite …

It's All in the Touch: Authenticating Users with HOST Gestures on Multi-Touch Screen Devices

C Wu, H Cao, G Xu, C Zhou, J Sun… - IEEE Transactions …, 2024 - ieeexplore.ieee.org
As smartphones proliferate, secure and user-friendly authentication methods are
increasingly critical. Existing behavioral biometrics, however, are often compromised by …

Security of Language Models for Code: A Systematic Literature Review

Y Chen, W Sun, C Fang, Z Chen, Y Ge, T Han… - arxiv preprint arxiv …, 2024 - arxiv.org
Language models for code (CodeLMs) have emerged as powerful tools for code-related
tasks, outperforming traditional methods and standard machine learning approaches …

Vulnerabilities in ai code generators: Exploring targeted data poisoning attacks

D Cotroneo, C Improta, P Liguori… - Proceedings of the 32nd …, 2024 - dl.acm.org
AI-based code generators have become pivotal in assisting developers in writing software
starting from natural language (NL). However, they are trained on large amounts of data …

Django: Detecting trojans in object detection models via gaussian focus calibration

G Shen, S Cheng, G Tao, K Zhang… - Advances in …, 2023 - proceedings.neurips.cc
Object detection models are vulnerable to backdoor or trojan attacks, where an attacker can
inject malicious triggers into the model, leading to altered behavior during inference. As a …

Assessing cybersecurity vulnerabilities in code large language models

MI Hossen, J Zhang, Y Cao, X Hei - arxiv preprint arxiv:2404.18567, 2024 - arxiv.org
Instruction-tuned Code Large Language Models (Code LLMs) are increasingly utilized as AI
coding assistants and integrated into various applications. However, the cybersecurity …