Backdoor Attack and Defense on Deep Learning: A Survey

Y Bai, G **ng, H Wu, Z Rao, C Ma… - IEEE Transactions …, 2024 - ieeexplore.ieee.org
Deep learning, as an important branch of machine learning, has been widely applied in
computer vision, natural language processing, speech recognition, and more. However …

Cooperative Backdoor Attack in Decentralized Reinforcement Learning with Theoretical Guarantee

M Gao, Y Zou, Z Zhang, X Cheng, D Yu - arxiv preprint arxiv:2405.15245, 2024 - arxiv.org
The safety of decentralized reinforcement learning (RL) is a challenging problem since
malicious agents can share their poisoned policies with benign agents. The paper …

SleeperNets: Universal Backdoor Poisoning Attacks Against Reinforcement Learning Agents

E Rathbun, C Amato, A Oprea - arxiv preprint arxiv:2405.20539, 2024 - arxiv.org
Reinforcement learning (RL) is an actively growing field that is seeing increased usage in
real-world, safety-critical applications--making it paramount to ensure the robustness of RL …

BLAST: A Stealthy Backdoor Leverage Attack against Cooperative Multi-Agent Deep Reinforcement Learning based Systems

Y Yu, S Yan, X Yin, J Fang, J Liu - arxiv preprint arxiv:2501.01593, 2025 - arxiv.org
Recent studies have shown that cooperative multi-agent deep reinforcement learning (c-
MADRL) is under the threat of backdoor attacks. Once a backdoor trigger is observed, it will …

Trading Devil RL: Backdoor attack via Stock market, Bayesian Optimization and Reinforcement Learning

O Mengara - arxiv preprint arxiv:2412.17908, 2024 - arxiv.org
With the rapid development of generative artificial intelligence, particularly large language
models, a number of sub-fields of deep learning have made significant progress and are …

Online Poisoning Attack Against Reinforcement Learning under Black-box Environments

J Li, B Zhang, J Wu - arxiv preprint arxiv:2412.00797, 2024 - arxiv.org
This paper proposes an online environment poisoning algorithm tailored for reinforcement
learning agents operating in a black-box setting, where an adversary deliberately …

A Disguised Wolf Is More Harmful Than a Toothless Tiger: Adaptive Malicious Code Injection Backdoor Attack Leveraging User Behavior as Triggers

S Wu, J Sang - arxiv preprint arxiv:2408.10334, 2024 - arxiv.org
In recent years, large language models (LLMs) have made significant progress in the field of
code generation. However, as more and more users rely on these models for software …

On the Robustness of Machine Learning Training in Security Sensitive Environments

G Severi - 2024 - search.proquest.com
Modern machine learning underpins a large variety of commercial software products,
including many cybersecurity solutions. Widely different models, from large transformers …