Neural polarizer: A lightweight and effective backdoor defense via purifying poisoned features

M Zhu, S Wei, H Zha, B Wu - Advances in Neural …, 2023 - proceedings.neurips.cc
Recent studies have demonstrated the susceptibility of deep neural networks to backdoor
attacks. Given a backdoored model, its prediction of a poisoned sample with trigger will be …

Enhancing fine-tuning based backdoor defense with sharpness-aware minimization

M Zhu, S Wei, L Shen, Y Fan… - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
Backdoor defense, which aims to detect or mitigate the effect of malicious triggers introduced
by attackers, is becoming increasingly critical for machine learning security and integrity …

Towards efficient adversarial training on vision transformers

B Wu, J Gu, Z Li, D Cai, X He, W Liu - European Conference on Computer …, 2022 - Springer
Abstract Vision Transformer (ViT), as a powerful alternative to Convolutional Neural Network
(CNN), has received much attention. Recent work showed that ViTs are also vulnerable to …

Boosting the transferability of adversarial attacks with reverse adversarial perturbation

Z Qin, Y Fan, Y Liu, L Shen, Y Zhang… - Advances in neural …, 2022 - proceedings.neurips.cc
Deep neural networks (DNNs) have been shown to be vulnerable to adversarial examples,
which can produce erroneous predictions by injecting imperceptible perturbations. In this …

Prior-guided adversarial initialization for fast adversarial training

X Jia, Y Zhang, X Wei, B Wu, K Ma, J Wang… - European Conference on …, 2022 - Springer
Fast adversarial training (FAT) effectively improves the efficiency of standard adversarial
training (SAT). However, initial FAT encounters catastrophic overfitting, ie, the robust …

Triangle attack: A query-efficient decision-based adversarial attack

X Wang, Z Zhang, K Tong, D Gong, K He, Z Li… - European conference on …, 2022 - Springer
Decision-based attack poses a severe threat to real-world applications since it regards the
target model as a black box and only accesses the hard prediction label. Great efforts have …

Revisiting backdoor attacks against large vision-language models

S Liang, J Liang, T Pang, C Du, A Liu… - arxiv preprint arxiv …, 2024 - arxiv.org
Instruction tuning enhances large vision-language models (LVLMs) but raises security risks
through potential backdoor attacks due to their openness. Previous backdoor studies focus …

Towards robust physical-world backdoor attacks on lane detection

X Zhang, A Liu, T Zhang, S Liang, X Liu - Proceedings of the 32nd ACM …, 2024 - dl.acm.org
Deep learning-based lane detection (LD) plays a critical role in autonomous driving
systems, such as adaptive cruise control. However, it is vulnerable to backdoor attacks …

A large-scale multiple-objective method for black-box attack against object detection

S Liang, L Li, Y Fan, X Jia, J Li, B Wu, X Cao - European Conference on …, 2022 - Springer
Recent studies have shown that detectors based on deep models are vulnerable to
adversarial examples, even in the black-box scenario where the attacker cannot access the …

Hide in thicket: Generating imperceptible and rational adversarial perturbations on 3d point clouds

T Lou, X Jia, J Gu, L Liu, S Liang… - Proceedings of the …, 2024 - openaccess.thecvf.com
Adversarial attack methods based on point manipulation for 3D point cloud classification
have revealed the fragility of 3D models yet the adversarial examples they produce are …