Suivre
Jianing Zhu
Jianing Zhu
Ph.D. student, HKBU
Adresse e-mail validée de comp.hkbu.edu.hk - Page d'accueil
Titre
Citée par
Citée par
Année
Geometry-aware instance-reweighted adversarial training
J Zhang, J Zhu, G Niu, B Han, M Sugiyama, M Kankanhalli
arXiv preprint arXiv:2010.01736, 2020
3212020
DeepInception: Hypnotize Large Language Model to Be Jailbreaker
X Li, Z Zhou, J Zhu, J Yao, T Liu, B Han
arXiv preprint arXiv:2311.03191, 2023
1242023
Reliable adversarial distillation with unreliable teachers
J Zhu, J Yao, B Han, J Zhang, T Liu, G Niu, J Zhou, J Xu, H Yang
arXiv preprint arXiv:2106.04928, 2021
742021
Understanding the interaction of adversarial training with noisy labels
J Zhu, J Zhang, B Han, T Liu, G Niu, H Yang, M Kankanhalli, M Sugiyama
arXiv preprint arXiv:2102.03482, 2021
302021
Diversified outlier exposure for out-of-distribution detection via informative extrapolation
J Zhu, Y Geng, J Yao, T Liu, G Niu, M Sugiyama, B Han
Advances in Neural Information Processing Systems 36, 22702-22734, 2023
192023
Unleashing Mask: Explore the Intrinsic Out-of-Distribution Detection Capability
J Zhu, H Li, J Yao, T Liu, J Xu, B Han
arXiv preprint arXiv:2306.03715, 2023
132023
Adversarial Training with Complementary Labels: On the Benefit of Gradually Informative Attacks
J Zhou, J Zhu, J Zhang, T Liu, G Niu, B Han, M Sugiyama
Advances in Neural Information Processing Systems 35, 23621-23633, 2022
102022
Combating Exacerbated Heterogeneity for Robust Models in Federated Learning
J Zhu, J Yao, T Liu, Q Yao, J Xu, B Han
arXiv preprint arXiv:2303.00250, 2023
92023
Unlearning with Control: Assessing Real-world Utility for Large Language Model Unlearning
Q Wang, B Han, P Yang, J Zhu, T Liu, M Sugiyama
arXiv preprint arXiv:2406.09179, 2024
42024
Decoupling the Class Label and the Target Concept in Machine Unlearning
J Zhu, B Han, J Yao, J Xu, G Niu, M Sugiyama
arXiv preprint arXiv:2406.08288, 2024
32024
Can Language Models Perform Robust Reasoning in Chain-of-thought Prompting with Noisy Rationales?
Z Zhou, R Tao, J Zhu, Y Luo, Z Wang, B Han
arXiv preprint arXiv:2410.23856, 2024
22024
Model Inversion Attacks: A Survey of Approaches and Countermeasures
Z Zhou, J Zhu, F Yu, X Li, X Peng, T Liu, B Han
arXiv preprint arXiv:2411.10023, 2024
12024
Can Large Language Models Reason Robustly with Noisy Rationales?
Z Zhou, R Tao, J Zhu, Y Luo, Z Wang, B Han
ICLR 2024 Workshop on Reliable and Responsible Foundation Models, 0
1
Self-Calibrated Tuning of Vision-Language Models for Out-of-Distribution Detection
G Yu, J Zhu, J Yao, B Han
arXiv preprint arXiv:2411.03359, 2024
2024
What If the Input is Expanded in OOD Detection?
B Zhang, J Zhu, Z Wang, T Liu, B Du, B Han
arXiv preprint arXiv:2410.18472, 2024
2024
Exploring model dynamics for accumulative poisoning discovery
J Zhu, X Guo, J Yao, C Du, L He, S Yuan, T Liu, L Wang, B Han
International Conference on Machine Learning, 42983-43004, 2023
2023
Le système ne peut pas réaliser cette opération maintenant. Veuillez réessayer plus tard.
Articles 1–16