Backdoor attacks and defenses targeting multi-domain ai models: A comprehensive review

S Zhang, Y Pan, Q Liu, Z Yan, KKR Choo… - ACM Computing …, 2024 - dl.acm.org
Since the emergence of security concerns in artificial intelligence (AI), there has been
significant attention devoted to the examination of backdoor attacks. Attackers can utilize …

Privacy in large language models: Attacks, defenses and future directions

H Li, Y Chen, J Luo, J Wang, H Peng, Y Kang… - arxiv preprint arxiv …, 2023 - arxiv.org
The advancement of large language models (LLMs) has significantly enhanced the ability to
effectively tackle various downstream NLP tasks and unify these tasks into generative …

Backdoorllm: A comprehensive benchmark for backdoor attacks on large language models

Y Li, H Huang, Y Zhao, X Ma, J Sun - arxiv preprint arxiv:2408.12798, 2024 - arxiv.org
Generative Large Language Models (LLMs) have made significant strides across various
tasks, but they remain vulnerable to backdoor attacks, where specific triggers in the prompt …

Cleangen: Mitigating backdoor attacks for generation tasks in large language models

Y Li, Z Xu, F Jiang, L Niu, D Sahabandu… - arxiv preprint arxiv …, 2024 - arxiv.org
The remarkable performance of large language models (LLMs) in generation tasks has
enabled practitioners to leverage publicly available models to power custom applications …

Text-tuple-table: Towards information integration in text-to-table generation via global tuple extraction

Z Deng, C Chan, W Wang, Y Sun, W Fan… - arxiv preprint arxiv …, 2024 - arxiv.org
The task of condensing large chunks of textual information into concise and structured tables
has gained attention recently due to the emergence of Large Language Models (LLMs) and …

Negotiationtom: A benchmark for stress-testing machine theory of mind on negotiation surrounding

C Chan, C Jiayang, Y Yim, Z Deng, W Fan, H Li… - arxiv preprint arxiv …, 2024 - arxiv.org
Large Language Models (LLMs) have sparked substantial interest and debate concerning
their potential emergence of Theory of Mind (ToM) ability. Theory of mind evaluations …

Safety at Scale: A Comprehensive Survey of Large Model Safety

X Ma, Y Gao, Y Wang, R Wang, X Wang, Y Sun… - arxiv preprint arxiv …, 2025 - arxiv.org
The rapid advancement of large models, driven by their exceptional abilities in learning and
generalization through large-scale pre-training, has reshaped the landscape of Artificial …

PoisonBench: Assessing Large Language Model Vulnerability to Data Poisoning

T Fu, M Sharma, P Torr, SB Cohen, D Krueger… - arxiv preprint arxiv …, 2024 - arxiv.org
Preference learning is a central component for aligning current LLMs, but this process can
be vulnerable to data poisoning attacks. To address this concern, we introduce …

ECON: On the Detection and Resolution of Evidence Conflicts

C Jiayang, C Chan, Q Zhuang, L Qiu, T Zhang… - arxiv preprint arxiv …, 2024 - arxiv.org
The rise of large language models (LLMs) has significantly influenced the quality of
information in decision-making systems, leading to the prevalence of AI-generated content …