Survey of vulnerabilities in large language models revealed by adversarial attacks

E Shayegani, MAA Mamun, Y Fu, P Zaree… - arxiv preprint arxiv …, 2023 - arxiv.org
Large Language Models (LLMs) are swiftly advancing in architecture and capability, and as
they integrate more deeply into complex systems, the urgency to scrutinize their security …

Defense strategies for adversarial machine learning: A survey

P Bountakas, A Zarras, A Lekidis, C Xenakis - Computer Science Review, 2023 - Elsevier
Abstract Adversarial Machine Learning (AML) is a recently introduced technique, aiming to
deceive Machine Learning (ML) models by providing falsified inputs to render those models …

Tree of attacks: Jailbreaking black-box llms automatically

A Mehrotra, M Zampetakis… - Advances in …, 2025 - proceedings.neurips.cc
Abstract While Large Language Models (LLMs) display versatile functionality, they continue
to generate harmful, biased, and toxic content, as demonstrated by the prevalence of human …

Not what you've signed up for: Compromising real-world llm-integrated applications with indirect prompt injection

K Greshake, S Abdelnabi, S Mishra, C Endres… - Proceedings of the 16th …, 2023 - dl.acm.org
Large Language Models (LLMs) are increasingly being integrated into applications, with
versatile functionalities that can be easily modulated via natural language prompts. So far, it …

Prompt Injection attack against LLM-integrated Applications

Y Liu, G Deng, Y Li, K Wang, Z Wang, X Wang… - arxiv preprint arxiv …, 2023 - arxiv.org
Large Language Models (LLMs), renowned for their superior proficiency in language
comprehension and generation, stimulate a vibrant ecosystem of applications around them …

Masterkey: Automated jailbreak across multiple large language model chatbots

G Deng, Y Liu, Y Li, K Wang, Y Zhang, Z Li… - arxiv preprint arxiv …, 2023 - arxiv.org
Large Language Models (LLMs) have revolutionized Artificial Intelligence (AI) services due
to their exceptional proficiency in understanding and generating human-like text. LLM …

[PDF][PDF] Adversarial machine learning

A Vassilev, A Oprea, A Fordyce, H Anderson - Gaithersburg, MD, 2024 - site.unibo.it
Abstract This NIST Trustworthy and Responsible AI report develops a taxonomy of concepts
and defines terminology in the field of adversarial machine learning (AML). The taxonomy is …

Sok: Pragmatic assessment of machine learning for network intrusion detection

G Apruzzese, P Laskov… - 2023 IEEE 8th European …, 2023 - ieeexplore.ieee.org
Machine Learning (ML) has become a valuable asset to solve many real-world tasks. For
Network Intrusion Detection (NID), however, scientific advances in ML are still seen with …

Do LLMs dream of elephants (when told not to)? Latent concept association and associative memory in transformers

Y Jiang, G Rajendran, P Ravikumar… - Advances in Neural …, 2025 - proceedings.neurips.cc
Abstract Large Language Models (LLMs) have the capacity to store and recall facts. Through
experimentation with open-source models, we observe that this ability to retrieve facts can …

A survey on malware detection with graph representation learning

T Bilot, N El Madhoun, K Al Agha, A Zouaoui - ACM Computing Surveys, 2024 - dl.acm.org
Malware detection has become a major concern due to the increasing number and
complexity of malware. Traditional detection methods based on signatures and heuristics …