Adversarial machine learning for network intrusion detection systems: A comprehensive survey
Network-based Intrusion Detection System (NIDS) forms the frontline defence against
network attacks that compromise the security of the data, systems, and networks. In recent …
network attacks that compromise the security of the data, systems, and networks. In recent …
When machine learning meets privacy: A survey and outlook
The newly emerged machine learning (eg, deep learning) methods have become a strong
driving force to revolutionize a wide range of industries, such as smart healthcare, financial …
driving force to revolutionize a wide range of industries, such as smart healthcare, financial …
Universal and transferable adversarial attacks on aligned language models
Because" out-of-the-box" large language models are capable of generating a great deal of
objectionable content, recent work has focused on aligning these models in an attempt to …
objectionable content, recent work has focused on aligning these models in an attempt to …
On evaluating adversarial robustness of large vision-language models
Large vision-language models (VLMs) such as GPT-4 have achieved unprecedented
performance in response generation, especially with visual inputs, enabling more creative …
performance in response generation, especially with visual inputs, enabling more creative …
Glaze: Protecting artists from style mimicry by {Text-to-Image} models
Recent text-to-image diffusion models such as MidJourney and Stable Diffusion threaten to
displace many in the professional artist community. In particular, models can learn to mimic …
displace many in the professional artist community. In particular, models can learn to mimic …
Foundational challenges in assuring alignment and safety of large language models
This work identifies 18 foundational challenges in assuring the alignment and safety of large
language models (LLMs). These challenges are organized into three different categories …
language models (LLMs). These challenges are organized into three different categories …
Red teaming language models with language models
Language Models (LMs) often cannot be deployed because of their potential to harm users
in hard-to-predict ways. Prior work identifies harmful behaviors before deployment by using …
in hard-to-predict ways. Prior work identifies harmful behaviors before deployment by using …
Autodan: Generating stealthy jailbreak prompts on aligned large language models
The aligned Large Language Models (LLMs) are powerful language understanding and
decision-making tools that are created through extensive alignment with human feedback …
decision-making tools that are created through extensive alignment with human feedback …
On adaptive attacks to adversarial example defenses
Adaptive attacks have (rightfully) become the de facto standard for evaluating defenses to
adversarial examples. We find, however, that typical adaptive evaluations are incomplete …
adversarial examples. We find, however, that typical adaptive evaluations are incomplete …
A survey on adversarial attacks and defences
Deep learning has evolved as a strong and efficient framework that can be applied to a
broad spectrum of complex learning problems which were difficult to solve using the …
broad spectrum of complex learning problems which were difficult to solve using the …