Explainable artificial intelligence applications in cyber security: State-of-the-art in research

Z Zhang, H Al Hamadi, E Damiani, CY Yeun… - IEEe …, 2022 - ieeexplore.ieee.org
This survey presents a comprehensive review of current literature on Explainable Artificial
Intelligence (XAI) methods for cyber security applications. Due to the rapid development of …

[HTML][HTML] Machine learning for Internet of Things data analysis: A survey

MS Mahdavinejad, M Rezvan, M Barekatain… - Digital Communications …, 2018 - Elsevier
Rapid developments in hardware, software, and communication technologies have
facilitated the emergence of Internet-connected sensory devices that provide observations …

Membership inference attacks from first principles

N Carlini, S Chien, M Nasr, S Song… - … IEEE symposium on …, 2022 - ieeexplore.ieee.org
A membership inference attack allows an adversary to query a trained machine learning
model to predict whether or not a particular example was contained in the model's training …

Efficient few-shot learning without prompts

L Tunstall, N Reimers, UES Jo, L Bates, D Korat… - arxiv preprint arxiv …, 2022 - arxiv.org
Recent few-shot methods, such as parameter-efficient fine-tuning (PEFT) and pattern
exploiting training (PET), have achieved impressive results in label-scarce settings …

Weight poisoning attacks on pre-trained models

K Kurita, P Michel, G Neubig - arxiv preprint arxiv:2004.06660, 2020 - arxiv.org
Recently, NLP has seen a surge in the usage of large pre-trained models. Users download
weights of models pre-trained on large datasets, then fine-tune the weights on a task of their …

Data shapley: Equitable valuation of data for machine learning

A Ghorbani, J Zou - International conference on machine …, 2019 - proceedings.mlr.press
As data becomes the fuel driving technological and economic growth, a fundamental
challenge is how to quantify the value of data in algorithmic predictions and decisions. For …

Machine unlearning of features and labels

A Warnecke, L Pirch, C Wressnegger… - arxiv preprint arxiv …, 2021 - arxiv.org
Removing information from a machine learning model is a non-trivial task that requires to
partially revert the training process. This task is unavoidable when sensitive data, such as …

Adversarial attacks on deep-learning models in natural language processing: A survey

WE Zhang, QZ Sheng, A Alhazmi, C Li - ACM Transactions on Intelligent …, 2020 - dl.acm.org
With the development of high computational devices, deep neural networks (DNNs), in
recent years, have gained significant popularity in many Artificial Intelligence (AI) …

Black-box generation of adversarial text sequences to evade deep learning classifiers

J Gao, J Lanchantin, ML Soffa… - 2018 IEEE Security and …, 2018 - ieeexplore.ieee.org
Although various techniques have been proposed to generate adversarial samples for white-
box attacks on text, little attention has been paid to a black-box attack, which is a more …

Understanding black-box predictions via influence functions

PW Koh, P Liang - International conference on machine …, 2017 - proceedings.mlr.press
How can we explain the predictions of a black-box model? In this paper, we use influence
functions—a classic technique from robust statistics—to trace a model's prediction through …