Adversarial attacks and defenses in explainable artificial intelligence: A survey

H Baniecki, P Biecek - Information Fusion, 2024 - Elsevier
Explainable artificial intelligence (XAI) methods are portrayed as a remedy for debugging
and trusting statistical and deep learning models, as well as interpreting their predictions …

Explainable artificial intelligence for cybersecurity: a literature survey

F Charmet, HC Tanuwidjaja, S Ayoubi… - Annals of …, 2022 - Springer
With the extensive application of deep learning (DL) algorithms in recent years, eg, for
detecting Android malware or vulnerable source code, artificial intelligence (AI) and …

SoK: Explainable machine learning in adversarial environments

M Noppel, C Wressnegger - 2024 IEEE Symposium on Security …, 2024 - ieeexplore.ieee.org
Modern deep learning methods have long been considered black boxes due to the lack of
insights into their decision-making process. However, recent advances in explainable …

[HTML][HTML] When explainability turns into a threat-using xAI to fool a fake news detection method

R Kozik, M Ficco, A Pawlicka, M Pawlicki, F Palmieri… - Computers & …, 2024 - Elsevier
The inclusion of Explainability of Artificial Intelligence (xAI) has become a mandatory
requirement for designing and implementing reliable, interpretable and ethical AI solutions …

Fooling explanations in text classifiers

A Ivankay, I Girardi, C Marchiori, P Frossard - arxiv preprint arxiv …, 2022 - arxiv.org
State-of-the-art text classification models are becoming increasingly reliant on deep neural
networks (DNNs). Due to their black-box nature, faithful and robust explanation methods …

Adversarial attacks in explainable machine learning: A survey of threats against models and humans

J Vadillo, R Santana, JA Lozano - … Reviews: Data Mining and …, 2025 - Wiley Online Library
Reliable deployment of machine learning models such as neural networks continues to be
challenging due to several limitations. Some of the main shortcomings are the lack of …

Are Your Explanations Reliable? Investigating the Stability of LIME in Explaining Text Classifiers by Marrying XAI and Adversarial Attack

C Burger, L Chen, T Le - arxiv preprint arxiv:2305.12351, 2023 - arxiv.org
LIME has emerged as one of the most commonly referenced tools in explainable AI (XAI)
frameworks that is integrated into critical machine learning applications--eg, healthcare and …

Revisiting the robustness of post-hoc interpretability methods

J Wei, H Turbé, G Mengaldo - arxiv preprint arxiv:2407.19683, 2024 - arxiv.org
Post-hoc interpretability methods play a critical role in explainable artificial intelligence (XAI),
as they pinpoint portions of data that a trained deep learning model deemed important to …

Large language models and sentiment analysis in financial markets: A review, datasets and case study

C Liu, A Arulappan, R Naha, A Mahanti… - IEEE …, 2024 - ieeexplore.ieee.org
This paper comprehensively examines Large Language Models (LLMs) in sentiment
analysis, specifically focusing on financial markets and exploring the correlation between …

Understanding and enhancing robustness of concept-based models

S Sinha, M Huai, J Sun, A Zhang - … of the AAAI Conference on Artificial …, 2023 - ojs.aaai.org
Rising usage of deep neural networks to perform decision making in critical applications like
medical diagnosis and fi-nancial analysis have raised concerns regarding their reliability …