Ml attack models: Adversarial attacks and data poisoning attacks

J Lin, L Dang, M Rahouti, K **ong - arxiv preprint arxiv:2112.02797, 2021 - arxiv.org
Many state-of-the-art ML models have outperformed humans in various tasks such as image
classification. With such outstanding performance, ML models are widely used today …

Defense against adversarial attacks based on stochastic descent sign activation networks on medical images

Y Yang, FY Shih, U Roshan - International Journal of Pattern …, 2022 - World Scientific
Machine learning techniques in medical imaging systems are accurate, but minor
perturbations in the data known as adversarial attacks can fool them. These attacks make …

Accurate and adversarially robust classification of medical images and ECG time-series with gradient-free trained sign activation neural networks

Z Yang, Y Yang, Y Xue, FY Shih, J Ady… - … on Bioinformatics and …, 2020 - ieeexplore.ieee.org
Adversarial attacks in medical AI imaging systems can lead to misdiagnosis and insurance
fraud as recently highlighted by Finlayson et. al. in Science 2019. They can also be carried …

[KNIHA][B] AI, machine learning and deep learning: a security perspective

F Hu, X Hei - 2023 - books.google.com
Today, Artificial Intelligence (AI) and Machine Learning/Deep Learning (ML/DL) have
become the hottest areas in information technology. In our society, many intelligent devices …

Machine learning attack models

J Lin, L Dang, M Rahouti, K **ong - AI, Machine Learning and …, 2023 - taylorfrancis.com
As machine learning (ML) systems have been dramatically integrated into a broad range of
decision-making-sensitive applications for the past years, adversarial attacks and data …

Accuracy of white box and black box adversarial attacks on a sign activation 01 loss neural network ensemble

Y Xue, U Roshan - 2023 - openreview.net
In this work we ask the question: is an ensemble of single hidden layer sign activation 01
loss networks more robust to white box and black box adversarial attacks than an ensemble …

Adversarial and data poisoning attacks against deep learning

J Lin - 2022 - search.proquest.com
Abstract Machine translation software, image captioning, grammar check (Grammarly),
chatbot, real-time captioning and translation, music genre classification, and document …

Accuracy of TextFooler black box adversarial attacks on 01 loss sign activation neural network ensemble

Y Xue, U Roshan - arxiv preprint arxiv:2402.07347, 2024 - arxiv.org
Recent work has shown the defense of 01 loss sign activation neural networks against
image classification adversarial attacks. A public challenge to attack the models on CIFAR10 …

Towards Adversarial Robustness with 01 Loss Models, and Novel Convolutional Neural Net Systems for Ultrasound Images

M **e - 2021 - search.proquest.com
TOWARDS ADVERSARIAL ROBUSTNESS WITH 01 LOSS MODELS, AND NOVEL
CONVOLUTIONAL NEURAL NET SYSTEMS FOR ULTRASOUND IMAGES by Page 1 …

Gradient Free Sign Activation Zero One Loss Neural Networks for Adversarially Robust Classification

Y Xue - 2021 - search.proquest.com
The zero-one loss function is less sensitive to outliers than convex surrogate losses such as
hinge and cross-entropy. However, as a non-convex function, it has a large number of local …