Rethinking lipschitz neural networks and certified robustness: A boolean function perspective

B Zhang, D Jiang, D He… - Advances in neural …, 2022 - proceedings.neurips.cc
Designing neural networks with bounded Lipschitz constant is a promising way to obtain
certifiably robust classifiers against adversarial examples. However, the relevant progress …

Certified robustness via dynamic margin maximization and improved lipschitz regularization

M Fazlyab, T Entesari, A Roy… - Advances in Neural …, 2023 - proceedings.neurips.cc
To improve the robustness of deep classifiers against adversarial perturbations, many
approaches have been proposed, such as designing new architectures with better …

Performance scaling via optimal transport: Enabling data selection from partially revealed sources

F Kang, HA Just, AK Sahu, R Jia - Advances in Neural …, 2023 - proceedings.neurips.cc
Traditionally, data selection has been studied in settings where all samples from prospective
sources are fully revealed to a machine learning developer. However, in practical data …

Posthoc privacy guarantees for collaborative inference with modified Propose-Test-Release

A Singh, P Vepakomma, V Sharma… - Advances in Neural …, 2023 - proceedings.neurips.cc
Cloud-based machine learning inference is an emerging paradigm where users query by
sending their data through a service provider who runs an ML model on that data and …

Input-relational verification of deep neural networks

D Banerjee, C Xu, G Singh - Proceedings of the ACM on Programming …, 2024 - dl.acm.org
We consider the verification of input-relational properties defined over deep neural networks
(DNNs) such as robustness against universal adversarial perturbations, monotonicity, etc …

Security and Privacy Issues in Deep Reinforcement Learning: Threats and Countermeasures

K Mo, P Ye, X Ren, S Wang, W Li, J Li - ACM Computing Surveys, 2024 - dl.acm.org
Deep Reinforcement Learning (DRL) is an essential subfield of Artificial Intelligence (AI),
where agents interact with environments to learn policies for solving complex tasks. In recent …

Aligning relational learning with lipschitz fairness

Y Jia, C Zhang, S Vosoughi - The Twelfth International Conference …, 2024 - openreview.net
Relational learning has gained significant attention, led by the expressiveness of Graph
Neural Networks (GNNs) on graph data. While the inherent biases in common graph data …

Improving the accuracy-robustness trade-off of classifiers via adaptive smoothing

Y Bai, BG Anderson, A Kim, S Sojoudi - SIAM Journal on Mathematics of Data …, 2024 - SIAM
While prior research has proposed a plethora of methods that build neural classifiers robust
against adversarial robustness, practitioners are still reluctant to adopt them due to their …

Efficient bound of Lipschitz constant for convolutional layers by gram iteration

B Delattre, Q Barthélemy, A Araujo… - … on Machine Learning, 2023 - proceedings.mlr.press
Since the control of the Lipschitz constant has a great impact on the training stability,
generalization, and robustness of neural networks, the estimation of this value is nowadays …

Verification of neural control barrier functions with symbolic derivative bounds propagation

H Hu, Y Yang, T Wei, C Liu - arxiv preprint arxiv:2410.16281, 2024 - arxiv.org
Control barrier functions (CBFs) are important in safety-critical systems and robot control
applications. Neural networks have been used to parameterize and synthesize CBFs with …