Surveying neuro-symbolic approaches for reliable artificial intelligence of things

Z Lu, I Afridi, HJ Kang, I Ruchkin, X Zheng - Journal of Reliable Intelligent …, 2024 - Springer
Abstract The integration of Artificial Intelligence (AI) with the Internet of Things (IoT), known
as the Artificial Intelligence of Things (AIoT), enhances the devices' processing and analysis …

Trustllm: Trustworthiness in large language models

Y Huang, L Sun, H Wang, S Wu, Q Zhang, Y Li… - arxiv preprint arxiv …, 2024 - arxiv.org
Large language models (LLMs), exemplified by ChatGPT, have gained considerable
attention for their excellent natural language processing capabilities. Nonetheless, these …

[HTML][HTML] Position: TrustLLM: Trustworthiness in large language models

Y Huang, L Sun, H Wang, S Wu… - International …, 2024 - proceedings.mlr.press
Large language models (LLMs) have gained considerable attention for their excellent
natural language processing capabilities. Nonetheless, these LLMs present many …

General cutting planes for bound-propagation-based neural network verification

H Zhang, S Wang, K Xu, L Li, B Li… - Advances in neural …, 2022 - proceedings.neurips.cc
Bound propagation methods, when combined with branch and bound, are among the most
effective methods to formally verify properties of deep neural networks such as correctness …

Rethinking lipschitz neural networks and certified robustness: A boolean function perspective

B Zhang, D Jiang, D He… - Advances in neural …, 2022 - proceedings.neurips.cc
Designing neural networks with bounded Lipschitz constant is a promising way to obtain
certifiably robust classifiers against adversarial examples. However, the relevant progress …

Sok: Certified robustness for deep neural networks

L Li, T **e, B Li - 2023 IEEE symposium on security and privacy …, 2023 - ieeexplore.ieee.org
Great advances in deep neural networks (DNNs) have led to state-of-the-art performance on
a wide range of tasks. However, recent studies have shown that DNNs are vulnerable to …

Certified training: Small boxes are all you need

MN Müller, F Eckert, M Fischer, M Vechev - arxiv preprint arxiv …, 2022 - arxiv.org
To obtain, deterministic guarantees of adversarial robustness, specialized training methods
are used. We propose, SABR, a novel such certified training method, based on the key …

Efficiently computing local lipschitz constants of neural networks via bound propagation

Z Shi, Y Wang, H Zhang, JZ Kolter… - Advances in Neural …, 2022 - proceedings.neurips.cc
Lipschitz constants are connected to many properties of neural networks, such as
robustness, fairness, and generalization. Existing methods for computing Lipschitz constants …

Connecting certified and adversarial training

Y Mao, M Müller, M Fischer… - Advances in Neural …, 2023 - proceedings.neurips.cc
Training certifiably robust neural networks remains a notoriously hard problem. While
adversarial training optimizes under-approximations of the worst-case loss, which leads to …

The Triangular Trade-off between Robustness, Accuracy and Fairness in Deep Neural Networks: A Survey

J Li, G Li - ACM Computing Surveys, 2024 - dl.acm.org
With the rapid development of deep learning, AI systems are being used more in complex
and important domains and necessitates the simultaneous fulfillment of multiple constraints …