Black-Box Adversarial Attack on Vision Language Models for Autonomous Driving

L Wang, T Zhang, Y Qu, S Liang, Y Chen, A Liu… - arxiv preprint arxiv …, 2025 - arxiv.org
Vision-language models (VLMs) have significantly advanced autonomous driving (AD) by
enhancing reasoning capabilities; however, these models remain highly susceptible to …

Are VLMs Ready for Autonomous Driving? An Empirical Study from the Reliability, Data, and Metric Perspectives

S **e, L Kong, Y Dong, C Sima, W Zhang… - arxiv preprint arxiv …, 2025 - arxiv.org
Recent advancements in Vision-Language Models (VLMs) have sparked interest in their use
for autonomous driving, particularly in generating interpretable driving decisions through …

VLM-Bench: A Closer Look at How Well VLMs Implicitly Link Explicit Matching Visual Cues

J Zhang, D Yao, R Pi, PP Liang - arxiv preprint arxiv:2502.12084, 2025 - arxiv.org
Visually linking matching cues is a crucial ability in daily life, such as identifying the same
person in multiple photos based on their cues, even without knowing who they are. Despite …