Recent advances of foundation language models-based continual learning: A survey

Y Yang, J Zhou, X Ding, T Huai, S Liu, Q Chen… - ACM Computing …, 2025 - dl.acm.org
Recently, foundation language models (LMs) have marked significant achievements in the
domains of natural language processing and computer vision. Unlike traditional neural …

Autonomous interactive correction MLLM for robust robotic manipulation

C **ong, C Shen, X Li, K Zhou, J Liu… - … Annual Conference on …, 2024 - openreview.net
The ability to reflect on and correct failures is crucial for robotic systems to interact stably with
real-life objects. Observing the generalization and reasoning capabilities of Multimodal …

AIC MLLM: Autonomous Interactive Correction MLLM for Robust Robotic Manipulation

C **ong, C Shen, X Li, K Zhou, J Liu, R Wang… - arxiv preprint arxiv …, 2024 - arxiv.org
The ability to reflect on and correct failures is crucial for robotic systems to interact stably with
real-life objects. Observing the generalization and reasoning capabilities of Multimodal …

LLM-controller: Dynamic robot control adaptation using large language models

R Zahedifar, MS Baghshah, A Taheri - Robotics and Autonomous Systems, 2025 - Elsevier
In this study, a dynamic adaptation of a robot controller is investigated using large language
models (LLMs). We propose our controller called the LLM-Controller, where, in response to …

Automating Robot Failure Recovery Using Vision-Language Models With Optimized Prompts

H Chen, Y Yao, R Liu, C Liu, J Ichnowski - arxiv preprint arxiv:2409.03966, 2024 - arxiv.org
Current robot autonomy struggles to operate beyond the assumed Operational Design
Domain (ODD), the specific set of conditions and environments in which the system is …

TrojanRobot: Physical-World Backdoor Attacks Against VLM-based Robotic Manipulation

X Wang, H Pan, H Zhang, M Li, S Hu, Z Zhou… - arxiv preprint arxiv …, 2024 - arxiv.org
Robotic manipulation in the physical world is increasingly empowered by\textit {large
language models}(LLMs) and\textit {vision-language models}(VLMs), leveraging their …

Diving into Self-Evolve Training for Multimodal Reasoning

W Liu, J Li, X Zhang, F Zhou, Y Cheng, J He - openreview.net
Reasoning ability is essential for Large Multimodal Models (LMMs). In the absence of
multimodal chain-of-thought annotated data, self-evolving training, where the model learns …