Picking the Cream of the Crop: Visual-Centric Data Selection with Collaborative Agents

Z Liu, Y Li, B Hu, W Luo, Y Wang, M Zhang - arxiv preprint arxiv …, 2025 - arxiv.org
To improve Multimodal Large Language Models'(MLLMs) ability to process images and
complex instructions, researchers predominantly curate large-scale visual instruction tuning …

RoVRM: A Robust Visual Reward Model Optimized via Auxiliary Textual Preference Data

C Wang, Y Gan, Y Huo, Y Mu, M Yang, Q He… - arxiv preprint arxiv …, 2024 - arxiv.org
Large vision-language models (LVLMs) often fail to align with human preferences, leading
to issues like generating misleading content without proper visual context (also known as …

[PDF][PDF] Continuous or Discrete, That Is the Question: A Survey on Large Multi-Modal Models from the Perspective of Input-Output Space Extension

Z Li, J Zhang, D Wang, Y Wang, X Huang, Z Wei - 2024 - preprints.org
With the success of large language models (LLMs) driving progress towards general-
purpose AI, there has been a growing focus on extending these models to multi-modal …