Unleashing the power of data tsunami: A comprehensive survey on data assessment and selection for instruction tuning of language models

Y Qin, Y Yang, P Guo, G Li, H Shao, Y Shi, Z Xu… - arxiv preprint arxiv …, 2024 - arxiv.org
Instruction tuning plays a critical role in aligning large language models (LLMs) with human
preference. Despite the vast amount of open instruction datasets, naively training a LLM on …

Enhancing visual-language modality alignment in large vision language models via self-improvement

X Wang, J Chen, Z Wang, Y Zhou, Y Zhou… - arxiv preprint arxiv …, 2024 - arxiv.org
Large vision-language models (LVLMs) have achieved impressive results in various visual
question-answering and reasoning tasks through vision instruction tuning on specific …

Tokenunify: Scalable autoregressive visual pre-training with mixture token prediction

Y Chen, H Shi, X Liu, T Shi, R Zhang, D Liu… - arxiv preprint arxiv …, 2024 - arxiv.org
Autoregressive next-token prediction is a standard pretraining method for large-scale
language models, but its application to vision tasks is hindered by the non-sequential nature …

Video-to-text pedestrian monitoring (VTPM): Leveraging computer vision and large Language Models for privacy-preserve pedestrian activity monitoring at …

AS Abdelrahman, M Abdel-Aty, D Wang - arxiv preprint arxiv:2408.11649, 2024 - arxiv.org
Computer vision has advanced research methodologies, enhancing system services across
various fields. It is a core component in traffic monitoring systems for improving road safety; …

Untangling the unrestricted web: Automatic identification of multilingual registers

E Henriksson, A Myntti, A Eskelinen… - arxiv preprint arxiv …, 2024 - arxiv.org
This article explores deep learning models for the automatic identification of registers-text
varieties such as news reports and discussion forums-in web-based datasets across 16 …

Residual-based language models are free boosters for biomedical imaging

Z Lai, J Wu, S Chen, Y Zhou, N Hovakimyan - arxiv preprint arxiv …, 2024 - arxiv.org
In this study, we uncover the unexpected efficacy of residual-based large language models
(LLMs) as part of encoders for biomedical imaging tasks, a domain traditionally devoid of …

Position paper: Data-centric ai in the age of large language models

X Xu, Z Wu, R Qiao, A Verma, Y Shu… - Findings of the …, 2024 - aclanthology.org
This position paper proposes a data-centric viewpoint of AI research, focusing on large
language models (LLMs). We start by making a key observation that data is instrumental in …

Data Management For Training Large Language Models: A Survey

Z Wang, W Zhong, Y Wang, Q Zhu, F Mi… - arxiv preprint arxiv …, 2023 - arxiv.org
Data plays a fundamental role in training Large Language Models (LLMs). Efficient data
management, particularly in formulating a well-suited training dataset, is significant for …

Data-centric ai in the age of large language models

X Xu, Z Wu, R Qiao, A Verma, Y Shu, J Wang… - arxiv preprint arxiv …, 2024 - arxiv.org
This position paper proposes a data-centric viewpoint of AI research, focusing on large
language models (LLMs). We start by making the key observation that data is instrumental in …

Csrec: Rethinking sequential recommendation from a causal perspective

X Liu, J Yuan, Y Zhou, J Li, F Huang, W Ai - arxiv preprint arxiv …, 2024 - arxiv.org
The essence of sequential recommender systems (RecSys) lies in understanding how users
make decisions. Most existing approaches frame the task as sequential prediction based on …