Turbo: Informativity-driven acceleration plug-in for vision-language large models

C Ju, H Wang, H Cheng, X Chen, Z Zhai… - … on Computer Vision, 2024 - Springer
Abstract Vision-Language Large Models (VLMs) recently become primary backbone of AI,
due to the impressive performance. However, their expensive computation costs, ie …

Long-tailed diffusion models with oriented calibration

T Zhang, H Zheng, J Yao, X Wang, M Zhou… - The twelfth …, 2024 - openreview.net
Diffusion models are acclaimed for generating high-quality and diverse images. However,
their performance notably degrades when trained on data with a long-tailed distribution. For …

Denoiser: Rethinking the robustness for open-vocabulary action recognition

H Cheng, C Ju, H Wang, J Liu, M Chen, Q Hu… - arxiv preprint arxiv …, 2024 - arxiv.org
As one of the fundamental video tasks in computer vision, Open-Vocabulary Action
Recognition (OVAR) recently gains increasing attention, with the development of vision …

Turbo: informativity-driven acceleration plug-in for vision-language models

C Ju, H Wang, Z Li, X Chen, Z Zhai, W Huang… - arxiv preprint arxiv …, 2023 - arxiv.org
Vision-Language Large Models (VLMs) have become primary backbone of AI, due to the
impressive performance. However, their expensive computation costs, ie, throughput and …

Advancing Myopia To Holism: Fully Contrastive Language-Image Pre-training

H Wang, C Ju, W Lin, S **ao, M Chen, Y Huang… - arxiv preprint arxiv …, 2024 - arxiv.org
In rapidly evolving field of vision-language models (VLMs), contrastive language-image pre-
training (CLIP) has made significant strides, becoming foundation for various downstream …

Cross-domain recommendation via knowledge distillation

X Li, Z Huang, Z Wu, C Wang, Y Chen - Knowledge-Based Systems, 2025 - Elsevier
Recommendation systems frequently suffer from data sparsity, resulting in less-than-ideal
recommendations. A prominent solution to this problem is Cross-Domain Recommendation …

FOLDER: Accelerating Multi-modal Large Language Models with Enhanced Performance

H Wang, Z Yu, G Spadaro, C Ju, V Quétu… - arxiv preprint arxiv …, 2025 - arxiv.org
Recently, Multi-modal Large Language Models (MLLMs) have shown remarkable
effectiveness for multi-modal tasks due to their abilities to generate and understand cross …

Counterfactual Learning-Driven Representation Disentanglement for Search-Enhanced Recommendation

J Cui, X Chen, S **ao, C Ju, J Lan, Q Liu… - arxiv preprint arxiv …, 2024 - arxiv.org
For recommender systems in internet platforms, search activities provide additional insights
into user interest through query-click interactions with items, and are thus widely used for …

FARM: Frequency-Aware Model for Cross-Domain Live-Streaming Recommendation

X Li, R Yang, S Wen, S Wang, Y Liu, G Wang… - arxiv preprint arxiv …, 2025 - arxiv.org
Live-streaming services have attracted widespread popularity due to their real-time
interactivity and entertainment value. Users can engage with live-streaming authors by …

DIIT: A Domain-Invariant Information Transfer Method for Industrial Cross-Domain Recommendation

H Huang, X Lou, C Chen, P Cheng, Y **n… - Proceedings of the 33rd …, 2024 - dl.acm.org
Cross-Domain Recommendation (CDR) have received widespread attention due to their
ability to utilize rich information across domains. However, most existing CDR methods …