Parameter-efficient fine-tuning for large models: A comprehensive survey

Z Han, C Gao, J Liu, J Zhang, SQ Zhang - arxiv preprint arxiv:2403.14608, 2024 - arxiv.org
Large models represent a groundbreaking advancement in multiple application fields,
enabling remarkable achievements across various tasks. However, their unprecedented …

A comprehensive survey on test-time adaptation under distribution shifts

J Liang, R He, T Tan - International Journal of Computer Vision, 2024 - Springer
Abstract Machine learning methods strive to acquire a robust model during the training
process that can effectively generalize to test samples, even in the presence of distribution …

Dual memory networks: A versatile adaptation approach for vision-language models

Y Zhang, W Zhu, H Tang, Z Ma… - Proceedings of the …, 2024 - openaccess.thecvf.com
With the emergence of pre-trained vision-language models like CLIP how to adapt them to
various downstream classification tasks has garnered significant attention in recent …

Multimodal Prompt Perceiver: Empower Adaptiveness Generalizability and Fidelity for All-in-One Image Restoration

Y Ai, H Huang, X Zhou, J Wang… - Proceedings of the IEEE …, 2024 - openaccess.thecvf.com
Despite substantial progress all-in-one image restoration (IR) grapples with persistent
challenges in handling intricate real-world degradations. This paper introduces MPerceiver …

On the test-time zero-shot generalization of vision-language models: Do we really need prompt learning?

M Zanella, I Ben Ayed - … of the IEEE/CVF Conference on …, 2024 - openaccess.thecvf.com
The development of large vision-language models notably CLIP has catalyzed research into
effective adaptation techniques with a particular focus on soft prompt tuning. Conjointly test …

Low-Rank Few-Shot Adaptation of Vision-Language Models

M Zanella, I Ben Ayed - … of the IEEE/CVF Conference on …, 2024 - openaccess.thecvf.com
Recent progress in the few-shot adaptation of Vision-Language Models (VLMs) has further
pushed their generalization capabilities at the expense of just a few labeled samples within …

Efficient Test-Time Adaptation of Vision-Language Models

A Karmanov, D Guan, S Lu… - Proceedings of the …, 2024 - openaccess.thecvf.com
Test-time adaptation with pre-trained vision-language models has attracted increasing
attention for tackling distribution shifts during the test time. Though prior studies have …

DiverGen: Improving Instance Segmentation by Learning Wider Data Distribution with More Diverse Generative Data

C Fan, M Zhu, H Chen, Y Liu, W Wu… - Proceedings of the …, 2024 - openaccess.thecvf.com
Instance segmentation is data-hungry and as model capacity increases data scale becomes
crucial for improving the accuracy. Most instance segmentation datasets today require costly …

Awt: Transferring vision-language models via augmentation, weighting, and transportation

Y Zhu, Y Ji, Z Zhao, G Wu, L Wang - arxiv preprint arxiv:2407.04603, 2024 - arxiv.org
Pre-trained vision-language models (VLMs) have shown impressive results in various visual
classification tasks. However, we often fail to fully unleash their potential when adapting …

Genview: Enhancing view quality with pretrained generative model for self-supervised learning

X Li, Y Yang, X Li, J Wu, Y Yu, B Ghanem… - European Conference on …, 2024 - Springer
Self-supervised learning has achieved remarkable success in acquiring high-quality
representations from unlabeled data. The widely adopted contrastive learning framework …