Robust fast adaptation from adversarially explicit task distribution generation

C Wang, Y Lv, Y Mao, Y Qu, Y Xu, X Ji - arxiv preprint arxiv:2407.19523, 2024 - arxiv.org
Meta-learning is a practical learning paradigm to transfer skills across tasks from a few
examples. Nevertheless, the existence of task distribution shifts tends to weaken meta …

Beyond model adaptation at test time: A survey

Z **ao, CGM Snoek - arxiv preprint arxiv:2411.03687, 2024 - arxiv.org
Machine learning algorithms have achieved remarkable success across various disciplines,
use cases and applications, under the prevailing assumption that training and test samples …

A study of test-time contrastive concepts for open-world, open-vocabulary semantic segmentation

M Wysoczańska, A Vobecky, A Cardiel… - arxiv preprint arxiv …, 2024 - arxiv.org
Recent VLMs, pre-trained on large amounts of image-text pairs to align both modalities,
have opened the way to open-vocabulary semantic segmentation. Given an arbitrary set of …

IPO: Interpretable Prompt Optimization for Vision-Language Models

Y Du, W Sun, CGM Snoek - arxiv preprint arxiv:2410.15397, 2024 - arxiv.org
Pre-trained vision-language models like CLIP have remarkably adapted to various
downstream tasks. Nonetheless, their performance heavily depends on the specificity of the …

Prompt Diffusion Robustifies Any-Modality Prompt Learning

Y Du, G Liu, Y Shang, Y Yao, R Kompella… - arxiv preprint arxiv …, 2024 - arxiv.org
Foundation models enable prompt-based classifiers for zero-shot and few-shot learning.
Nonetheless, the conventional method of employing fixed prompts suffers from distributional …

Advances in Multimodal Adaptation and Generalization: From Traditional Approaches to Foundation Models

H Dong, M Liu, K Zhou, E Chatzi, J Kannala… - arxiv preprint arxiv …, 2025 - arxiv.org
In real-world scenarios, achieving domain adaptation and generalization poses significant
challenges, as models must adapt to or generalize across unknown target distributions …

DynaPrompt: Dynamic Test-Time Prompt Tuning

Z **ao, S Yan, J Hong, J Cai, X Jiang, Y Hu… - arxiv preprint arxiv …, 2025 - arxiv.org
Test-time prompt tuning enhances zero-shot generalization of vision-language models but
tends to ignore the relatedness among test samples during inference. Online test-time …

Parameter-Efficient Fine-Tuning for Foundation Models

D Zhang, T Feng, L Xue, Y Wang, Y Dong… - arxiv preprint arxiv …, 2025 - arxiv.org
This survey delves into the realm of Parameter-Efficient Fine-Tuning (PEFT) within the
context of Foundation Models (FMs). PEFT, a cost-effective fine-tuning technique, minimizes …

Technical note on calibrating vision-language models under covariate shift

B Khan, R Qureshi, T Syed - arxiv preprint arxiv:2502.07847, 2025 - arxiv.org
Despite being a successful example of emerging capability, vision-language foundation
models for low-shot vision classification have a limited ability to sufficiently generalize to the …

Complementary Subspace Low-Rank Adaptation of Vision-Language Models for Few-Shot Classification

Z Wang, J Dai, K Li, X Li, Y Guo, M **ang - arxiv preprint arxiv:2501.15040, 2025 - arxiv.org
Vision language model (VLM) has been designed for large scale image-text alignment as a
pretrained foundation model. For downstream few shot classification tasks, parameter …