Fundamental limits of prompt tuning transformers: Universality, capacity and efficiency
We investigate the statistical and computational limits of prompt tuning for transformer-based
foundation models. Our key contributions are prompt tuning on\textit {single-head} …
foundation models. Our key contributions are prompt tuning on\textit {single-head} …
Gita: Graph to visual and textual integration for vision-language graph reasoning
Large Language Models (LLMs) are increasingly used for various tasks with graph
structures. Though LLMs can process graph information in a textual format, they overlook the …
structures. Though LLMs can process graph information in a textual format, they overlook the …
Multi-Perspective Test-Time Prompt Tuning for Global, Local Visuals, and Language
Recent advances in vision-language models (VLMs) have demonstrated significant
generalization across a broad range of tasks through prompt learning. However, bridging …
generalization across a broad range of tasks through prompt learning. However, bridging …