Dora: Weight-decomposed low-rank adaptation

SY Liu, CY Wang, H Yin, P Molchanov… - … on Machine Learning, 2024 - openreview.net
Among the widely used parameter-efficient fine-tuning (PEFT) methods, LoRA and its
variants have gained considerable popularity because of avoiding additional inference …

A survey on stability of learning with limited labelled data and its sensitivity to the effects of randomness

B Pecher, I Srba, M Bielikova - ACM Computing Surveys, 2024 - dl.acm.org
Learning with limited labelled data, such as prompting, in-context learning, fine-tuning, meta-
learning, or few-shot learning, aims to effectively train a model using only a small amount of …

Infoprompt: Information-theoretic soft prompt tuning for natural language understanding

J Wu, T Yu, R Wang, Z Song, R Zhang… - Advances in …, 2023 - proceedings.neurips.cc
Soft prompt tuning achieves superior performances across a wide range of few-shot tasks.
However, the performances of prompt tuning can be highly sensitive to the initialization of …

Aprompt: Attention prompt tuning for efficient adaptation of pre-trained language models

Q Wang, Y Mao, J Wang, H Yu, S Nie… - Proceedings of the …, 2023 - aclanthology.org
With the continuous growth of large language models, the process of fine-tuning these
models for new tasks has become increasingly parameter-intensive. Prompt tuning, a …

Prompting language-informed distribution for compositional zero-shot learning

W Bao, L Chen, H Huang, Y Kong - European Conference on Computer …, 2024 - Springer
Compositional zero-shot learning (CZSL) task aims to recognize unseen compositional
visual concepts, eg., sliced tomatoes, where the model is learned only from the seen …

Extending Whisper with prompt tuning to target-speaker ASR

H Ma, Z Peng, M Shao, J Li, J Liu - ICASSP 2024-2024 IEEE …, 2024 - ieeexplore.ieee.org
Target-speaker automatic speech recognition (ASR) aims to transcribe the desired speech
of a target speaker from multi-talker overlapped utterances. Most of the existing target …

[PDF][PDF] Flora: Low-rank core space for n-dimension

C Si, X Wang, X Yang, Z Xu, Q Li, J Dai… - arxiv preprint arxiv …, 2024 - researchgate.net
Adapting pre-trained foundation models for various downstream tasks has been prevalent in
artificial intelligence. Due to the vast number of tasks and high costs, adjusting all …

Promptintern: Saving inference costs by internalizing recurrent prompt during large language model fine-tuning

J Zou, M Zhou, T Li, S Han, D Zhang - arxiv preprint arxiv:2407.02211, 2024 - arxiv.org
Recent advances in fine-tuning large language models (LLMs) have greatly enhanced their
usage in domain-specific tasks. Despite the success, fine-tuning continues to rely on …

Dean: Deactivating the coupled neurons to mitigate fairness-privacy conflicts in large language models

C Qian, D Liu, J Zhang, Y Liu, J Shao - arxiv preprint arxiv:2410.16672, 2024 - arxiv.org
Ensuring awareness of fairness and privacy in Large Language Models (LLMs) is critical.
Interestingly, we discover a counter-intuitive trade-off phenomenon that enhancing an LLM's …

Non-intrusive adaptation: Input-centric parameter-efficient fine-tuning for versatile multimodal modeling

Y Wang, J Wu, T Dabral, J Zhang, G Brown… - arxiv preprint arxiv …, 2023 - arxiv.org
Large language models (LLMs) and vision language models (VLMs) demonstrate excellent
performance on a wide range of tasks by scaling up parameter counts from O (10^ 9) to O …