Parameter-efficient fine-tuning for large models: A comprehensive survey

Z Han, C Gao, J Liu, J Zhang, SQ Zhang - arxiv preprint arxiv:2403.14608, 2024 - arxiv.org
Large models represent a groundbreaking advancement in multiple application fields,
enabling remarkable achievements across various tasks. However, their unprecedented …

A survey on lora of large language models

Y Mao, Y Ge, Y Fan, W Xu, Y Mi, Z Hu… - Frontiers of Computer …, 2025 - Springer
Abstract Low-Rank Adaptation (LoRA), which updates the dense neural network layers with
pluggable low-rank matrices, is one of the best performed parameter efficient fine-tuning …

Lora learns less and forgets less

D Biderman, J Portes, JJG Ortiz, M Paul… - arxiv preprint arxiv …, 2024 - arxiv.org
Low-Rank Adaptation (LoRA) is a widely-used parameter-efficient finetuning method for
large language models. LoRA saves memory by training only low rank perturbations to …

Reft: Representation finetuning for language models

Z Wu, A Arora, Z Wang, A Geiger, D Jurafsky… - arxiv preprint arxiv …, 2024 - arxiv.org
Parameter-efficient fine-tuning (PEFT) methods seek to adapt large models via updates to a
small number of weights. However, much prior interpretability work has shown that …

MELoRA: mini-ensemble low-rank adapters for parameter-efficient fine-tuning

P Ren, C Shi, S Wu, M Zhang, Z Ren… - Proceedings of the …, 2024 - aclanthology.org
Parameter-efficient fine-tuning (PEFT) is a popular method for tailoring pre-trained large
language models (LLMs), especially as the models' scale and the diversity of tasks increase …

Tied-lora: Enhacing parameter efficiency of lora with weight tying

A Renduchintala, T Konuk, O Kuchaiev - arxiv preprint arxiv:2311.09578, 2023 - arxiv.org
We propose Tied-LoRA, a simple paradigm utilizes weight tying and selective training to
further increase parameter efficiency of the Low-rank adaptation (LoRA) method. Our …

Lottery ticket adaptation: Mitigating destructive interference in llms

A Panda, B Isik, X Qi, S Koyejo, T Weissman… - arxiv preprint arxiv …, 2024 - arxiv.org
Existing methods for adapting large language models (LLMs) to new tasks are not suited to
multi-task adaptation because they modify all the model weights--causing destructive …

Lora+: Efficient low rank adaptation of large models

S Hayou, N Ghosh, B Yu - arxiv preprint arxiv:2402.12354, 2024 - arxiv.org
In this paper, we show that Low Rank Adaptation (LoRA) as originally introduced in Hu et
al.(2021) leads to suboptimal finetuning of models with large width (embedding dimension) …

Survey of different large language model architectures: Trends, benchmarks, and challenges

M Shao, A Basit, R Karri, M Shafique - IEEE Access, 2024 - ieeexplore.ieee.org
Large Language Models (LLMs) represent a class of deep learning models adept at
understanding natural language and generating coherent responses to various prompts or …

A Practitioner's Guide to Continual Multimodal Pretraining

K Roth, V Udandarao, S Dziadzio, A Prabhu… - arxiv preprint arxiv …, 2024 - arxiv.org
Multimodal foundation models serve numerous applications at the intersection of vision and
language. Still, despite being pretrained on extensive data, they become outdated over time …