Parameter-efficient fine-tuning for large models: A comprehensive survey

Z Han, C Gao, J Liu, J Zhang, SQ Zhang - arxiv preprint arxiv:2403.14608, 2024‏ - arxiv.org
Large models represent a groundbreaking advancement in multiple application fields,
enabling remarkable achievements across various tasks. However, their unprecedented …

Parameter-efficient fine-tuning methods for pretrained language models: A critical review and assessment

L Xu, H **e, SZJ Qin, X Tao, FL Wang - arxiv preprint arxiv:2312.12148, 2023‏ - arxiv.org
With the continuous growth in the number of parameters of transformer-based pretrained
language models (PLMs), particularly the emergence of large language models (LLMs) with …

Efficient large language models: A survey

Z Wan, X Wang, C Liu, S Alam, Y Zheng, J Liu… - arxiv preprint arxiv …, 2023‏ - arxiv.org
Large Language Models (LLMs) have demonstrated remarkable capabilities in important
tasks such as natural language understanding and language generation, and thus have the …

Galore: Memory-efficient llm training by gradient low-rank projection

J Zhao, Z Zhang, B Chen, Z Wang… - arxiv preprint arxiv …, 2024‏ - arxiv.org
Training Large Language Models (LLMs) presents significant memory challenges,
predominantly due to the growing size of weights and optimizer states. Common memory …

Lorahub: Efficient cross-task generalization via dynamic lora composition

C Huang, Q Liu, BY Lin, T Pang, C Du, M Lin - arxiv preprint arxiv …, 2023‏ - arxiv.org
Low-rank adaptations (LoRA) are often employed to fine-tune large language models
(LLMs) for new tasks. This paper investigates LoRA composability for cross-task …

End-edge-cloud collaborative computing for deep learning: A comprehensive survey

Y Wang, C Yang, S Lan, L Zhu… - … Surveys & Tutorials, 2024‏ - ieeexplore.ieee.org
The booming development of deep learning applications and services heavily relies on
large deep learning models and massive data in the cloud. However, cloud-based deep …

Efficient multimodal large language models: A survey

Y **, J Li, Y Liu, T Gu, K Wu, Z Jiang, M He… - arxiv preprint arxiv …, 2024‏ - arxiv.org
In the past year, Multimodal Large Language Models (MLLMs) have demonstrated
remarkable performance in tasks such as visual question answering, visual understanding …

Hydralora: An asymmetric lora architecture for efficient fine-tuning

C Tian, Z Shi, Z Guo, L Li, CZ Xu - Advances in Neural …, 2025‏ - proceedings.neurips.cc
Abstract Adapting Large Language Models (LLMs) to new tasks through fine-tuning has
been made more efficient by the introduction of Parameter-Efficient Fine-Tuning (PEFT) …

Delta-lora: Fine-tuning high-rank parameters with the delta of low-rank matrices

B Zi, X Qi, L Wang, J Wang, KF Wong… - arxiv preprint arxiv …, 2023‏ - arxiv.org
In this paper, we present Delta-LoRA, which is a novel parameter-efficient approach to fine-
tune large language models (LLMs). In contrast to LoRA and other low-rank adaptation …

A survey on lora of large language models

Y Mao, Y Ge, Y Fan, W Xu, Y Mi, Z Hu… - Frontiers of Computer …, 2025‏ - Springer
Abstract Low-Rank Adaptation (LoRA), which updates the dense neural network layers with
pluggable low-rank matrices, is one of the best performed parameter efficient fine-tuning …