A survey on lora of large language models

Y Mao, Y Ge, Y Fan, W Xu, Y Mi, Z Hu… - Frontiers of Computer …, 2025 - Springer
Abstract Low-Rank Adaptation (LoRA), which updates the dense neural network layers with
pluggable low-rank matrices, is one of the best performed parameter efficient fine-tuning …

Localize-and-stitch: Efficient model merging via sparse task arithmetic

Y He, Y Hu, Y Lin, T Zhang, H Zhao - arxiv preprint arxiv:2408.13656, 2024 - arxiv.org
Model merging offers an effective strategy to combine the strengths of multiple finetuned
models into a unified model that preserves the specialized capabilities of each. Existing …

Compress then serve: Serving thousands of lora adapters with little overhead

R Brüel-Gabrielsson, J Zhu, O Bhardwaj… - arxiv preprint arxiv …, 2024 - arxiv.org
Fine-tuning large language models (LLMs) with low-rank adaptations (LoRAs) has become
common practice, often yielding numerous copies of the same LLM differing only in their …

Channel Merging: Preserving Specialization for Merged Experts

M Zhang, J Liu, G Ding, X Yu, L Ou… - arxiv preprint arxiv …, 2024 - arxiv.org
Lately, the practice of utilizing task-specific fine-tuning has been implemented to improve the
performance of large language models (LLM) in subsequent tasks. Through the integration …

[PDF][PDF] Low-Rank Adaptation for Scalable Fine-Tuning of Pre-Trained Language Models

H Dong, J Shun - 2025 - preprints.org
Low-Rank Adaptation (LoRA) is a computationally efficient approach for fine-tuning large
pre-trained language models, designed to reduce memory and computational overhead by …

Meta-Learning with Complex Tasks

W Jiang - 2024 - search.proquest.com
Meta-Learning aims at extracting shared knowledge (meta-knowledge) from historical tasks
to accelerate learning on new tasks. It has achieved promising performance in various …